diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2009-09-14 20:53:36 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2009-09-14 20:53:36 -0400 |
commit | 39695224bd84dc4be29abad93a0ec232a16fc519 (patch) | |
tree | 2bfa5cb50788a4c8be9f2e9f4412e47a565f4508 /drivers | |
parent | a9bbd210a44102cc50b30a5f3d111dbf5f2f9cd4 (diff) | |
parent | ea038f63ac52439e7816295fa6064fe95e6c1f51 (diff) |
Merge git://git.kernel.org/pub/scm/linux/kernel/git/jejb/scsi-misc-2.6
* git://git.kernel.org/pub/scm/linux/kernel/git/jejb/scsi-misc-2.6: (209 commits)
[SCSI] fix oops during scsi scanning
[SCSI] libsrp: fix memory leak in srp_ring_free()
[SCSI] libiscsi, bnx2i: make bound ep check common
[SCSI] libiscsi: add completion function for drivers that do not need pdu processing
[SCSI] scsi_dh_rdac: changes for rdac debug logging
[SCSI] scsi_dh_rdac: changes to collect the rdac debug information during the initialization
[SCSI] scsi_dh_rdac: move the init code from rdac_activate to rdac_bus_attach
[SCSI] sg: fix oops in the error path in sg_build_indirect()
[SCSI] mptsas : Bump version to 3.04.12
[SCSI] mptsas : FW event thread and scsi mid layer deadlock in SYNCHRONIZE CACHE command
[SCSI] mptsas : Send DID_NO_CONNECT for pending IOs of removed device
[SCSI] mptsas : PAE Kernel more than 4 GB kernel panic
[SCSI] mptsas : NULL pointer on big endian systems causing Expander not to tear off
[SCSI] mptsas : Sanity check for phyinfo is added
[SCSI] scsi_dh_rdac: Add support for Sun StorageTek ST2500, ST2510 and ST2530
[SCSI] pmcraid: PMC-Sierra MaxRAID driver to support 6Gb/s SAS RAID controller
[SCSI] qla2xxx: Update version number to 8.03.01-k6.
[SCSI] qla2xxx: Properly delete rports attached to a vport.
[SCSI] qla2xxx: Correct various NPIV issues.
[SCSI] qla2xxx: Correct qla2x00_eh_wait_on_command() to wait correctly.
...
Diffstat (limited to 'drivers')
99 files changed, 14174 insertions, 4714 deletions
diff --git a/drivers/md/dm-mpath.c b/drivers/md/dm-mpath.c index 6f0d90d4a541..32d0b878eccc 100644 --- a/drivers/md/dm-mpath.c +++ b/drivers/md/dm-mpath.c | |||
@@ -64,6 +64,7 @@ struct multipath { | |||
64 | spinlock_t lock; | 64 | spinlock_t lock; |
65 | 65 | ||
66 | const char *hw_handler_name; | 66 | const char *hw_handler_name; |
67 | char *hw_handler_params; | ||
67 | unsigned nr_priority_groups; | 68 | unsigned nr_priority_groups; |
68 | struct list_head priority_groups; | 69 | struct list_head priority_groups; |
69 | unsigned pg_init_required; /* pg_init needs calling? */ | 70 | unsigned pg_init_required; /* pg_init needs calling? */ |
@@ -219,6 +220,7 @@ static void free_multipath(struct multipath *m) | |||
219 | } | 220 | } |
220 | 221 | ||
221 | kfree(m->hw_handler_name); | 222 | kfree(m->hw_handler_name); |
223 | kfree(m->hw_handler_params); | ||
222 | mempool_destroy(m->mpio_pool); | 224 | mempool_destroy(m->mpio_pool); |
223 | kfree(m); | 225 | kfree(m); |
224 | } | 226 | } |
@@ -615,6 +617,17 @@ static struct pgpath *parse_path(struct arg_set *as, struct path_selector *ps, | |||
615 | dm_put_device(ti, p->path.dev); | 617 | dm_put_device(ti, p->path.dev); |
616 | goto bad; | 618 | goto bad; |
617 | } | 619 | } |
620 | |||
621 | if (m->hw_handler_params) { | ||
622 | r = scsi_dh_set_params(q, m->hw_handler_params); | ||
623 | if (r < 0) { | ||
624 | ti->error = "unable to set hardware " | ||
625 | "handler parameters"; | ||
626 | scsi_dh_detach(q); | ||
627 | dm_put_device(ti, p->path.dev); | ||
628 | goto bad; | ||
629 | } | ||
630 | } | ||
618 | } | 631 | } |
619 | 632 | ||
620 | r = ps->type->add_path(ps, &p->path, as->argc, as->argv, &ti->error); | 633 | r = ps->type->add_path(ps, &p->path, as->argc, as->argv, &ti->error); |
@@ -705,6 +718,7 @@ static struct priority_group *parse_priority_group(struct arg_set *as, | |||
705 | static int parse_hw_handler(struct arg_set *as, struct multipath *m) | 718 | static int parse_hw_handler(struct arg_set *as, struct multipath *m) |
706 | { | 719 | { |
707 | unsigned hw_argc; | 720 | unsigned hw_argc; |
721 | int ret; | ||
708 | struct dm_target *ti = m->ti; | 722 | struct dm_target *ti = m->ti; |
709 | 723 | ||
710 | static struct param _params[] = { | 724 | static struct param _params[] = { |
@@ -726,17 +740,33 @@ static int parse_hw_handler(struct arg_set *as, struct multipath *m) | |||
726 | request_module("scsi_dh_%s", m->hw_handler_name); | 740 | request_module("scsi_dh_%s", m->hw_handler_name); |
727 | if (scsi_dh_handler_exist(m->hw_handler_name) == 0) { | 741 | if (scsi_dh_handler_exist(m->hw_handler_name) == 0) { |
728 | ti->error = "unknown hardware handler type"; | 742 | ti->error = "unknown hardware handler type"; |
729 | kfree(m->hw_handler_name); | 743 | ret = -EINVAL; |
730 | m->hw_handler_name = NULL; | 744 | goto fail; |
731 | return -EINVAL; | ||
732 | } | 745 | } |
733 | 746 | ||
734 | if (hw_argc > 1) | 747 | if (hw_argc > 1) { |
735 | DMWARN("Ignoring user-specified arguments for " | 748 | char *p; |
736 | "hardware handler \"%s\"", m->hw_handler_name); | 749 | int i, j, len = 4; |
750 | |||
751 | for (i = 0; i <= hw_argc - 2; i++) | ||
752 | len += strlen(as->argv[i]) + 1; | ||
753 | p = m->hw_handler_params = kzalloc(len, GFP_KERNEL); | ||
754 | if (!p) { | ||
755 | ti->error = "memory allocation failed"; | ||
756 | ret = -ENOMEM; | ||
757 | goto fail; | ||
758 | } | ||
759 | j = sprintf(p, "%d", hw_argc - 1); | ||
760 | for (i = 0, p+=j+1; i <= hw_argc - 2; i++, p+=j+1) | ||
761 | j = sprintf(p, "%s", as->argv[i]); | ||
762 | } | ||
737 | consume(as, hw_argc - 1); | 763 | consume(as, hw_argc - 1); |
738 | 764 | ||
739 | return 0; | 765 | return 0; |
766 | fail: | ||
767 | kfree(m->hw_handler_name); | ||
768 | m->hw_handler_name = NULL; | ||
769 | return ret; | ||
740 | } | 770 | } |
741 | 771 | ||
742 | static int parse_features(struct arg_set *as, struct multipath *m) | 772 | static int parse_features(struct arg_set *as, struct multipath *m) |
diff --git a/drivers/message/fusion/mptbase.c b/drivers/message/fusion/mptbase.c index 5d0ba4f5924c..76fa2ee0b574 100644 --- a/drivers/message/fusion/mptbase.c +++ b/drivers/message/fusion/mptbase.c | |||
@@ -1015,9 +1015,9 @@ mpt_add_sge_64bit(void *pAddr, u32 flagslength, dma_addr_t dma_addr) | |||
1015 | { | 1015 | { |
1016 | SGESimple64_t *pSge = (SGESimple64_t *) pAddr; | 1016 | SGESimple64_t *pSge = (SGESimple64_t *) pAddr; |
1017 | pSge->Address.Low = cpu_to_le32 | 1017 | pSge->Address.Low = cpu_to_le32 |
1018 | (lower_32_bits((unsigned long)(dma_addr))); | 1018 | (lower_32_bits(dma_addr)); |
1019 | pSge->Address.High = cpu_to_le32 | 1019 | pSge->Address.High = cpu_to_le32 |
1020 | (upper_32_bits((unsigned long)dma_addr)); | 1020 | (upper_32_bits(dma_addr)); |
1021 | pSge->FlagsLength = cpu_to_le32 | 1021 | pSge->FlagsLength = cpu_to_le32 |
1022 | ((flagslength | MPT_SGE_FLAGS_64_BIT_ADDRESSING)); | 1022 | ((flagslength | MPT_SGE_FLAGS_64_BIT_ADDRESSING)); |
1023 | } | 1023 | } |
@@ -1038,8 +1038,8 @@ mpt_add_sge_64bit_1078(void *pAddr, u32 flagslength, dma_addr_t dma_addr) | |||
1038 | u32 tmp; | 1038 | u32 tmp; |
1039 | 1039 | ||
1040 | pSge->Address.Low = cpu_to_le32 | 1040 | pSge->Address.Low = cpu_to_le32 |
1041 | (lower_32_bits((unsigned long)(dma_addr))); | 1041 | (lower_32_bits(dma_addr)); |
1042 | tmp = (u32)(upper_32_bits((unsigned long)dma_addr)); | 1042 | tmp = (u32)(upper_32_bits(dma_addr)); |
1043 | 1043 | ||
1044 | /* | 1044 | /* |
1045 | * 1078 errata workaround for the 36GB limitation | 1045 | * 1078 errata workaround for the 36GB limitation |
@@ -1101,7 +1101,7 @@ mpt_add_chain_64bit(void *pAddr, u8 next, u16 length, dma_addr_t dma_addr) | |||
1101 | pChain->NextChainOffset = next; | 1101 | pChain->NextChainOffset = next; |
1102 | 1102 | ||
1103 | pChain->Address.Low = cpu_to_le32(tmp); | 1103 | pChain->Address.Low = cpu_to_le32(tmp); |
1104 | tmp = (u32)(upper_32_bits((unsigned long)dma_addr)); | 1104 | tmp = (u32)(upper_32_bits(dma_addr)); |
1105 | pChain->Address.High = cpu_to_le32(tmp); | 1105 | pChain->Address.High = cpu_to_le32(tmp); |
1106 | } | 1106 | } |
1107 | 1107 | ||
@@ -1297,12 +1297,8 @@ mpt_host_page_alloc(MPT_ADAPTER *ioc, pIOCInit_t ioc_init) | |||
1297 | psge = (char *)&ioc_init->HostPageBufferSGE; | 1297 | psge = (char *)&ioc_init->HostPageBufferSGE; |
1298 | flags_length = MPI_SGE_FLAGS_SIMPLE_ELEMENT | | 1298 | flags_length = MPI_SGE_FLAGS_SIMPLE_ELEMENT | |
1299 | MPI_SGE_FLAGS_SYSTEM_ADDRESS | | 1299 | MPI_SGE_FLAGS_SYSTEM_ADDRESS | |
1300 | MPI_SGE_FLAGS_32_BIT_ADDRESSING | | ||
1301 | MPI_SGE_FLAGS_HOST_TO_IOC | | 1300 | MPI_SGE_FLAGS_HOST_TO_IOC | |
1302 | MPI_SGE_FLAGS_END_OF_BUFFER; | 1301 | MPI_SGE_FLAGS_END_OF_BUFFER; |
1303 | if (sizeof(dma_addr_t) == sizeof(u64)) { | ||
1304 | flags_length |= MPI_SGE_FLAGS_64_BIT_ADDRESSING; | ||
1305 | } | ||
1306 | flags_length = flags_length << MPI_SGE_FLAGS_SHIFT; | 1302 | flags_length = flags_length << MPI_SGE_FLAGS_SHIFT; |
1307 | flags_length |= ioc->HostPageBuffer_sz; | 1303 | flags_length |= ioc->HostPageBuffer_sz; |
1308 | ioc->add_sge(psge, flags_length, ioc->HostPageBuffer_dma); | 1304 | ioc->add_sge(psge, flags_length, ioc->HostPageBuffer_dma); |
@@ -2224,8 +2220,6 @@ mpt_do_ioc_recovery(MPT_ADAPTER *ioc, u32 reason, int sleepFlag) | |||
2224 | int hard; | 2220 | int hard; |
2225 | int rc=0; | 2221 | int rc=0; |
2226 | int ii; | 2222 | int ii; |
2227 | u8 cb_idx; | ||
2228 | int handlers; | ||
2229 | int ret = 0; | 2223 | int ret = 0; |
2230 | int reset_alt_ioc_active = 0; | 2224 | int reset_alt_ioc_active = 0; |
2231 | int irq_allocated = 0; | 2225 | int irq_allocated = 0; |
@@ -2548,34 +2542,6 @@ mpt_do_ioc_recovery(MPT_ADAPTER *ioc, u32 reason, int sleepFlag) | |||
2548 | mpt_get_manufacturing_pg_0(ioc); | 2542 | mpt_get_manufacturing_pg_0(ioc); |
2549 | } | 2543 | } |
2550 | 2544 | ||
2551 | /* | ||
2552 | * Call each currently registered protocol IOC reset handler | ||
2553 | * with post-reset indication. | ||
2554 | * NOTE: If we're doing _IOC_BRINGUP, there can be no | ||
2555 | * MptResetHandlers[] registered yet. | ||
2556 | */ | ||
2557 | if (hard_reset_done) { | ||
2558 | rc = handlers = 0; | ||
2559 | for (cb_idx = MPT_MAX_PROTOCOL_DRIVERS-1; cb_idx; cb_idx--) { | ||
2560 | if ((ret == 0) && MptResetHandlers[cb_idx]) { | ||
2561 | dprintk(ioc, printk(MYIOC_s_DEBUG_FMT | ||
2562 | "Calling IOC post_reset handler #%d\n", | ||
2563 | ioc->name, cb_idx)); | ||
2564 | rc += mpt_signal_reset(cb_idx, ioc, MPT_IOC_POST_RESET); | ||
2565 | handlers++; | ||
2566 | } | ||
2567 | |||
2568 | if (alt_ioc_ready && MptResetHandlers[cb_idx]) { | ||
2569 | drsprintk(ioc, printk(MYIOC_s_DEBUG_FMT | ||
2570 | "Calling IOC post_reset handler #%d\n", | ||
2571 | ioc->alt_ioc->name, cb_idx)); | ||
2572 | rc += mpt_signal_reset(cb_idx, ioc->alt_ioc, MPT_IOC_POST_RESET); | ||
2573 | handlers++; | ||
2574 | } | ||
2575 | } | ||
2576 | /* FIXME? Examine results here? */ | ||
2577 | } | ||
2578 | |||
2579 | out: | 2545 | out: |
2580 | if ((ret != 0) && irq_allocated) { | 2546 | if ((ret != 0) && irq_allocated) { |
2581 | free_irq(ioc->pci_irq, ioc); | 2547 | free_irq(ioc->pci_irq, ioc); |
@@ -3938,6 +3904,7 @@ mpt_diag_reset(MPT_ADAPTER *ioc, int ignore, int sleepFlag) | |||
3938 | int count = 0; | 3904 | int count = 0; |
3939 | u32 diag1val = 0; | 3905 | u32 diag1val = 0; |
3940 | MpiFwHeader_t *cached_fw; /* Pointer to FW */ | 3906 | MpiFwHeader_t *cached_fw; /* Pointer to FW */ |
3907 | u8 cb_idx; | ||
3941 | 3908 | ||
3942 | /* Clear any existing interrupts */ | 3909 | /* Clear any existing interrupts */ |
3943 | CHIPREG_WRITE32(&ioc->chip->IntStatus, 0); | 3910 | CHIPREG_WRITE32(&ioc->chip->IntStatus, 0); |
@@ -3956,6 +3923,18 @@ mpt_diag_reset(MPT_ADAPTER *ioc, int ignore, int sleepFlag) | |||
3956 | else | 3923 | else |
3957 | mdelay(1); | 3924 | mdelay(1); |
3958 | 3925 | ||
3926 | /* | ||
3927 | * Call each currently registered protocol IOC reset handler | ||
3928 | * with pre-reset indication. | ||
3929 | * NOTE: If we're doing _IOC_BRINGUP, there can be no | ||
3930 | * MptResetHandlers[] registered yet. | ||
3931 | */ | ||
3932 | for (cb_idx = MPT_MAX_PROTOCOL_DRIVERS-1; cb_idx; cb_idx--) { | ||
3933 | if (MptResetHandlers[cb_idx]) | ||
3934 | (*(MptResetHandlers[cb_idx]))(ioc, | ||
3935 | MPT_IOC_PRE_RESET); | ||
3936 | } | ||
3937 | |||
3959 | for (count = 0; count < 60; count ++) { | 3938 | for (count = 0; count < 60; count ++) { |
3960 | doorbell = CHIPREG_READ32(&ioc->chip->Doorbell); | 3939 | doorbell = CHIPREG_READ32(&ioc->chip->Doorbell); |
3961 | doorbell &= MPI_IOC_STATE_MASK; | 3940 | doorbell &= MPI_IOC_STATE_MASK; |
@@ -4052,25 +4031,15 @@ mpt_diag_reset(MPT_ADAPTER *ioc, int ignore, int sleepFlag) | |||
4052 | * NOTE: If we're doing _IOC_BRINGUP, there can be no | 4031 | * NOTE: If we're doing _IOC_BRINGUP, there can be no |
4053 | * MptResetHandlers[] registered yet. | 4032 | * MptResetHandlers[] registered yet. |
4054 | */ | 4033 | */ |
4055 | { | 4034 | for (cb_idx = MPT_MAX_PROTOCOL_DRIVERS-1; cb_idx; cb_idx--) { |
4056 | u8 cb_idx; | 4035 | if (MptResetHandlers[cb_idx]) { |
4057 | int r = 0; | 4036 | mpt_signal_reset(cb_idx, |
4058 | 4037 | ioc, MPT_IOC_PRE_RESET); | |
4059 | for (cb_idx = MPT_MAX_PROTOCOL_DRIVERS-1; cb_idx; cb_idx--) { | 4038 | if (ioc->alt_ioc) { |
4060 | if (MptResetHandlers[cb_idx]) { | 4039 | mpt_signal_reset(cb_idx, |
4061 | dprintk(ioc, printk(MYIOC_s_DEBUG_FMT | 4040 | ioc->alt_ioc, MPT_IOC_PRE_RESET); |
4062 | "Calling IOC pre_reset handler #%d\n", | ||
4063 | ioc->name, cb_idx)); | ||
4064 | r += mpt_signal_reset(cb_idx, ioc, MPT_IOC_PRE_RESET); | ||
4065 | if (ioc->alt_ioc) { | ||
4066 | dprintk(ioc, printk(MYIOC_s_DEBUG_FMT | ||
4067 | "Calling alt-%s pre_reset handler #%d\n", | ||
4068 | ioc->name, ioc->alt_ioc->name, cb_idx)); | ||
4069 | r += mpt_signal_reset(cb_idx, ioc->alt_ioc, MPT_IOC_PRE_RESET); | ||
4070 | } | ||
4071 | } | 4041 | } |
4072 | } | 4042 | } |
4073 | /* FIXME? Examine results here? */ | ||
4074 | } | 4043 | } |
4075 | 4044 | ||
4076 | if (ioc->cached_fw) | 4045 | if (ioc->cached_fw) |
@@ -6956,7 +6925,7 @@ EXPORT_SYMBOL(mpt_halt_firmware); | |||
6956 | int | 6925 | int |
6957 | mpt_HardResetHandler(MPT_ADAPTER *ioc, int sleepFlag) | 6926 | mpt_HardResetHandler(MPT_ADAPTER *ioc, int sleepFlag) |
6958 | { | 6927 | { |
6959 | int rc; | 6928 | int rc; |
6960 | u8 cb_idx; | 6929 | u8 cb_idx; |
6961 | unsigned long flags; | 6930 | unsigned long flags; |
6962 | unsigned long time_count; | 6931 | unsigned long time_count; |
@@ -6982,8 +6951,6 @@ mpt_HardResetHandler(MPT_ADAPTER *ioc, int sleepFlag) | |||
6982 | ioc->alt_ioc->ioc_reset_in_progress = 1; | 6951 | ioc->alt_ioc->ioc_reset_in_progress = 1; |
6983 | spin_unlock_irqrestore(&ioc->taskmgmt_lock, flags); | 6952 | spin_unlock_irqrestore(&ioc->taskmgmt_lock, flags); |
6984 | 6953 | ||
6985 | /* FIXME: If do_ioc_recovery fails, repeat.... | ||
6986 | */ | ||
6987 | 6954 | ||
6988 | /* The SCSI driver needs to adjust timeouts on all current | 6955 | /* The SCSI driver needs to adjust timeouts on all current |
6989 | * commands prior to the diagnostic reset being issued. | 6956 | * commands prior to the diagnostic reset being issued. |
@@ -7020,6 +6987,15 @@ mpt_HardResetHandler(MPT_ADAPTER *ioc, int sleepFlag) | |||
7020 | } | 6987 | } |
7021 | spin_unlock_irqrestore(&ioc->taskmgmt_lock, flags); | 6988 | spin_unlock_irqrestore(&ioc->taskmgmt_lock, flags); |
7022 | 6989 | ||
6990 | for (cb_idx = MPT_MAX_PROTOCOL_DRIVERS-1; cb_idx; cb_idx--) { | ||
6991 | if (MptResetHandlers[cb_idx]) { | ||
6992 | mpt_signal_reset(cb_idx, ioc, MPT_IOC_POST_RESET); | ||
6993 | if (ioc->alt_ioc) | ||
6994 | mpt_signal_reset(cb_idx, | ||
6995 | ioc->alt_ioc, MPT_IOC_POST_RESET); | ||
6996 | } | ||
6997 | } | ||
6998 | |||
7023 | dtmprintk(ioc, | 6999 | dtmprintk(ioc, |
7024 | printk(MYIOC_s_DEBUG_FMT | 7000 | printk(MYIOC_s_DEBUG_FMT |
7025 | "HardResetHandler: completed (%d seconds): %s\n", ioc->name, | 7001 | "HardResetHandler: completed (%d seconds): %s\n", ioc->name, |
diff --git a/drivers/message/fusion/mptbase.h b/drivers/message/fusion/mptbase.h index 1c8514dc31ca..8dd4d219e433 100644 --- a/drivers/message/fusion/mptbase.h +++ b/drivers/message/fusion/mptbase.h | |||
@@ -76,8 +76,8 @@ | |||
76 | #define COPYRIGHT "Copyright (c) 1999-2008 " MODULEAUTHOR | 76 | #define COPYRIGHT "Copyright (c) 1999-2008 " MODULEAUTHOR |
77 | #endif | 77 | #endif |
78 | 78 | ||
79 | #define MPT_LINUX_VERSION_COMMON "3.04.10" | 79 | #define MPT_LINUX_VERSION_COMMON "3.04.12" |
80 | #define MPT_LINUX_PACKAGE_NAME "@(#)mptlinux-3.04.09" | 80 | #define MPT_LINUX_PACKAGE_NAME "@(#)mptlinux-3.04.12" |
81 | #define WHAT_MAGIC_STRING "@" "(" "#" ")" | 81 | #define WHAT_MAGIC_STRING "@" "(" "#" ")" |
82 | 82 | ||
83 | #define show_mptmod_ver(s,ver) \ | 83 | #define show_mptmod_ver(s,ver) \ |
@@ -157,8 +157,9 @@ | |||
157 | /* | 157 | /* |
158 | * Try to keep these at 2^N-1 | 158 | * Try to keep these at 2^N-1 |
159 | */ | 159 | */ |
160 | #define MPT_FC_CAN_QUEUE 127 | 160 | #define MPT_FC_CAN_QUEUE 1024 |
161 | #define MPT_SCSI_CAN_QUEUE 127 | 161 | #define MPT_SCSI_CAN_QUEUE 127 |
162 | #define MPT_SAS_CAN_QUEUE 127 | ||
162 | 163 | ||
163 | /* | 164 | /* |
164 | * Set the MAX_SGE value based on user input. | 165 | * Set the MAX_SGE value based on user input. |
@@ -879,23 +880,9 @@ typedef enum { | |||
879 | 880 | ||
880 | typedef struct _MPT_SCSI_HOST { | 881 | typedef struct _MPT_SCSI_HOST { |
881 | MPT_ADAPTER *ioc; | 882 | MPT_ADAPTER *ioc; |
882 | int port; | ||
883 | u32 pad0; | ||
884 | MPT_LOCAL_REPLY *pLocal; /* used for internal commands */ | ||
885 | struct timer_list timer; | ||
886 | /* Pool of memory for holding SCpnts before doing | ||
887 | * OS callbacks. freeQ is the free pool. | ||
888 | */ | ||
889 | u8 negoNvram; /* DV disabled, nego NVRAM */ | ||
890 | u8 pad1; | ||
891 | u8 rsvd[2]; | ||
892 | MPT_FRAME_HDR *cmdPtr; /* Ptr to nonOS request */ | ||
893 | struct scsi_cmnd *abortSCpnt; | ||
894 | MPT_LOCAL_REPLY localReply; /* internal cmd reply struct */ | ||
895 | ushort sel_timeout[MPT_MAX_FC_DEVICES]; | 883 | ushort sel_timeout[MPT_MAX_FC_DEVICES]; |
896 | char *info_kbuf; | 884 | char *info_kbuf; |
897 | long last_queue_full; | 885 | long last_queue_full; |
898 | u16 tm_iocstatus; | ||
899 | u16 spi_pending; | 886 | u16 spi_pending; |
900 | struct list_head target_reset_list; | 887 | struct list_head target_reset_list; |
901 | } MPT_SCSI_HOST; | 888 | } MPT_SCSI_HOST; |
diff --git a/drivers/message/fusion/mptfc.c b/drivers/message/fusion/mptfc.c index e61df133a59e..ebf6ae024da4 100644 --- a/drivers/message/fusion/mptfc.c +++ b/drivers/message/fusion/mptfc.c | |||
@@ -1288,25 +1288,6 @@ mptfc_probe(struct pci_dev *pdev, const struct pci_device_id *id) | |||
1288 | dprintk(ioc, printk(MYIOC_s_DEBUG_FMT "ScsiLookup @ %p\n", | 1288 | dprintk(ioc, printk(MYIOC_s_DEBUG_FMT "ScsiLookup @ %p\n", |
1289 | ioc->name, ioc->ScsiLookup)); | 1289 | ioc->name, ioc->ScsiLookup)); |
1290 | 1290 | ||
1291 | /* Clear the TM flags | ||
1292 | */ | ||
1293 | hd->abortSCpnt = NULL; | ||
1294 | |||
1295 | /* Clear the pointer used to store | ||
1296 | * single-threaded commands, i.e., those | ||
1297 | * issued during a bus scan, dv and | ||
1298 | * configuration pages. | ||
1299 | */ | ||
1300 | hd->cmdPtr = NULL; | ||
1301 | |||
1302 | /* Initialize this SCSI Hosts' timers | ||
1303 | * To use, set the timer expires field | ||
1304 | * and add_timer | ||
1305 | */ | ||
1306 | init_timer(&hd->timer); | ||
1307 | hd->timer.data = (unsigned long) hd; | ||
1308 | hd->timer.function = mptscsih_timer_expired; | ||
1309 | |||
1310 | hd->last_queue_full = 0; | 1291 | hd->last_queue_full = 0; |
1311 | 1292 | ||
1312 | sh->transportt = mptfc_transport_template; | 1293 | sh->transportt = mptfc_transport_template; |
diff --git a/drivers/message/fusion/mptsas.c b/drivers/message/fusion/mptsas.c index 55ff25244af4..83873e3d0ce7 100644 --- a/drivers/message/fusion/mptsas.c +++ b/drivers/message/fusion/mptsas.c | |||
@@ -72,6 +72,7 @@ | |||
72 | */ | 72 | */ |
73 | #define MPTSAS_RAID_CHANNEL 1 | 73 | #define MPTSAS_RAID_CHANNEL 1 |
74 | 74 | ||
75 | #define SAS_CONFIG_PAGE_TIMEOUT 30 | ||
75 | MODULE_AUTHOR(MODULEAUTHOR); | 76 | MODULE_AUTHOR(MODULEAUTHOR); |
76 | MODULE_DESCRIPTION(my_NAME); | 77 | MODULE_DESCRIPTION(my_NAME); |
77 | MODULE_LICENSE("GPL"); | 78 | MODULE_LICENSE("GPL"); |
@@ -324,7 +325,6 @@ mptsas_cleanup_fw_event_q(MPT_ADAPTER *ioc) | |||
324 | { | 325 | { |
325 | struct fw_event_work *fw_event, *next; | 326 | struct fw_event_work *fw_event, *next; |
326 | struct mptsas_target_reset_event *target_reset_list, *n; | 327 | struct mptsas_target_reset_event *target_reset_list, *n; |
327 | u8 flush_q; | ||
328 | MPT_SCSI_HOST *hd = shost_priv(ioc->sh); | 328 | MPT_SCSI_HOST *hd = shost_priv(ioc->sh); |
329 | 329 | ||
330 | /* flush the target_reset_list */ | 330 | /* flush the target_reset_list */ |
@@ -344,15 +344,10 @@ mptsas_cleanup_fw_event_q(MPT_ADAPTER *ioc) | |||
344 | !ioc->fw_event_q || in_interrupt()) | 344 | !ioc->fw_event_q || in_interrupt()) |
345 | return; | 345 | return; |
346 | 346 | ||
347 | flush_q = 0; | ||
348 | list_for_each_entry_safe(fw_event, next, &ioc->fw_event_list, list) { | 347 | list_for_each_entry_safe(fw_event, next, &ioc->fw_event_list, list) { |
349 | if (cancel_delayed_work(&fw_event->work)) | 348 | if (cancel_delayed_work(&fw_event->work)) |
350 | mptsas_free_fw_event(ioc, fw_event); | 349 | mptsas_free_fw_event(ioc, fw_event); |
351 | else | ||
352 | flush_q = 1; | ||
353 | } | 350 | } |
354 | if (flush_q) | ||
355 | flush_workqueue(ioc->fw_event_q); | ||
356 | } | 351 | } |
357 | 352 | ||
358 | 353 | ||
@@ -661,7 +656,7 @@ mptsas_add_device_component_starget_ir(MPT_ADAPTER *ioc, | |||
661 | cfg.pageAddr = starget->id; | 656 | cfg.pageAddr = starget->id; |
662 | cfg.cfghdr.hdr = &hdr; | 657 | cfg.cfghdr.hdr = &hdr; |
663 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; | 658 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; |
664 | cfg.timeout = 10; | 659 | cfg.timeout = SAS_CONFIG_PAGE_TIMEOUT; |
665 | 660 | ||
666 | if (mpt_config(ioc, &cfg) != 0) | 661 | if (mpt_config(ioc, &cfg) != 0) |
667 | goto out; | 662 | goto out; |
@@ -851,7 +846,13 @@ mptsas_setup_wide_ports(MPT_ADAPTER *ioc, struct mptsas_portinfo *port_info) | |||
851 | port_details->num_phys--; | 846 | port_details->num_phys--; |
852 | port_details->phy_bitmask &= ~ (1 << phy_info->phy_id); | 847 | port_details->phy_bitmask &= ~ (1 << phy_info->phy_id); |
853 | memset(&phy_info->attached, 0, sizeof(struct mptsas_devinfo)); | 848 | memset(&phy_info->attached, 0, sizeof(struct mptsas_devinfo)); |
854 | sas_port_delete_phy(port_details->port, phy_info->phy); | 849 | if (phy_info->phy) { |
850 | devtprintk(ioc, dev_printk(KERN_DEBUG, | ||
851 | &phy_info->phy->dev, MYIOC_s_FMT | ||
852 | "delete phy %d, phy-obj (0x%p)\n", ioc->name, | ||
853 | phy_info->phy_id, phy_info->phy)); | ||
854 | sas_port_delete_phy(port_details->port, phy_info->phy); | ||
855 | } | ||
855 | phy_info->port_details = NULL; | 856 | phy_info->port_details = NULL; |
856 | } | 857 | } |
857 | 858 | ||
@@ -1272,7 +1273,6 @@ mptsas_ioc_reset(MPT_ADAPTER *ioc, int reset_phase) | |||
1272 | } | 1273 | } |
1273 | mptsas_cleanup_fw_event_q(ioc); | 1274 | mptsas_cleanup_fw_event_q(ioc); |
1274 | mptsas_queue_rescan(ioc); | 1275 | mptsas_queue_rescan(ioc); |
1275 | mptsas_fw_event_on(ioc); | ||
1276 | break; | 1276 | break; |
1277 | default: | 1277 | default: |
1278 | break; | 1278 | break; |
@@ -1318,7 +1318,7 @@ mptsas_sas_enclosure_pg0(MPT_ADAPTER *ioc, struct mptsas_enclosure *enclosure, | |||
1318 | cfg.pageAddr = form + form_specific; | 1318 | cfg.pageAddr = form + form_specific; |
1319 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; | 1319 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; |
1320 | cfg.dir = 0; /* read */ | 1320 | cfg.dir = 0; /* read */ |
1321 | cfg.timeout = 10; | 1321 | cfg.timeout = SAS_CONFIG_PAGE_TIMEOUT; |
1322 | 1322 | ||
1323 | error = mpt_config(ioc, &cfg); | 1323 | error = mpt_config(ioc, &cfg); |
1324 | if (error) | 1324 | if (error) |
@@ -1592,6 +1592,7 @@ mptsas_firmware_event_work(struct work_struct *work) | |||
1592 | mptsas_scan_sas_topology(ioc); | 1592 | mptsas_scan_sas_topology(ioc); |
1593 | ioc->in_rescan = 0; | 1593 | ioc->in_rescan = 0; |
1594 | mptsas_free_fw_event(ioc, fw_event); | 1594 | mptsas_free_fw_event(ioc, fw_event); |
1595 | mptsas_fw_event_on(ioc); | ||
1595 | return; | 1596 | return; |
1596 | } | 1597 | } |
1597 | 1598 | ||
@@ -1891,7 +1892,7 @@ static struct scsi_host_template mptsas_driver_template = { | |||
1891 | .eh_bus_reset_handler = mptscsih_bus_reset, | 1892 | .eh_bus_reset_handler = mptscsih_bus_reset, |
1892 | .eh_host_reset_handler = mptscsih_host_reset, | 1893 | .eh_host_reset_handler = mptscsih_host_reset, |
1893 | .bios_param = mptscsih_bios_param, | 1894 | .bios_param = mptscsih_bios_param, |
1894 | .can_queue = MPT_FC_CAN_QUEUE, | 1895 | .can_queue = MPT_SAS_CAN_QUEUE, |
1895 | .this_id = -1, | 1896 | .this_id = -1, |
1896 | .sg_tablesize = MPT_SCSI_SG_DEPTH, | 1897 | .sg_tablesize = MPT_SCSI_SG_DEPTH, |
1897 | .max_sectors = 8192, | 1898 | .max_sectors = 8192, |
@@ -1926,7 +1927,7 @@ static int mptsas_get_linkerrors(struct sas_phy *phy) | |||
1926 | cfg.pageAddr = phy->identify.phy_identifier; | 1927 | cfg.pageAddr = phy->identify.phy_identifier; |
1927 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; | 1928 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; |
1928 | cfg.dir = 0; /* read */ | 1929 | cfg.dir = 0; /* read */ |
1929 | cfg.timeout = 10; | 1930 | cfg.timeout = SAS_CONFIG_PAGE_TIMEOUT; |
1930 | 1931 | ||
1931 | error = mpt_config(ioc, &cfg); | 1932 | error = mpt_config(ioc, &cfg); |
1932 | if (error) | 1933 | if (error) |
@@ -2278,7 +2279,7 @@ mptsas_sas_io_unit_pg0(MPT_ADAPTER *ioc, struct mptsas_portinfo *port_info) | |||
2278 | cfg.pageAddr = 0; | 2279 | cfg.pageAddr = 0; |
2279 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; | 2280 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; |
2280 | cfg.dir = 0; /* read */ | 2281 | cfg.dir = 0; /* read */ |
2281 | cfg.timeout = 10; | 2282 | cfg.timeout = SAS_CONFIG_PAGE_TIMEOUT; |
2282 | 2283 | ||
2283 | error = mpt_config(ioc, &cfg); | 2284 | error = mpt_config(ioc, &cfg); |
2284 | if (error) | 2285 | if (error) |
@@ -2349,7 +2350,7 @@ mptsas_sas_io_unit_pg1(MPT_ADAPTER *ioc) | |||
2349 | 2350 | ||
2350 | cfg.cfghdr.ehdr = &hdr; | 2351 | cfg.cfghdr.ehdr = &hdr; |
2351 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; | 2352 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; |
2352 | cfg.timeout = 10; | 2353 | cfg.timeout = SAS_CONFIG_PAGE_TIMEOUT; |
2353 | cfg.cfghdr.ehdr->PageType = MPI_CONFIG_PAGETYPE_EXTENDED; | 2354 | cfg.cfghdr.ehdr->PageType = MPI_CONFIG_PAGETYPE_EXTENDED; |
2354 | cfg.cfghdr.ehdr->ExtPageType = MPI_CONFIG_EXTPAGETYPE_SAS_IO_UNIT; | 2355 | cfg.cfghdr.ehdr->ExtPageType = MPI_CONFIG_EXTPAGETYPE_SAS_IO_UNIT; |
2355 | cfg.cfghdr.ehdr->PageVersion = MPI_SASIOUNITPAGE1_PAGEVERSION; | 2356 | cfg.cfghdr.ehdr->PageVersion = MPI_SASIOUNITPAGE1_PAGEVERSION; |
@@ -2411,7 +2412,7 @@ mptsas_sas_phy_pg0(MPT_ADAPTER *ioc, struct mptsas_phyinfo *phy_info, | |||
2411 | 2412 | ||
2412 | cfg.cfghdr.ehdr = &hdr; | 2413 | cfg.cfghdr.ehdr = &hdr; |
2413 | cfg.dir = 0; /* read */ | 2414 | cfg.dir = 0; /* read */ |
2414 | cfg.timeout = 10; | 2415 | cfg.timeout = SAS_CONFIG_PAGE_TIMEOUT; |
2415 | 2416 | ||
2416 | /* Get Phy Pg 0 for each Phy. */ | 2417 | /* Get Phy Pg 0 for each Phy. */ |
2417 | cfg.physAddr = -1; | 2418 | cfg.physAddr = -1; |
@@ -2479,7 +2480,7 @@ mptsas_sas_device_pg0(MPT_ADAPTER *ioc, struct mptsas_devinfo *device_info, | |||
2479 | cfg.physAddr = -1; | 2480 | cfg.physAddr = -1; |
2480 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; | 2481 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; |
2481 | cfg.dir = 0; /* read */ | 2482 | cfg.dir = 0; /* read */ |
2482 | cfg.timeout = 10; | 2483 | cfg.timeout = SAS_CONFIG_PAGE_TIMEOUT; |
2483 | 2484 | ||
2484 | memset(device_info, 0, sizeof(struct mptsas_devinfo)); | 2485 | memset(device_info, 0, sizeof(struct mptsas_devinfo)); |
2485 | error = mpt_config(ioc, &cfg); | 2486 | error = mpt_config(ioc, &cfg); |
@@ -2554,7 +2555,7 @@ mptsas_sas_expander_pg0(MPT_ADAPTER *ioc, struct mptsas_portinfo *port_info, | |||
2554 | cfg.pageAddr = form + form_specific; | 2555 | cfg.pageAddr = form + form_specific; |
2555 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; | 2556 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; |
2556 | cfg.dir = 0; /* read */ | 2557 | cfg.dir = 0; /* read */ |
2557 | cfg.timeout = 10; | 2558 | cfg.timeout = SAS_CONFIG_PAGE_TIMEOUT; |
2558 | 2559 | ||
2559 | memset(port_info, 0, sizeof(struct mptsas_portinfo)); | 2560 | memset(port_info, 0, sizeof(struct mptsas_portinfo)); |
2560 | error = mpt_config(ioc, &cfg); | 2561 | error = mpt_config(ioc, &cfg); |
@@ -2635,7 +2636,7 @@ mptsas_sas_expander_pg1(MPT_ADAPTER *ioc, struct mptsas_phyinfo *phy_info, | |||
2635 | cfg.pageAddr = form + form_specific; | 2636 | cfg.pageAddr = form + form_specific; |
2636 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; | 2637 | cfg.action = MPI_CONFIG_ACTION_PAGE_HEADER; |
2637 | cfg.dir = 0; /* read */ | 2638 | cfg.dir = 0; /* read */ |
2638 | cfg.timeout = 10; | 2639 | cfg.timeout = SAS_CONFIG_PAGE_TIMEOUT; |
2639 | 2640 | ||
2640 | error = mpt_config(ioc, &cfg); | 2641 | error = mpt_config(ioc, &cfg); |
2641 | if (error) | 2642 | if (error) |
@@ -3307,6 +3308,7 @@ mptsas_send_expander_event(struct fw_event_work *fw_event) | |||
3307 | expander_data = (MpiEventDataSasExpanderStatusChange_t *) | 3308 | expander_data = (MpiEventDataSasExpanderStatusChange_t *) |
3308 | fw_event->event_data; | 3309 | fw_event->event_data; |
3309 | memcpy(&sas_address, &expander_data->SASAddress, sizeof(__le64)); | 3310 | memcpy(&sas_address, &expander_data->SASAddress, sizeof(__le64)); |
3311 | sas_address = le64_to_cpu(sas_address); | ||
3310 | port_info = mptsas_find_portinfo_by_sas_address(ioc, sas_address); | 3312 | port_info = mptsas_find_portinfo_by_sas_address(ioc, sas_address); |
3311 | 3313 | ||
3312 | if (expander_data->ReasonCode == MPI_EVENT_SAS_EXP_RC_ADDED) { | 3314 | if (expander_data->ReasonCode == MPI_EVENT_SAS_EXP_RC_ADDED) { |
@@ -4760,10 +4762,9 @@ mptsas_probe(struct pci_dev *pdev, const struct pci_device_id *id) | |||
4760 | 4762 | ||
4761 | /* set 16 byte cdb's */ | 4763 | /* set 16 byte cdb's */ |
4762 | sh->max_cmd_len = 16; | 4764 | sh->max_cmd_len = 16; |
4763 | 4765 | sh->can_queue = min_t(int, ioc->req_depth - 10, sh->can_queue); | |
4764 | sh->max_id = ioc->pfacts[0].PortSCSIID; | 4766 | sh->max_id = -1; |
4765 | sh->max_lun = max_lun; | 4767 | sh->max_lun = max_lun; |
4766 | |||
4767 | sh->transportt = mptsas_transport_template; | 4768 | sh->transportt = mptsas_transport_template; |
4768 | 4769 | ||
4769 | /* Required entry. | 4770 | /* Required entry. |
@@ -4821,25 +4822,6 @@ mptsas_probe(struct pci_dev *pdev, const struct pci_device_id *id) | |||
4821 | dprintk(ioc, printk(MYIOC_s_DEBUG_FMT "ScsiLookup @ %p\n", | 4822 | dprintk(ioc, printk(MYIOC_s_DEBUG_FMT "ScsiLookup @ %p\n", |
4822 | ioc->name, ioc->ScsiLookup)); | 4823 | ioc->name, ioc->ScsiLookup)); |
4823 | 4824 | ||
4824 | /* Clear the TM flags | ||
4825 | */ | ||
4826 | hd->abortSCpnt = NULL; | ||
4827 | |||
4828 | /* Clear the pointer used to store | ||
4829 | * single-threaded commands, i.e., those | ||
4830 | * issued during a bus scan, dv and | ||
4831 | * configuration pages. | ||
4832 | */ | ||
4833 | hd->cmdPtr = NULL; | ||
4834 | |||
4835 | /* Initialize this SCSI Hosts' timers | ||
4836 | * To use, set the timer expires field | ||
4837 | * and add_timer | ||
4838 | */ | ||
4839 | init_timer(&hd->timer); | ||
4840 | hd->timer.data = (unsigned long) hd; | ||
4841 | hd->timer.function = mptscsih_timer_expired; | ||
4842 | |||
4843 | ioc->sas_data.ptClear = mpt_pt_clear; | 4825 | ioc->sas_data.ptClear = mpt_pt_clear; |
4844 | 4826 | ||
4845 | hd->last_queue_full = 0; | 4827 | hd->last_queue_full = 0; |
diff --git a/drivers/message/fusion/mptscsih.c b/drivers/message/fusion/mptscsih.c index 8440f78f6969..c29578614504 100644 --- a/drivers/message/fusion/mptscsih.c +++ b/drivers/message/fusion/mptscsih.c | |||
@@ -628,6 +628,16 @@ mptscsih_io_done(MPT_ADAPTER *ioc, MPT_FRAME_HDR *mf, MPT_FRAME_HDR *mr) | |||
628 | return 1; | 628 | return 1; |
629 | } | 629 | } |
630 | 630 | ||
631 | if (ioc->bus_type == SAS) { | ||
632 | VirtDevice *vdevice = sc->device->hostdata; | ||
633 | |||
634 | if (!vdevice || !vdevice->vtarget || | ||
635 | vdevice->vtarget->deleted) { | ||
636 | sc->result = DID_NO_CONNECT << 16; | ||
637 | goto out; | ||
638 | } | ||
639 | } | ||
640 | |||
631 | sc->host_scribble = NULL; | 641 | sc->host_scribble = NULL; |
632 | sc->result = DID_OK << 16; /* Set default reply as OK */ | 642 | sc->result = DID_OK << 16; /* Set default reply as OK */ |
633 | pScsiReq = (SCSIIORequest_t *) mf; | 643 | pScsiReq = (SCSIIORequest_t *) mf; |
@@ -689,6 +699,7 @@ mptscsih_io_done(MPT_ADAPTER *ioc, MPT_FRAME_HDR *mf, MPT_FRAME_HDR *mr) | |||
689 | 699 | ||
690 | switch(status) { | 700 | switch(status) { |
691 | case MPI_IOCSTATUS_BUSY: /* 0x0002 */ | 701 | case MPI_IOCSTATUS_BUSY: /* 0x0002 */ |
702 | case MPI_IOCSTATUS_INSUFFICIENT_RESOURCES: /* 0x0006 */ | ||
692 | /* CHECKME! | 703 | /* CHECKME! |
693 | * Maybe: DRIVER_BUSY | SUGGEST_RETRY | DID_SOFT_ERROR (retry) | 704 | * Maybe: DRIVER_BUSY | SUGGEST_RETRY | DID_SOFT_ERROR (retry) |
694 | * But not: DID_BUS_BUSY lest one risk | 705 | * But not: DID_BUS_BUSY lest one risk |
@@ -872,7 +883,6 @@ mptscsih_io_done(MPT_ADAPTER *ioc, MPT_FRAME_HDR *mf, MPT_FRAME_HDR *mr) | |||
872 | case MPI_IOCSTATUS_INVALID_SGL: /* 0x0003 */ | 883 | case MPI_IOCSTATUS_INVALID_SGL: /* 0x0003 */ |
873 | case MPI_IOCSTATUS_INTERNAL_ERROR: /* 0x0004 */ | 884 | case MPI_IOCSTATUS_INTERNAL_ERROR: /* 0x0004 */ |
874 | case MPI_IOCSTATUS_RESERVED: /* 0x0005 */ | 885 | case MPI_IOCSTATUS_RESERVED: /* 0x0005 */ |
875 | case MPI_IOCSTATUS_INSUFFICIENT_RESOURCES: /* 0x0006 */ | ||
876 | case MPI_IOCSTATUS_INVALID_FIELD: /* 0x0007 */ | 886 | case MPI_IOCSTATUS_INVALID_FIELD: /* 0x0007 */ |
877 | case MPI_IOCSTATUS_INVALID_STATE: /* 0x0008 */ | 887 | case MPI_IOCSTATUS_INVALID_STATE: /* 0x0008 */ |
878 | case MPI_IOCSTATUS_SCSI_IO_DATA_ERROR: /* 0x0046 */ | 888 | case MPI_IOCSTATUS_SCSI_IO_DATA_ERROR: /* 0x0046 */ |
@@ -892,7 +902,7 @@ mptscsih_io_done(MPT_ADAPTER *ioc, MPT_FRAME_HDR *mf, MPT_FRAME_HDR *mr) | |||
892 | #endif | 902 | #endif |
893 | 903 | ||
894 | } /* end of address reply case */ | 904 | } /* end of address reply case */ |
895 | 905 | out: | |
896 | /* Unmap the DMA buffers, if any. */ | 906 | /* Unmap the DMA buffers, if any. */ |
897 | scsi_dma_unmap(sc); | 907 | scsi_dma_unmap(sc); |
898 | 908 | ||
@@ -1729,9 +1739,6 @@ mptscsih_abort(struct scsi_cmnd * SCpnt) | |||
1729 | */ | 1739 | */ |
1730 | mf = MPT_INDEX_2_MFPTR(ioc, scpnt_idx); | 1740 | mf = MPT_INDEX_2_MFPTR(ioc, scpnt_idx); |
1731 | ctx2abort = mf->u.frame.hwhdr.msgctxu.MsgContext; | 1741 | ctx2abort = mf->u.frame.hwhdr.msgctxu.MsgContext; |
1732 | |||
1733 | hd->abortSCpnt = SCpnt; | ||
1734 | |||
1735 | retval = mptscsih_IssueTaskMgmt(hd, | 1742 | retval = mptscsih_IssueTaskMgmt(hd, |
1736 | MPI_SCSITASKMGMT_TASKTYPE_ABORT_TASK, | 1743 | MPI_SCSITASKMGMT_TASKTYPE_ABORT_TASK, |
1737 | vdevice->vtarget->channel, | 1744 | vdevice->vtarget->channel, |
@@ -2293,7 +2300,10 @@ mptscsih_change_queue_depth(struct scsi_device *sdev, int qdepth) | |||
2293 | else | 2300 | else |
2294 | max_depth = MPT_SCSI_CMD_PER_DEV_LOW; | 2301 | max_depth = MPT_SCSI_CMD_PER_DEV_LOW; |
2295 | } else | 2302 | } else |
2296 | max_depth = MPT_SCSI_CMD_PER_DEV_HIGH; | 2303 | max_depth = ioc->sh->can_queue; |
2304 | |||
2305 | if (!sdev->tagged_supported) | ||
2306 | max_depth = 1; | ||
2297 | 2307 | ||
2298 | if (qdepth > max_depth) | 2308 | if (qdepth > max_depth) |
2299 | qdepth = max_depth; | 2309 | qdepth = max_depth; |
@@ -2627,50 +2637,6 @@ mptscsih_scandv_complete(MPT_ADAPTER *ioc, MPT_FRAME_HDR *req, | |||
2627 | return 1; | 2637 | return 1; |
2628 | } | 2638 | } |
2629 | 2639 | ||
2630 | /*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/ | ||
2631 | /* mptscsih_timer_expired - Call back for timer process. | ||
2632 | * Used only for dv functionality. | ||
2633 | * @data: Pointer to MPT_SCSI_HOST recast as an unsigned long | ||
2634 | * | ||
2635 | */ | ||
2636 | void | ||
2637 | mptscsih_timer_expired(unsigned long data) | ||
2638 | { | ||
2639 | MPT_SCSI_HOST *hd = (MPT_SCSI_HOST *) data; | ||
2640 | MPT_ADAPTER *ioc = hd->ioc; | ||
2641 | |||
2642 | ddvprintk(ioc, printk(MYIOC_s_DEBUG_FMT "Timer Expired! Cmd %p\n", ioc->name, hd->cmdPtr)); | ||
2643 | |||
2644 | if (hd->cmdPtr) { | ||
2645 | MPIHeader_t *cmd = (MPIHeader_t *)hd->cmdPtr; | ||
2646 | |||
2647 | if (cmd->Function == MPI_FUNCTION_SCSI_IO_REQUEST) { | ||
2648 | /* Desire to issue a task management request here. | ||
2649 | * TM requests MUST be single threaded. | ||
2650 | * If old eh code and no TM current, issue request. | ||
2651 | * If new eh code, do nothing. Wait for OS cmd timeout | ||
2652 | * for bus reset. | ||
2653 | */ | ||
2654 | } else { | ||
2655 | /* Perform a FW reload */ | ||
2656 | if (mpt_HardResetHandler(ioc, NO_SLEEP) < 0) { | ||
2657 | printk(MYIOC_s_WARN_FMT "Firmware Reload FAILED!\n", ioc->name); | ||
2658 | } | ||
2659 | } | ||
2660 | } else { | ||
2661 | /* This should NEVER happen */ | ||
2662 | printk(MYIOC_s_WARN_FMT "Null cmdPtr!!!!\n", ioc->name); | ||
2663 | } | ||
2664 | |||
2665 | /* No more processing. | ||
2666 | * TM call will generate an interrupt for SCSI TM Management. | ||
2667 | * The FW will reply to all outstanding commands, callback will finish cleanup. | ||
2668 | * Hard reset clean-up will free all resources. | ||
2669 | */ | ||
2670 | ddvprintk(ioc, printk(MYIOC_s_DEBUG_FMT "Timer Expired Complete!\n", ioc->name)); | ||
2671 | |||
2672 | return; | ||
2673 | } | ||
2674 | 2640 | ||
2675 | /** | 2641 | /** |
2676 | * mptscsih_get_completion_code - | 2642 | * mptscsih_get_completion_code - |
@@ -3265,6 +3231,5 @@ EXPORT_SYMBOL(mptscsih_scandv_complete); | |||
3265 | EXPORT_SYMBOL(mptscsih_event_process); | 3231 | EXPORT_SYMBOL(mptscsih_event_process); |
3266 | EXPORT_SYMBOL(mptscsih_ioc_reset); | 3232 | EXPORT_SYMBOL(mptscsih_ioc_reset); |
3267 | EXPORT_SYMBOL(mptscsih_change_queue_depth); | 3233 | EXPORT_SYMBOL(mptscsih_change_queue_depth); |
3268 | EXPORT_SYMBOL(mptscsih_timer_expired); | ||
3269 | 3234 | ||
3270 | /*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/ | 3235 | /*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/ |
diff --git a/drivers/message/fusion/mptscsih.h b/drivers/message/fusion/mptscsih.h index eb3f677528ac..e0b33e04a33b 100644 --- a/drivers/message/fusion/mptscsih.h +++ b/drivers/message/fusion/mptscsih.h | |||
@@ -129,7 +129,6 @@ extern int mptscsih_scandv_complete(MPT_ADAPTER *ioc, MPT_FRAME_HDR *mf, MPT_FRA | |||
129 | extern int mptscsih_event_process(MPT_ADAPTER *ioc, EventNotificationReply_t *pEvReply); | 129 | extern int mptscsih_event_process(MPT_ADAPTER *ioc, EventNotificationReply_t *pEvReply); |
130 | extern int mptscsih_ioc_reset(MPT_ADAPTER *ioc, int post_reset); | 130 | extern int mptscsih_ioc_reset(MPT_ADAPTER *ioc, int post_reset); |
131 | extern int mptscsih_change_queue_depth(struct scsi_device *sdev, int qdepth); | 131 | extern int mptscsih_change_queue_depth(struct scsi_device *sdev, int qdepth); |
132 | extern void mptscsih_timer_expired(unsigned long data); | ||
133 | extern u8 mptscsih_raid_id_to_num(MPT_ADAPTER *ioc, u8 channel, u8 id); | 132 | extern u8 mptscsih_raid_id_to_num(MPT_ADAPTER *ioc, u8 channel, u8 id); |
134 | extern int mptscsih_is_phys_disk(MPT_ADAPTER *ioc, u8 channel, u8 id); | 133 | extern int mptscsih_is_phys_disk(MPT_ADAPTER *ioc, u8 channel, u8 id); |
135 | extern struct device_attribute *mptscsih_host_attrs[]; | 134 | extern struct device_attribute *mptscsih_host_attrs[]; |
diff --git a/drivers/message/fusion/mptspi.c b/drivers/message/fusion/mptspi.c index c5b808fd55ba..69f4257419b5 100644 --- a/drivers/message/fusion/mptspi.c +++ b/drivers/message/fusion/mptspi.c | |||
@@ -1472,28 +1472,7 @@ mptspi_probe(struct pci_dev *pdev, const struct pci_device_id *id) | |||
1472 | dprintk(ioc, printk(MYIOC_s_DEBUG_FMT "ScsiLookup @ %p\n", | 1472 | dprintk(ioc, printk(MYIOC_s_DEBUG_FMT "ScsiLookup @ %p\n", |
1473 | ioc->name, ioc->ScsiLookup)); | 1473 | ioc->name, ioc->ScsiLookup)); |
1474 | 1474 | ||
1475 | /* Clear the TM flags | ||
1476 | */ | ||
1477 | hd->abortSCpnt = NULL; | ||
1478 | |||
1479 | /* Clear the pointer used to store | ||
1480 | * single-threaded commands, i.e., those | ||
1481 | * issued during a bus scan, dv and | ||
1482 | * configuration pages. | ||
1483 | */ | ||
1484 | hd->cmdPtr = NULL; | ||
1485 | |||
1486 | /* Initialize this SCSI Hosts' timers | ||
1487 | * To use, set the timer expires field | ||
1488 | * and add_timer | ||
1489 | */ | ||
1490 | init_timer(&hd->timer); | ||
1491 | hd->timer.data = (unsigned long) hd; | ||
1492 | hd->timer.function = mptscsih_timer_expired; | ||
1493 | |||
1494 | ioc->spi_data.Saf_Te = mpt_saf_te; | 1475 | ioc->spi_data.Saf_Te = mpt_saf_te; |
1495 | |||
1496 | hd->negoNvram = MPT_SCSICFG_USE_NVRAM; | ||
1497 | ddvprintk(ioc, printk(MYIOC_s_DEBUG_FMT | 1476 | ddvprintk(ioc, printk(MYIOC_s_DEBUG_FMT |
1498 | "saf_te %x\n", | 1477 | "saf_te %x\n", |
1499 | ioc->name, | 1478 | ioc->name, |
diff --git a/drivers/misc/enclosure.c b/drivers/misc/enclosure.c index 348443bdb23b..7b039306037f 100644 --- a/drivers/misc/enclosure.c +++ b/drivers/misc/enclosure.c | |||
@@ -33,24 +33,44 @@ static DEFINE_MUTEX(container_list_lock); | |||
33 | static struct class enclosure_class; | 33 | static struct class enclosure_class; |
34 | 34 | ||
35 | /** | 35 | /** |
36 | * enclosure_find - find an enclosure given a device | 36 | * enclosure_find - find an enclosure given a parent device |
37 | * @dev: the device to find for | 37 | * @dev: the parent to match against |
38 | * @start: Optional enclosure device to start from (NULL if none) | ||
38 | * | 39 | * |
39 | * Looks through the list of registered enclosures to see | 40 | * Looks through the list of registered enclosures to find all those |
40 | * if it can find a match for a device. Returns NULL if no | 41 | * with @dev as a parent. Returns NULL if no enclosure is |
41 | * enclosure is found. Obtains a reference to the enclosure class | 42 | * found. @start can be used as a starting point to obtain multiple |
42 | * device which must be released with device_put(). | 43 | * enclosures per parent (should begin with NULL and then be set to |
44 | * each returned enclosure device). Obtains a reference to the | ||
45 | * enclosure class device which must be released with device_put(). | ||
46 | * If @start is not NULL, a reference must be taken on it which is | ||
47 | * released before returning (this allows a loop through all | ||
48 | * enclosures to exit with only the reference on the enclosure of | ||
49 | * interest held). Note that the @dev may correspond to the actual | ||
50 | * device housing the enclosure, in which case no iteration via @start | ||
51 | * is required. | ||
43 | */ | 52 | */ |
44 | struct enclosure_device *enclosure_find(struct device *dev) | 53 | struct enclosure_device *enclosure_find(struct device *dev, |
54 | struct enclosure_device *start) | ||
45 | { | 55 | { |
46 | struct enclosure_device *edev; | 56 | struct enclosure_device *edev; |
47 | 57 | ||
48 | mutex_lock(&container_list_lock); | 58 | mutex_lock(&container_list_lock); |
49 | list_for_each_entry(edev, &container_list, node) { | 59 | edev = list_prepare_entry(start, &container_list, node); |
50 | if (edev->edev.parent == dev) { | 60 | if (start) |
51 | get_device(&edev->edev); | 61 | put_device(&start->edev); |
52 | mutex_unlock(&container_list_lock); | 62 | |
53 | return edev; | 63 | list_for_each_entry_continue(edev, &container_list, node) { |
64 | struct device *parent = edev->edev.parent; | ||
65 | /* parent might not be immediate, so iterate up to | ||
66 | * the root of the tree if necessary */ | ||
67 | while (parent) { | ||
68 | if (parent == dev) { | ||
69 | get_device(&edev->edev); | ||
70 | mutex_unlock(&container_list_lock); | ||
71 | return edev; | ||
72 | } | ||
73 | parent = parent->parent; | ||
54 | } | 74 | } |
55 | } | 75 | } |
56 | mutex_unlock(&container_list_lock); | 76 | mutex_unlock(&container_list_lock); |
@@ -295,6 +315,9 @@ int enclosure_add_device(struct enclosure_device *edev, int component, | |||
295 | 315 | ||
296 | cdev = &edev->component[component]; | 316 | cdev = &edev->component[component]; |
297 | 317 | ||
318 | if (cdev->dev == dev) | ||
319 | return -EEXIST; | ||
320 | |||
298 | if (cdev->dev) | 321 | if (cdev->dev) |
299 | enclosure_remove_links(cdev); | 322 | enclosure_remove_links(cdev); |
300 | 323 | ||
@@ -312,19 +335,25 @@ EXPORT_SYMBOL_GPL(enclosure_add_device); | |||
312 | * Returns zero on success or an error. | 335 | * Returns zero on success or an error. |
313 | * | 336 | * |
314 | */ | 337 | */ |
315 | int enclosure_remove_device(struct enclosure_device *edev, int component) | 338 | int enclosure_remove_device(struct enclosure_device *edev, struct device *dev) |
316 | { | 339 | { |
317 | struct enclosure_component *cdev; | 340 | struct enclosure_component *cdev; |
341 | int i; | ||
318 | 342 | ||
319 | if (!edev || component >= edev->components) | 343 | if (!edev || !dev) |
320 | return -EINVAL; | 344 | return -EINVAL; |
321 | 345 | ||
322 | cdev = &edev->component[component]; | 346 | for (i = 0; i < edev->components; i++) { |
323 | 347 | cdev = &edev->component[i]; | |
324 | device_del(&cdev->cdev); | 348 | if (cdev->dev == dev) { |
325 | put_device(cdev->dev); | 349 | enclosure_remove_links(cdev); |
326 | cdev->dev = NULL; | 350 | device_del(&cdev->cdev); |
327 | return device_add(&cdev->cdev); | 351 | put_device(dev); |
352 | cdev->dev = NULL; | ||
353 | return device_add(&cdev->cdev); | ||
354 | } | ||
355 | } | ||
356 | return -ENODEV; | ||
328 | } | 357 | } |
329 | EXPORT_SYMBOL_GPL(enclosure_remove_device); | 358 | EXPORT_SYMBOL_GPL(enclosure_remove_device); |
330 | 359 | ||
diff --git a/drivers/s390/scsi/zfcp_aux.c b/drivers/s390/scsi/zfcp_aux.c index 2ccbd185a5fb..1be6bf7e8ce6 100644 --- a/drivers/s390/scsi/zfcp_aux.c +++ b/drivers/s390/scsi/zfcp_aux.c | |||
@@ -42,6 +42,12 @@ static char *init_device; | |||
42 | module_param_named(device, init_device, charp, 0400); | 42 | module_param_named(device, init_device, charp, 0400); |
43 | MODULE_PARM_DESC(device, "specify initial device"); | 43 | MODULE_PARM_DESC(device, "specify initial device"); |
44 | 44 | ||
45 | static struct kmem_cache *zfcp_cache_hw_align(const char *name, | ||
46 | unsigned long size) | ||
47 | { | ||
48 | return kmem_cache_create(name, size, roundup_pow_of_two(size), 0, NULL); | ||
49 | } | ||
50 | |||
45 | static int zfcp_reqlist_alloc(struct zfcp_adapter *adapter) | 51 | static int zfcp_reqlist_alloc(struct zfcp_adapter *adapter) |
46 | { | 52 | { |
47 | int idx; | 53 | int idx; |
@@ -78,7 +84,7 @@ static void __init zfcp_init_device_configure(char *busid, u64 wwpn, u64 lun) | |||
78 | struct zfcp_port *port; | 84 | struct zfcp_port *port; |
79 | struct zfcp_unit *unit; | 85 | struct zfcp_unit *unit; |
80 | 86 | ||
81 | down(&zfcp_data.config_sema); | 87 | mutex_lock(&zfcp_data.config_mutex); |
82 | read_lock_irq(&zfcp_data.config_lock); | 88 | read_lock_irq(&zfcp_data.config_lock); |
83 | adapter = zfcp_get_adapter_by_busid(busid); | 89 | adapter = zfcp_get_adapter_by_busid(busid); |
84 | if (adapter) | 90 | if (adapter) |
@@ -93,31 +99,23 @@ static void __init zfcp_init_device_configure(char *busid, u64 wwpn, u64 lun) | |||
93 | unit = zfcp_unit_enqueue(port, lun); | 99 | unit = zfcp_unit_enqueue(port, lun); |
94 | if (IS_ERR(unit)) | 100 | if (IS_ERR(unit)) |
95 | goto out_unit; | 101 | goto out_unit; |
96 | up(&zfcp_data.config_sema); | 102 | mutex_unlock(&zfcp_data.config_mutex); |
97 | ccw_device_set_online(adapter->ccw_device); | 103 | ccw_device_set_online(adapter->ccw_device); |
98 | 104 | ||
99 | zfcp_erp_wait(adapter); | 105 | zfcp_erp_wait(adapter); |
100 | flush_work(&unit->scsi_work); | 106 | flush_work(&unit->scsi_work); |
101 | 107 | ||
102 | down(&zfcp_data.config_sema); | 108 | mutex_lock(&zfcp_data.config_mutex); |
103 | zfcp_unit_put(unit); | 109 | zfcp_unit_put(unit); |
104 | out_unit: | 110 | out_unit: |
105 | zfcp_port_put(port); | 111 | zfcp_port_put(port); |
106 | out_port: | 112 | out_port: |
107 | zfcp_adapter_put(adapter); | 113 | zfcp_adapter_put(adapter); |
108 | out_adapter: | 114 | out_adapter: |
109 | up(&zfcp_data.config_sema); | 115 | mutex_unlock(&zfcp_data.config_mutex); |
110 | return; | 116 | return; |
111 | } | 117 | } |
112 | 118 | ||
113 | static struct kmem_cache *zfcp_cache_create(int size, char *name) | ||
114 | { | ||
115 | int align = 1; | ||
116 | while ((size - align) > 0) | ||
117 | align <<= 1; | ||
118 | return kmem_cache_create(name , size, align, 0, NULL); | ||
119 | } | ||
120 | |||
121 | static void __init zfcp_init_device_setup(char *devstr) | 119 | static void __init zfcp_init_device_setup(char *devstr) |
122 | { | 120 | { |
123 | char *token; | 121 | char *token; |
@@ -158,24 +156,27 @@ static int __init zfcp_module_init(void) | |||
158 | { | 156 | { |
159 | int retval = -ENOMEM; | 157 | int retval = -ENOMEM; |
160 | 158 | ||
161 | zfcp_data.fsf_req_qtcb_cache = zfcp_cache_create( | 159 | zfcp_data.gpn_ft_cache = zfcp_cache_hw_align("zfcp_gpn", |
162 | sizeof(struct zfcp_fsf_req_qtcb), "zfcp_fsf"); | 160 | sizeof(struct ct_iu_gpn_ft_req)); |
163 | if (!zfcp_data.fsf_req_qtcb_cache) | 161 | if (!zfcp_data.gpn_ft_cache) |
164 | goto out; | 162 | goto out; |
165 | 163 | ||
166 | zfcp_data.sr_buffer_cache = zfcp_cache_create( | 164 | zfcp_data.qtcb_cache = zfcp_cache_hw_align("zfcp_qtcb", |
167 | sizeof(struct fsf_status_read_buffer), "zfcp_sr"); | 165 | sizeof(struct fsf_qtcb)); |
166 | if (!zfcp_data.qtcb_cache) | ||
167 | goto out_qtcb_cache; | ||
168 | |||
169 | zfcp_data.sr_buffer_cache = zfcp_cache_hw_align("zfcp_sr", | ||
170 | sizeof(struct fsf_status_read_buffer)); | ||
168 | if (!zfcp_data.sr_buffer_cache) | 171 | if (!zfcp_data.sr_buffer_cache) |
169 | goto out_sr_cache; | 172 | goto out_sr_cache; |
170 | 173 | ||
171 | zfcp_data.gid_pn_cache = zfcp_cache_create( | 174 | zfcp_data.gid_pn_cache = zfcp_cache_hw_align("zfcp_gid", |
172 | sizeof(struct zfcp_gid_pn_data), "zfcp_gid"); | 175 | sizeof(struct zfcp_gid_pn_data)); |
173 | if (!zfcp_data.gid_pn_cache) | 176 | if (!zfcp_data.gid_pn_cache) |
174 | goto out_gid_cache; | 177 | goto out_gid_cache; |
175 | 178 | ||
176 | zfcp_data.work_queue = create_singlethread_workqueue("zfcp_wq"); | 179 | mutex_init(&zfcp_data.config_mutex); |
177 | |||
178 | sema_init(&zfcp_data.config_sema, 1); | ||
179 | rwlock_init(&zfcp_data.config_lock); | 180 | rwlock_init(&zfcp_data.config_lock); |
180 | 181 | ||
181 | zfcp_data.scsi_transport_template = | 182 | zfcp_data.scsi_transport_template = |
@@ -209,7 +210,9 @@ out_transport: | |||
209 | out_gid_cache: | 210 | out_gid_cache: |
210 | kmem_cache_destroy(zfcp_data.sr_buffer_cache); | 211 | kmem_cache_destroy(zfcp_data.sr_buffer_cache); |
211 | out_sr_cache: | 212 | out_sr_cache: |
212 | kmem_cache_destroy(zfcp_data.fsf_req_qtcb_cache); | 213 | kmem_cache_destroy(zfcp_data.qtcb_cache); |
214 | out_qtcb_cache: | ||
215 | kmem_cache_destroy(zfcp_data.gpn_ft_cache); | ||
213 | out: | 216 | out: |
214 | return retval; | 217 | return retval; |
215 | } | 218 | } |
@@ -263,7 +266,7 @@ static void zfcp_sysfs_unit_release(struct device *dev) | |||
263 | * @port: pointer to port where unit is added | 266 | * @port: pointer to port where unit is added |
264 | * @fcp_lun: FCP LUN of unit to be enqueued | 267 | * @fcp_lun: FCP LUN of unit to be enqueued |
265 | * Returns: pointer to enqueued unit on success, ERR_PTR on error | 268 | * Returns: pointer to enqueued unit on success, ERR_PTR on error |
266 | * Locks: config_sema must be held to serialize changes to the unit list | 269 | * Locks: config_mutex must be held to serialize changes to the unit list |
267 | * | 270 | * |
268 | * Sets up some unit internal structures and creates sysfs entry. | 271 | * Sets up some unit internal structures and creates sysfs entry. |
269 | */ | 272 | */ |
@@ -271,6 +274,13 @@ struct zfcp_unit *zfcp_unit_enqueue(struct zfcp_port *port, u64 fcp_lun) | |||
271 | { | 274 | { |
272 | struct zfcp_unit *unit; | 275 | struct zfcp_unit *unit; |
273 | 276 | ||
277 | read_lock_irq(&zfcp_data.config_lock); | ||
278 | if (zfcp_get_unit_by_lun(port, fcp_lun)) { | ||
279 | read_unlock_irq(&zfcp_data.config_lock); | ||
280 | return ERR_PTR(-EINVAL); | ||
281 | } | ||
282 | read_unlock_irq(&zfcp_data.config_lock); | ||
283 | |||
274 | unit = kzalloc(sizeof(struct zfcp_unit), GFP_KERNEL); | 284 | unit = kzalloc(sizeof(struct zfcp_unit), GFP_KERNEL); |
275 | if (!unit) | 285 | if (!unit) |
276 | return ERR_PTR(-ENOMEM); | 286 | return ERR_PTR(-ENOMEM); |
@@ -282,8 +292,11 @@ struct zfcp_unit *zfcp_unit_enqueue(struct zfcp_port *port, u64 fcp_lun) | |||
282 | unit->port = port; | 292 | unit->port = port; |
283 | unit->fcp_lun = fcp_lun; | 293 | unit->fcp_lun = fcp_lun; |
284 | 294 | ||
285 | dev_set_name(&unit->sysfs_device, "0x%016llx", | 295 | if (dev_set_name(&unit->sysfs_device, "0x%016llx", |
286 | (unsigned long long) fcp_lun); | 296 | (unsigned long long) fcp_lun)) { |
297 | kfree(unit); | ||
298 | return ERR_PTR(-ENOMEM); | ||
299 | } | ||
287 | unit->sysfs_device.parent = &port->sysfs_device; | 300 | unit->sysfs_device.parent = &port->sysfs_device; |
288 | unit->sysfs_device.release = zfcp_sysfs_unit_release; | 301 | unit->sysfs_device.release = zfcp_sysfs_unit_release; |
289 | dev_set_drvdata(&unit->sysfs_device, unit); | 302 | dev_set_drvdata(&unit->sysfs_device, unit); |
@@ -299,20 +312,15 @@ struct zfcp_unit *zfcp_unit_enqueue(struct zfcp_port *port, u64 fcp_lun) | |||
299 | unit->latencies.cmd.channel.min = 0xFFFFFFFF; | 312 | unit->latencies.cmd.channel.min = 0xFFFFFFFF; |
300 | unit->latencies.cmd.fabric.min = 0xFFFFFFFF; | 313 | unit->latencies.cmd.fabric.min = 0xFFFFFFFF; |
301 | 314 | ||
302 | read_lock_irq(&zfcp_data.config_lock); | 315 | if (device_register(&unit->sysfs_device)) { |
303 | if (zfcp_get_unit_by_lun(port, fcp_lun)) { | 316 | put_device(&unit->sysfs_device); |
304 | read_unlock_irq(&zfcp_data.config_lock); | 317 | return ERR_PTR(-EINVAL); |
305 | goto err_out_free; | ||
306 | } | 318 | } |
307 | read_unlock_irq(&zfcp_data.config_lock); | ||
308 | |||
309 | if (device_register(&unit->sysfs_device)) | ||
310 | goto err_out_free; | ||
311 | 319 | ||
312 | if (sysfs_create_group(&unit->sysfs_device.kobj, | 320 | if (sysfs_create_group(&unit->sysfs_device.kobj, |
313 | &zfcp_sysfs_unit_attrs)) { | 321 | &zfcp_sysfs_unit_attrs)) { |
314 | device_unregister(&unit->sysfs_device); | 322 | device_unregister(&unit->sysfs_device); |
315 | return ERR_PTR(-EIO); | 323 | return ERR_PTR(-EINVAL); |
316 | } | 324 | } |
317 | 325 | ||
318 | zfcp_unit_get(unit); | 326 | zfcp_unit_get(unit); |
@@ -327,10 +335,6 @@ struct zfcp_unit *zfcp_unit_enqueue(struct zfcp_port *port, u64 fcp_lun) | |||
327 | zfcp_port_get(port); | 335 | zfcp_port_get(port); |
328 | 336 | ||
329 | return unit; | 337 | return unit; |
330 | |||
331 | err_out_free: | ||
332 | kfree(unit); | ||
333 | return ERR_PTR(-EINVAL); | ||
334 | } | 338 | } |
335 | 339 | ||
336 | /** | 340 | /** |
@@ -353,37 +357,47 @@ void zfcp_unit_dequeue(struct zfcp_unit *unit) | |||
353 | 357 | ||
354 | static int zfcp_allocate_low_mem_buffers(struct zfcp_adapter *adapter) | 358 | static int zfcp_allocate_low_mem_buffers(struct zfcp_adapter *adapter) |
355 | { | 359 | { |
356 | /* must only be called with zfcp_data.config_sema taken */ | 360 | /* must only be called with zfcp_data.config_mutex taken */ |
357 | adapter->pool.fsf_req_erp = | 361 | adapter->pool.erp_req = |
358 | mempool_create_slab_pool(1, zfcp_data.fsf_req_qtcb_cache); | 362 | mempool_create_kmalloc_pool(1, sizeof(struct zfcp_fsf_req)); |
359 | if (!adapter->pool.fsf_req_erp) | 363 | if (!adapter->pool.erp_req) |
364 | return -ENOMEM; | ||
365 | |||
366 | adapter->pool.gid_pn_req = | ||
367 | mempool_create_kmalloc_pool(1, sizeof(struct zfcp_fsf_req)); | ||
368 | if (!adapter->pool.gid_pn_req) | ||
360 | return -ENOMEM; | 369 | return -ENOMEM; |
361 | 370 | ||
362 | adapter->pool.fsf_req_scsi = | 371 | adapter->pool.scsi_req = |
363 | mempool_create_slab_pool(1, zfcp_data.fsf_req_qtcb_cache); | 372 | mempool_create_kmalloc_pool(1, sizeof(struct zfcp_fsf_req)); |
364 | if (!adapter->pool.fsf_req_scsi) | 373 | if (!adapter->pool.scsi_req) |
365 | return -ENOMEM; | 374 | return -ENOMEM; |
366 | 375 | ||
367 | adapter->pool.fsf_req_abort = | 376 | adapter->pool.scsi_abort = |
368 | mempool_create_slab_pool(1, zfcp_data.fsf_req_qtcb_cache); | 377 | mempool_create_kmalloc_pool(1, sizeof(struct zfcp_fsf_req)); |
369 | if (!adapter->pool.fsf_req_abort) | 378 | if (!adapter->pool.scsi_abort) |
370 | return -ENOMEM; | 379 | return -ENOMEM; |
371 | 380 | ||
372 | adapter->pool.fsf_req_status_read = | 381 | adapter->pool.status_read_req = |
373 | mempool_create_kmalloc_pool(FSF_STATUS_READS_RECOM, | 382 | mempool_create_kmalloc_pool(FSF_STATUS_READS_RECOM, |
374 | sizeof(struct zfcp_fsf_req)); | 383 | sizeof(struct zfcp_fsf_req)); |
375 | if (!adapter->pool.fsf_req_status_read) | 384 | if (!adapter->pool.status_read_req) |
376 | return -ENOMEM; | 385 | return -ENOMEM; |
377 | 386 | ||
378 | adapter->pool.data_status_read = | 387 | adapter->pool.qtcb_pool = |
388 | mempool_create_slab_pool(4, zfcp_data.qtcb_cache); | ||
389 | if (!adapter->pool.qtcb_pool) | ||
390 | return -ENOMEM; | ||
391 | |||
392 | adapter->pool.status_read_data = | ||
379 | mempool_create_slab_pool(FSF_STATUS_READS_RECOM, | 393 | mempool_create_slab_pool(FSF_STATUS_READS_RECOM, |
380 | zfcp_data.sr_buffer_cache); | 394 | zfcp_data.sr_buffer_cache); |
381 | if (!adapter->pool.data_status_read) | 395 | if (!adapter->pool.status_read_data) |
382 | return -ENOMEM; | 396 | return -ENOMEM; |
383 | 397 | ||
384 | adapter->pool.data_gid_pn = | 398 | adapter->pool.gid_pn_data = |
385 | mempool_create_slab_pool(1, zfcp_data.gid_pn_cache); | 399 | mempool_create_slab_pool(1, zfcp_data.gid_pn_cache); |
386 | if (!adapter->pool.data_gid_pn) | 400 | if (!adapter->pool.gid_pn_data) |
387 | return -ENOMEM; | 401 | return -ENOMEM; |
388 | 402 | ||
389 | return 0; | 403 | return 0; |
@@ -391,19 +405,21 @@ static int zfcp_allocate_low_mem_buffers(struct zfcp_adapter *adapter) | |||
391 | 405 | ||
392 | static void zfcp_free_low_mem_buffers(struct zfcp_adapter *adapter) | 406 | static void zfcp_free_low_mem_buffers(struct zfcp_adapter *adapter) |
393 | { | 407 | { |
394 | /* zfcp_data.config_sema must be held */ | 408 | /* zfcp_data.config_mutex must be held */ |
395 | if (adapter->pool.fsf_req_erp) | 409 | if (adapter->pool.erp_req) |
396 | mempool_destroy(adapter->pool.fsf_req_erp); | 410 | mempool_destroy(adapter->pool.erp_req); |
397 | if (adapter->pool.fsf_req_scsi) | 411 | if (adapter->pool.scsi_req) |
398 | mempool_destroy(adapter->pool.fsf_req_scsi); | 412 | mempool_destroy(adapter->pool.scsi_req); |
399 | if (adapter->pool.fsf_req_abort) | 413 | if (adapter->pool.scsi_abort) |
400 | mempool_destroy(adapter->pool.fsf_req_abort); | 414 | mempool_destroy(adapter->pool.scsi_abort); |
401 | if (adapter->pool.fsf_req_status_read) | 415 | if (adapter->pool.qtcb_pool) |
402 | mempool_destroy(adapter->pool.fsf_req_status_read); | 416 | mempool_destroy(adapter->pool.qtcb_pool); |
403 | if (adapter->pool.data_status_read) | 417 | if (adapter->pool.status_read_req) |
404 | mempool_destroy(adapter->pool.data_status_read); | 418 | mempool_destroy(adapter->pool.status_read_req); |
405 | if (adapter->pool.data_gid_pn) | 419 | if (adapter->pool.status_read_data) |
406 | mempool_destroy(adapter->pool.data_gid_pn); | 420 | mempool_destroy(adapter->pool.status_read_data); |
421 | if (adapter->pool.gid_pn_data) | ||
422 | mempool_destroy(adapter->pool.gid_pn_data); | ||
407 | } | 423 | } |
408 | 424 | ||
409 | /** | 425 | /** |
@@ -418,7 +434,7 @@ static void zfcp_free_low_mem_buffers(struct zfcp_adapter *adapter) | |||
418 | int zfcp_status_read_refill(struct zfcp_adapter *adapter) | 434 | int zfcp_status_read_refill(struct zfcp_adapter *adapter) |
419 | { | 435 | { |
420 | while (atomic_read(&adapter->stat_miss) > 0) | 436 | while (atomic_read(&adapter->stat_miss) > 0) |
421 | if (zfcp_fsf_status_read(adapter)) { | 437 | if (zfcp_fsf_status_read(adapter->qdio)) { |
422 | if (atomic_read(&adapter->stat_miss) >= 16) { | 438 | if (atomic_read(&adapter->stat_miss) >= 16) { |
423 | zfcp_erp_adapter_reopen(adapter, 0, "axsref1", | 439 | zfcp_erp_adapter_reopen(adapter, 0, "axsref1", |
424 | NULL); | 440 | NULL); |
@@ -446,6 +462,27 @@ static void zfcp_print_sl(struct seq_file *m, struct service_level *sl) | |||
446 | adapter->fsf_lic_version); | 462 | adapter->fsf_lic_version); |
447 | } | 463 | } |
448 | 464 | ||
465 | static int zfcp_setup_adapter_work_queue(struct zfcp_adapter *adapter) | ||
466 | { | ||
467 | char name[TASK_COMM_LEN]; | ||
468 | |||
469 | snprintf(name, sizeof(name), "zfcp_q_%s", | ||
470 | dev_name(&adapter->ccw_device->dev)); | ||
471 | adapter->work_queue = create_singlethread_workqueue(name); | ||
472 | |||
473 | if (adapter->work_queue) | ||
474 | return 0; | ||
475 | return -ENOMEM; | ||
476 | } | ||
477 | |||
478 | static void zfcp_destroy_adapter_work_queue(struct zfcp_adapter *adapter) | ||
479 | { | ||
480 | if (adapter->work_queue) | ||
481 | destroy_workqueue(adapter->work_queue); | ||
482 | adapter->work_queue = NULL; | ||
483 | |||
484 | } | ||
485 | |||
449 | /** | 486 | /** |
450 | * zfcp_adapter_enqueue - enqueue a new adapter to the list | 487 | * zfcp_adapter_enqueue - enqueue a new adapter to the list |
451 | * @ccw_device: pointer to the struct cc_device | 488 | * @ccw_device: pointer to the struct cc_device |
@@ -455,7 +492,7 @@ static void zfcp_print_sl(struct seq_file *m, struct service_level *sl) | |||
455 | * Enqueues an adapter at the end of the adapter list in the driver data. | 492 | * Enqueues an adapter at the end of the adapter list in the driver data. |
456 | * All adapter internal structures are set up. | 493 | * All adapter internal structures are set up. |
457 | * Proc-fs entries are also created. | 494 | * Proc-fs entries are also created. |
458 | * locks: config_sema must be held to serialise changes to the adapter list | 495 | * locks: config_mutex must be held to serialize changes to the adapter list |
459 | */ | 496 | */ |
460 | int zfcp_adapter_enqueue(struct ccw_device *ccw_device) | 497 | int zfcp_adapter_enqueue(struct ccw_device *ccw_device) |
461 | { | 498 | { |
@@ -463,37 +500,37 @@ int zfcp_adapter_enqueue(struct ccw_device *ccw_device) | |||
463 | 500 | ||
464 | /* | 501 | /* |
465 | * Note: It is safe to release the list_lock, as any list changes | 502 | * Note: It is safe to release the list_lock, as any list changes |
466 | * are protected by the config_sema, which must be held to get here | 503 | * are protected by the config_mutex, which must be held to get here |
467 | */ | 504 | */ |
468 | 505 | ||
469 | adapter = kzalloc(sizeof(struct zfcp_adapter), GFP_KERNEL); | 506 | adapter = kzalloc(sizeof(struct zfcp_adapter), GFP_KERNEL); |
470 | if (!adapter) | 507 | if (!adapter) |
471 | return -ENOMEM; | 508 | return -ENOMEM; |
472 | 509 | ||
473 | adapter->gs = kzalloc(sizeof(struct zfcp_wka_ports), GFP_KERNEL); | ||
474 | if (!adapter->gs) { | ||
475 | kfree(adapter); | ||
476 | return -ENOMEM; | ||
477 | } | ||
478 | |||
479 | ccw_device->handler = NULL; | 510 | ccw_device->handler = NULL; |
480 | adapter->ccw_device = ccw_device; | 511 | adapter->ccw_device = ccw_device; |
481 | atomic_set(&adapter->refcount, 0); | 512 | atomic_set(&adapter->refcount, 0); |
482 | 513 | ||
483 | if (zfcp_qdio_allocate(adapter)) | 514 | if (zfcp_qdio_setup(adapter)) |
484 | goto qdio_allocate_failed; | 515 | goto qdio_failed; |
485 | 516 | ||
486 | if (zfcp_allocate_low_mem_buffers(adapter)) | 517 | if (zfcp_allocate_low_mem_buffers(adapter)) |
487 | goto failed_low_mem_buffers; | 518 | goto low_mem_buffers_failed; |
488 | 519 | ||
489 | if (zfcp_reqlist_alloc(adapter)) | 520 | if (zfcp_reqlist_alloc(adapter)) |
490 | goto failed_low_mem_buffers; | 521 | goto low_mem_buffers_failed; |
491 | 522 | ||
492 | if (zfcp_adapter_debug_register(adapter)) | 523 | if (zfcp_dbf_adapter_register(adapter)) |
493 | goto debug_register_failed; | 524 | goto debug_register_failed; |
494 | 525 | ||
526 | if (zfcp_setup_adapter_work_queue(adapter)) | ||
527 | goto work_queue_failed; | ||
528 | |||
529 | if (zfcp_fc_gs_setup(adapter)) | ||
530 | goto generic_services_failed; | ||
531 | |||
495 | init_waitqueue_head(&adapter->remove_wq); | 532 | init_waitqueue_head(&adapter->remove_wq); |
496 | init_waitqueue_head(&adapter->erp_thread_wqh); | 533 | init_waitqueue_head(&adapter->erp_ready_wq); |
497 | init_waitqueue_head(&adapter->erp_done_wqh); | 534 | init_waitqueue_head(&adapter->erp_done_wqh); |
498 | 535 | ||
499 | INIT_LIST_HEAD(&adapter->port_list_head); | 536 | INIT_LIST_HEAD(&adapter->port_list_head); |
@@ -502,20 +539,14 @@ int zfcp_adapter_enqueue(struct ccw_device *ccw_device) | |||
502 | 539 | ||
503 | spin_lock_init(&adapter->req_list_lock); | 540 | spin_lock_init(&adapter->req_list_lock); |
504 | 541 | ||
505 | spin_lock_init(&adapter->hba_dbf_lock); | ||
506 | spin_lock_init(&adapter->san_dbf_lock); | ||
507 | spin_lock_init(&adapter->scsi_dbf_lock); | ||
508 | spin_lock_init(&adapter->rec_dbf_lock); | ||
509 | spin_lock_init(&adapter->req_q_lock); | ||
510 | spin_lock_init(&adapter->qdio_stat_lock); | ||
511 | |||
512 | rwlock_init(&adapter->erp_lock); | 542 | rwlock_init(&adapter->erp_lock); |
513 | rwlock_init(&adapter->abort_lock); | 543 | rwlock_init(&adapter->abort_lock); |
514 | 544 | ||
515 | sema_init(&adapter->erp_ready_sem, 0); | 545 | if (zfcp_erp_thread_setup(adapter)) |
546 | goto erp_thread_failed; | ||
516 | 547 | ||
517 | INIT_WORK(&adapter->stat_work, _zfcp_status_read_scheduler); | 548 | INIT_WORK(&adapter->stat_work, _zfcp_status_read_scheduler); |
518 | INIT_WORK(&adapter->scan_work, _zfcp_scan_ports_later); | 549 | INIT_WORK(&adapter->scan_work, _zfcp_fc_scan_ports_later); |
519 | 550 | ||
520 | adapter->service_level.seq_print = zfcp_print_sl; | 551 | adapter->service_level.seq_print = zfcp_print_sl; |
521 | 552 | ||
@@ -529,20 +560,25 @@ int zfcp_adapter_enqueue(struct ccw_device *ccw_device) | |||
529 | goto sysfs_failed; | 560 | goto sysfs_failed; |
530 | 561 | ||
531 | atomic_clear_mask(ZFCP_STATUS_COMMON_REMOVE, &adapter->status); | 562 | atomic_clear_mask(ZFCP_STATUS_COMMON_REMOVE, &adapter->status); |
532 | zfcp_fc_wka_ports_init(adapter); | ||
533 | 563 | ||
534 | if (!zfcp_adapter_scsi_register(adapter)) | 564 | if (!zfcp_adapter_scsi_register(adapter)) |
535 | return 0; | 565 | return 0; |
536 | 566 | ||
537 | sysfs_failed: | 567 | sysfs_failed: |
538 | zfcp_adapter_debug_unregister(adapter); | 568 | zfcp_erp_thread_kill(adapter); |
569 | erp_thread_failed: | ||
570 | zfcp_fc_gs_destroy(adapter); | ||
571 | generic_services_failed: | ||
572 | zfcp_destroy_adapter_work_queue(adapter); | ||
573 | work_queue_failed: | ||
574 | zfcp_dbf_adapter_unregister(adapter->dbf); | ||
539 | debug_register_failed: | 575 | debug_register_failed: |
540 | dev_set_drvdata(&ccw_device->dev, NULL); | 576 | dev_set_drvdata(&ccw_device->dev, NULL); |
541 | kfree(adapter->req_list); | 577 | kfree(adapter->req_list); |
542 | failed_low_mem_buffers: | 578 | low_mem_buffers_failed: |
543 | zfcp_free_low_mem_buffers(adapter); | 579 | zfcp_free_low_mem_buffers(adapter); |
544 | qdio_allocate_failed: | 580 | qdio_failed: |
545 | zfcp_qdio_free(adapter); | 581 | zfcp_qdio_destroy(adapter->qdio); |
546 | kfree(adapter); | 582 | kfree(adapter); |
547 | return -ENOMEM; | 583 | return -ENOMEM; |
548 | } | 584 | } |
@@ -559,6 +595,7 @@ void zfcp_adapter_dequeue(struct zfcp_adapter *adapter) | |||
559 | 595 | ||
560 | cancel_work_sync(&adapter->scan_work); | 596 | cancel_work_sync(&adapter->scan_work); |
561 | cancel_work_sync(&adapter->stat_work); | 597 | cancel_work_sync(&adapter->stat_work); |
598 | zfcp_fc_wka_ports_force_offline(adapter->gs); | ||
562 | zfcp_adapter_scsi_unregister(adapter); | 599 | zfcp_adapter_scsi_unregister(adapter); |
563 | sysfs_remove_group(&adapter->ccw_device->dev.kobj, | 600 | sysfs_remove_group(&adapter->ccw_device->dev.kobj, |
564 | &zfcp_sysfs_adapter_attrs); | 601 | &zfcp_sysfs_adapter_attrs); |
@@ -570,13 +607,15 @@ void zfcp_adapter_dequeue(struct zfcp_adapter *adapter) | |||
570 | if (!retval) | 607 | if (!retval) |
571 | return; | 608 | return; |
572 | 609 | ||
573 | zfcp_adapter_debug_unregister(adapter); | 610 | zfcp_fc_gs_destroy(adapter); |
574 | zfcp_qdio_free(adapter); | 611 | zfcp_erp_thread_kill(adapter); |
612 | zfcp_destroy_adapter_work_queue(adapter); | ||
613 | zfcp_dbf_adapter_unregister(adapter->dbf); | ||
575 | zfcp_free_low_mem_buffers(adapter); | 614 | zfcp_free_low_mem_buffers(adapter); |
615 | zfcp_qdio_destroy(adapter->qdio); | ||
576 | kfree(adapter->req_list); | 616 | kfree(adapter->req_list); |
577 | kfree(adapter->fc_stats); | 617 | kfree(adapter->fc_stats); |
578 | kfree(adapter->stats_reset_data); | 618 | kfree(adapter->stats_reset_data); |
579 | kfree(adapter->gs); | ||
580 | kfree(adapter); | 619 | kfree(adapter); |
581 | } | 620 | } |
582 | 621 | ||
@@ -592,7 +631,7 @@ static void zfcp_sysfs_port_release(struct device *dev) | |||
592 | * @status: initial status for the port | 631 | * @status: initial status for the port |
593 | * @d_id: destination id of the remote port to be enqueued | 632 | * @d_id: destination id of the remote port to be enqueued |
594 | * Returns: pointer to enqueued port on success, ERR_PTR on error | 633 | * Returns: pointer to enqueued port on success, ERR_PTR on error |
595 | * Locks: config_sema must be held to serialize changes to the port list | 634 | * Locks: config_mutex must be held to serialize changes to the port list |
596 | * | 635 | * |
597 | * All port internal structures are set up and the sysfs entry is generated. | 636 | * All port internal structures are set up and the sysfs entry is generated. |
598 | * d_id is used to enqueue ports with a well known address like the Directory | 637 | * d_id is used to enqueue ports with a well known address like the Directory |
@@ -602,7 +641,13 @@ struct zfcp_port *zfcp_port_enqueue(struct zfcp_adapter *adapter, u64 wwpn, | |||
602 | u32 status, u32 d_id) | 641 | u32 status, u32 d_id) |
603 | { | 642 | { |
604 | struct zfcp_port *port; | 643 | struct zfcp_port *port; |
605 | int retval; | 644 | |
645 | read_lock_irq(&zfcp_data.config_lock); | ||
646 | if (zfcp_get_port_by_wwpn(adapter, wwpn)) { | ||
647 | read_unlock_irq(&zfcp_data.config_lock); | ||
648 | return ERR_PTR(-EINVAL); | ||
649 | } | ||
650 | read_unlock_irq(&zfcp_data.config_lock); | ||
606 | 651 | ||
607 | port = kzalloc(sizeof(struct zfcp_port), GFP_KERNEL); | 652 | port = kzalloc(sizeof(struct zfcp_port), GFP_KERNEL); |
608 | if (!port) | 653 | if (!port) |
@@ -610,7 +655,7 @@ struct zfcp_port *zfcp_port_enqueue(struct zfcp_adapter *adapter, u64 wwpn, | |||
610 | 655 | ||
611 | init_waitqueue_head(&port->remove_wq); | 656 | init_waitqueue_head(&port->remove_wq); |
612 | INIT_LIST_HEAD(&port->unit_list_head); | 657 | INIT_LIST_HEAD(&port->unit_list_head); |
613 | INIT_WORK(&port->gid_pn_work, zfcp_erp_port_strategy_open_lookup); | 658 | INIT_WORK(&port->gid_pn_work, zfcp_fc_port_did_lookup); |
614 | INIT_WORK(&port->test_link_work, zfcp_fc_link_test_work); | 659 | INIT_WORK(&port->test_link_work, zfcp_fc_link_test_work); |
615 | INIT_WORK(&port->rport_work, zfcp_scsi_rport_work); | 660 | INIT_WORK(&port->rport_work, zfcp_scsi_rport_work); |
616 | 661 | ||
@@ -623,29 +668,24 @@ struct zfcp_port *zfcp_port_enqueue(struct zfcp_adapter *adapter, u64 wwpn, | |||
623 | atomic_set_mask(status | ZFCP_STATUS_COMMON_REMOVE, &port->status); | 668 | atomic_set_mask(status | ZFCP_STATUS_COMMON_REMOVE, &port->status); |
624 | atomic_set(&port->refcount, 0); | 669 | atomic_set(&port->refcount, 0); |
625 | 670 | ||
626 | dev_set_name(&port->sysfs_device, "0x%016llx", | 671 | if (dev_set_name(&port->sysfs_device, "0x%016llx", |
627 | (unsigned long long)wwpn); | 672 | (unsigned long long)wwpn)) { |
673 | kfree(port); | ||
674 | return ERR_PTR(-ENOMEM); | ||
675 | } | ||
628 | port->sysfs_device.parent = &adapter->ccw_device->dev; | 676 | port->sysfs_device.parent = &adapter->ccw_device->dev; |
629 | |||
630 | port->sysfs_device.release = zfcp_sysfs_port_release; | 677 | port->sysfs_device.release = zfcp_sysfs_port_release; |
631 | dev_set_drvdata(&port->sysfs_device, port); | 678 | dev_set_drvdata(&port->sysfs_device, port); |
632 | 679 | ||
633 | read_lock_irq(&zfcp_data.config_lock); | 680 | if (device_register(&port->sysfs_device)) { |
634 | if (zfcp_get_port_by_wwpn(adapter, wwpn)) { | 681 | put_device(&port->sysfs_device); |
635 | read_unlock_irq(&zfcp_data.config_lock); | 682 | return ERR_PTR(-EINVAL); |
636 | goto err_out_free; | ||
637 | } | 683 | } |
638 | read_unlock_irq(&zfcp_data.config_lock); | ||
639 | 684 | ||
640 | if (device_register(&port->sysfs_device)) | 685 | if (sysfs_create_group(&port->sysfs_device.kobj, |
641 | goto err_out_free; | 686 | &zfcp_sysfs_port_attrs)) { |
642 | |||
643 | retval = sysfs_create_group(&port->sysfs_device.kobj, | ||
644 | &zfcp_sysfs_port_attrs); | ||
645 | |||
646 | if (retval) { | ||
647 | device_unregister(&port->sysfs_device); | 687 | device_unregister(&port->sysfs_device); |
648 | goto err_out; | 688 | return ERR_PTR(-EINVAL); |
649 | } | 689 | } |
650 | 690 | ||
651 | zfcp_port_get(port); | 691 | zfcp_port_get(port); |
@@ -659,11 +699,6 @@ struct zfcp_port *zfcp_port_enqueue(struct zfcp_adapter *adapter, u64 wwpn, | |||
659 | 699 | ||
660 | zfcp_adapter_get(adapter); | 700 | zfcp_adapter_get(adapter); |
661 | return port; | 701 | return port; |
662 | |||
663 | err_out_free: | ||
664 | kfree(port); | ||
665 | err_out: | ||
666 | return ERR_PTR(-EINVAL); | ||
667 | } | 702 | } |
668 | 703 | ||
669 | /** | 704 | /** |
@@ -672,12 +707,11 @@ err_out: | |||
672 | */ | 707 | */ |
673 | void zfcp_port_dequeue(struct zfcp_port *port) | 708 | void zfcp_port_dequeue(struct zfcp_port *port) |
674 | { | 709 | { |
675 | wait_event(port->remove_wq, atomic_read(&port->refcount) == 0); | ||
676 | write_lock_irq(&zfcp_data.config_lock); | 710 | write_lock_irq(&zfcp_data.config_lock); |
677 | list_del(&port->list); | 711 | list_del(&port->list); |
678 | write_unlock_irq(&zfcp_data.config_lock); | 712 | write_unlock_irq(&zfcp_data.config_lock); |
679 | if (port->rport) | 713 | wait_event(port->remove_wq, atomic_read(&port->refcount) == 0); |
680 | port->rport->dd_data = NULL; | 714 | cancel_work_sync(&port->rport_work); /* usually not necessary */ |
681 | zfcp_adapter_put(port->adapter); | 715 | zfcp_adapter_put(port->adapter); |
682 | sysfs_remove_group(&port->sysfs_device.kobj, &zfcp_sysfs_port_attrs); | 716 | sysfs_remove_group(&port->sysfs_device.kobj, &zfcp_sysfs_port_attrs); |
683 | device_unregister(&port->sysfs_device); | 717 | device_unregister(&port->sysfs_device); |
diff --git a/drivers/s390/scsi/zfcp_ccw.c b/drivers/s390/scsi/zfcp_ccw.c index d9da5c42ccbe..0c90f8e71605 100644 --- a/drivers/s390/scsi/zfcp_ccw.c +++ b/drivers/s390/scsi/zfcp_ccw.c | |||
@@ -18,12 +18,15 @@ static int zfcp_ccw_suspend(struct ccw_device *cdev) | |||
18 | { | 18 | { |
19 | struct zfcp_adapter *adapter = dev_get_drvdata(&cdev->dev); | 19 | struct zfcp_adapter *adapter = dev_get_drvdata(&cdev->dev); |
20 | 20 | ||
21 | down(&zfcp_data.config_sema); | 21 | if (!adapter) |
22 | return 0; | ||
23 | |||
24 | mutex_lock(&zfcp_data.config_mutex); | ||
22 | 25 | ||
23 | zfcp_erp_adapter_shutdown(adapter, 0, "ccsusp1", NULL); | 26 | zfcp_erp_adapter_shutdown(adapter, 0, "ccsusp1", NULL); |
24 | zfcp_erp_wait(adapter); | 27 | zfcp_erp_wait(adapter); |
25 | 28 | ||
26 | up(&zfcp_data.config_sema); | 29 | mutex_unlock(&zfcp_data.config_mutex); |
27 | 30 | ||
28 | return 0; | 31 | return 0; |
29 | } | 32 | } |
@@ -33,6 +36,9 @@ static int zfcp_ccw_activate(struct ccw_device *cdev) | |||
33 | { | 36 | { |
34 | struct zfcp_adapter *adapter = dev_get_drvdata(&cdev->dev); | 37 | struct zfcp_adapter *adapter = dev_get_drvdata(&cdev->dev); |
35 | 38 | ||
39 | if (!adapter) | ||
40 | return 0; | ||
41 | |||
36 | zfcp_erp_modify_adapter_status(adapter, "ccresu1", NULL, | 42 | zfcp_erp_modify_adapter_status(adapter, "ccresu1", NULL, |
37 | ZFCP_STATUS_COMMON_RUNNING, ZFCP_SET); | 43 | ZFCP_STATUS_COMMON_RUNNING, ZFCP_SET); |
38 | zfcp_erp_adapter_reopen(adapter, ZFCP_STATUS_COMMON_ERP_FAILED, | 44 | zfcp_erp_adapter_reopen(adapter, ZFCP_STATUS_COMMON_ERP_FAILED, |
@@ -63,25 +69,14 @@ int zfcp_ccw_priv_sch(struct zfcp_adapter *adapter) | |||
63 | * zfcp_ccw_probe - probe function of zfcp driver | 69 | * zfcp_ccw_probe - probe function of zfcp driver |
64 | * @ccw_device: pointer to belonging ccw device | 70 | * @ccw_device: pointer to belonging ccw device |
65 | * | 71 | * |
66 | * This function gets called by the common i/o layer and sets up the initial | 72 | * This function gets called by the common i/o layer for each FCP |
67 | * data structures for each fcp adapter, which was detected by the system. | 73 | * device found on the current system. This is only a stub to make cio |
68 | * Also the sysfs files for this adapter will be created by this function. | 74 | * work: To only allocate adapter resources for devices actually used, |
69 | * In addition the nameserver port will be added to the ports of the adapter | 75 | * the allocation is deferred to the first call to ccw_set_online. |
70 | * and its sysfs representation will be created too. | ||
71 | */ | 76 | */ |
72 | static int zfcp_ccw_probe(struct ccw_device *ccw_device) | 77 | static int zfcp_ccw_probe(struct ccw_device *ccw_device) |
73 | { | 78 | { |
74 | int retval = 0; | 79 | return 0; |
75 | |||
76 | down(&zfcp_data.config_sema); | ||
77 | if (zfcp_adapter_enqueue(ccw_device)) { | ||
78 | dev_err(&ccw_device->dev, | ||
79 | "Setting up data structures for the " | ||
80 | "FCP adapter failed\n"); | ||
81 | retval = -EINVAL; | ||
82 | } | ||
83 | up(&zfcp_data.config_sema); | ||
84 | return retval; | ||
85 | } | 80 | } |
86 | 81 | ||
87 | /** | 82 | /** |
@@ -102,8 +97,11 @@ static void zfcp_ccw_remove(struct ccw_device *ccw_device) | |||
102 | LIST_HEAD(port_remove_lh); | 97 | LIST_HEAD(port_remove_lh); |
103 | 98 | ||
104 | ccw_device_set_offline(ccw_device); | 99 | ccw_device_set_offline(ccw_device); |
105 | down(&zfcp_data.config_sema); | 100 | |
101 | mutex_lock(&zfcp_data.config_mutex); | ||
106 | adapter = dev_get_drvdata(&ccw_device->dev); | 102 | adapter = dev_get_drvdata(&ccw_device->dev); |
103 | if (!adapter) | ||
104 | goto out; | ||
107 | 105 | ||
108 | write_lock_irq(&zfcp_data.config_lock); | 106 | write_lock_irq(&zfcp_data.config_lock); |
109 | list_for_each_entry_safe(port, p, &adapter->port_list_head, list) { | 107 | list_for_each_entry_safe(port, p, &adapter->port_list_head, list) { |
@@ -129,29 +127,41 @@ static void zfcp_ccw_remove(struct ccw_device *ccw_device) | |||
129 | wait_event(adapter->remove_wq, atomic_read(&adapter->refcount) == 0); | 127 | wait_event(adapter->remove_wq, atomic_read(&adapter->refcount) == 0); |
130 | zfcp_adapter_dequeue(adapter); | 128 | zfcp_adapter_dequeue(adapter); |
131 | 129 | ||
132 | up(&zfcp_data.config_sema); | 130 | out: |
131 | mutex_unlock(&zfcp_data.config_mutex); | ||
133 | } | 132 | } |
134 | 133 | ||
135 | /** | 134 | /** |
136 | * zfcp_ccw_set_online - set_online function of zfcp driver | 135 | * zfcp_ccw_set_online - set_online function of zfcp driver |
137 | * @ccw_device: pointer to belonging ccw device | 136 | * @ccw_device: pointer to belonging ccw device |
138 | * | 137 | * |
139 | * This function gets called by the common i/o layer and sets an adapter | 138 | * This function gets called by the common i/o layer and sets an |
140 | * into state online. Setting an fcp device online means that it will be | 139 | * adapter into state online. The first call will allocate all |
141 | * registered with the SCSI stack, that the QDIO queues will be set up | 140 | * adapter resources that will be retained until the device is removed |
142 | * and that the adapter will be opened (asynchronously). | 141 | * via zfcp_ccw_remove. |
142 | * | ||
143 | * Setting an fcp device online means that it will be registered with | ||
144 | * the SCSI stack, that the QDIO queues will be set up and that the | ||
145 | * adapter will be opened. | ||
143 | */ | 146 | */ |
144 | static int zfcp_ccw_set_online(struct ccw_device *ccw_device) | 147 | static int zfcp_ccw_set_online(struct ccw_device *ccw_device) |
145 | { | 148 | { |
146 | struct zfcp_adapter *adapter; | 149 | struct zfcp_adapter *adapter; |
147 | int retval; | 150 | int ret = 0; |
148 | 151 | ||
149 | down(&zfcp_data.config_sema); | 152 | mutex_lock(&zfcp_data.config_mutex); |
150 | adapter = dev_get_drvdata(&ccw_device->dev); | 153 | adapter = dev_get_drvdata(&ccw_device->dev); |
151 | 154 | ||
152 | retval = zfcp_erp_thread_setup(adapter); | 155 | if (!adapter) { |
153 | if (retval) | 156 | ret = zfcp_adapter_enqueue(ccw_device); |
154 | goto out; | 157 | if (ret) { |
158 | dev_err(&ccw_device->dev, | ||
159 | "Setting up data structures for the " | ||
160 | "FCP adapter failed\n"); | ||
161 | goto out; | ||
162 | } | ||
163 | adapter = dev_get_drvdata(&ccw_device->dev); | ||
164 | } | ||
155 | 165 | ||
156 | /* initialize request counter */ | 166 | /* initialize request counter */ |
157 | BUG_ON(!zfcp_reqlist_isempty(adapter)); | 167 | BUG_ON(!zfcp_reqlist_isempty(adapter)); |
@@ -162,13 +172,11 @@ static int zfcp_ccw_set_online(struct ccw_device *ccw_device) | |||
162 | zfcp_erp_adapter_reopen(adapter, ZFCP_STATUS_COMMON_ERP_FAILED, | 172 | zfcp_erp_adapter_reopen(adapter, ZFCP_STATUS_COMMON_ERP_FAILED, |
163 | "ccsonl2", NULL); | 173 | "ccsonl2", NULL); |
164 | zfcp_erp_wait(adapter); | 174 | zfcp_erp_wait(adapter); |
165 | up(&zfcp_data.config_sema); | 175 | out: |
166 | flush_work(&adapter->scan_work); | 176 | mutex_unlock(&zfcp_data.config_mutex); |
167 | return 0; | 177 | if (!ret) |
168 | 178 | flush_work(&adapter->scan_work); | |
169 | out: | 179 | return ret; |
170 | up(&zfcp_data.config_sema); | ||
171 | return retval; | ||
172 | } | 180 | } |
173 | 181 | ||
174 | /** | 182 | /** |
@@ -182,12 +190,15 @@ static int zfcp_ccw_set_offline(struct ccw_device *ccw_device) | |||
182 | { | 190 | { |
183 | struct zfcp_adapter *adapter; | 191 | struct zfcp_adapter *adapter; |
184 | 192 | ||
185 | down(&zfcp_data.config_sema); | 193 | mutex_lock(&zfcp_data.config_mutex); |
186 | adapter = dev_get_drvdata(&ccw_device->dev); | 194 | adapter = dev_get_drvdata(&ccw_device->dev); |
195 | if (!adapter) | ||
196 | goto out; | ||
197 | |||
187 | zfcp_erp_adapter_shutdown(adapter, 0, "ccsoff1", NULL); | 198 | zfcp_erp_adapter_shutdown(adapter, 0, "ccsoff1", NULL); |
188 | zfcp_erp_wait(adapter); | 199 | zfcp_erp_wait(adapter); |
189 | zfcp_erp_thread_kill(adapter); | 200 | mutex_unlock(&zfcp_data.config_mutex); |
190 | up(&zfcp_data.config_sema); | 201 | out: |
191 | return 0; | 202 | return 0; |
192 | } | 203 | } |
193 | 204 | ||
@@ -240,11 +251,12 @@ static void zfcp_ccw_shutdown(struct ccw_device *cdev) | |||
240 | { | 251 | { |
241 | struct zfcp_adapter *adapter; | 252 | struct zfcp_adapter *adapter; |
242 | 253 | ||
243 | down(&zfcp_data.config_sema); | 254 | mutex_lock(&zfcp_data.config_mutex); |
244 | adapter = dev_get_drvdata(&cdev->dev); | 255 | adapter = dev_get_drvdata(&cdev->dev); |
245 | zfcp_erp_adapter_shutdown(adapter, 0, "ccshut1", NULL); | 256 | zfcp_erp_adapter_shutdown(adapter, 0, "ccshut1", NULL); |
246 | zfcp_erp_wait(adapter); | 257 | zfcp_erp_wait(adapter); |
247 | up(&zfcp_data.config_sema); | 258 | zfcp_erp_thread_kill(adapter); |
259 | mutex_unlock(&zfcp_data.config_mutex); | ||
248 | } | 260 | } |
249 | 261 | ||
250 | static struct ccw_driver zfcp_ccw_driver = { | 262 | static struct ccw_driver zfcp_ccw_driver = { |
diff --git a/drivers/s390/scsi/zfcp_dbf.c b/drivers/s390/scsi/zfcp_dbf.c index b99b87ce5a39..215b70749e95 100644 --- a/drivers/s390/scsi/zfcp_dbf.c +++ b/drivers/s390/scsi/zfcp_dbf.c | |||
@@ -3,7 +3,7 @@ | |||
3 | * | 3 | * |
4 | * Debug traces for zfcp. | 4 | * Debug traces for zfcp. |
5 | * | 5 | * |
6 | * Copyright IBM Corporation 2002, 2008 | 6 | * Copyright IBM Corporation 2002, 2009 |
7 | */ | 7 | */ |
8 | 8 | ||
9 | #define KMSG_COMPONENT "zfcp" | 9 | #define KMSG_COMPONENT "zfcp" |
@@ -11,6 +11,7 @@ | |||
11 | 11 | ||
12 | #include <linux/ctype.h> | 12 | #include <linux/ctype.h> |
13 | #include <asm/debug.h> | 13 | #include <asm/debug.h> |
14 | #include "zfcp_dbf.h" | ||
14 | #include "zfcp_ext.h" | 15 | #include "zfcp_ext.h" |
15 | 16 | ||
16 | static u32 dbfsize = 4; | 17 | static u32 dbfsize = 4; |
@@ -37,19 +38,6 @@ static void zfcp_dbf_hexdump(debug_info_t *dbf, void *to, int to_len, | |||
37 | } | 38 | } |
38 | } | 39 | } |
39 | 40 | ||
40 | /* FIXME: this duplicate this code in s390 debug feature */ | ||
41 | static void zfcp_dbf_timestamp(unsigned long long stck, struct timespec *time) | ||
42 | { | ||
43 | unsigned long long sec; | ||
44 | |||
45 | stck -= 0x8126d60e46000000LL - (0x3c26700LL * 1000000 * 4096); | ||
46 | sec = stck >> 12; | ||
47 | do_div(sec, 1000000); | ||
48 | time->tv_sec = sec; | ||
49 | stck -= (sec * 1000000) << 12; | ||
50 | time->tv_nsec = ((stck * 1000) >> 12); | ||
51 | } | ||
52 | |||
53 | static void zfcp_dbf_tag(char **p, const char *label, const char *tag) | 41 | static void zfcp_dbf_tag(char **p, const char *label, const char *tag) |
54 | { | 42 | { |
55 | int i; | 43 | int i; |
@@ -106,7 +94,7 @@ static int zfcp_dbf_view_header(debug_info_t *id, struct debug_view *view, | |||
106 | char *p = out_buf; | 94 | char *p = out_buf; |
107 | 95 | ||
108 | if (strncmp(dump->tag, "dump", ZFCP_DBF_TAG_SIZE) != 0) { | 96 | if (strncmp(dump->tag, "dump", ZFCP_DBF_TAG_SIZE) != 0) { |
109 | zfcp_dbf_timestamp(entry->id.stck, &t); | 97 | stck_to_timespec(entry->id.stck, &t); |
110 | zfcp_dbf_out(&p, "timestamp", "%011lu:%06lu", | 98 | zfcp_dbf_out(&p, "timestamp", "%011lu:%06lu", |
111 | t.tv_sec, t.tv_nsec); | 99 | t.tv_sec, t.tv_nsec); |
112 | zfcp_dbf_out(&p, "cpu", "%02i", entry->id.fields.cpuid); | 100 | zfcp_dbf_out(&p, "cpu", "%02i", entry->id.fields.cpuid); |
@@ -119,13 +107,10 @@ static int zfcp_dbf_view_header(debug_info_t *id, struct debug_view *view, | |||
119 | return p - out_buf; | 107 | return p - out_buf; |
120 | } | 108 | } |
121 | 109 | ||
122 | /** | 110 | void _zfcp_dbf_hba_fsf_response(const char *tag2, int level, |
123 | * zfcp_hba_dbf_event_fsf_response - trace event for request completion | 111 | struct zfcp_fsf_req *fsf_req, |
124 | * @fsf_req: request that has been completed | 112 | struct zfcp_dbf *dbf) |
125 | */ | ||
126 | void zfcp_hba_dbf_event_fsf_response(struct zfcp_fsf_req *fsf_req) | ||
127 | { | 113 | { |
128 | struct zfcp_adapter *adapter = fsf_req->adapter; | ||
129 | struct fsf_qtcb *qtcb = fsf_req->qtcb; | 114 | struct fsf_qtcb *qtcb = fsf_req->qtcb; |
130 | union fsf_prot_status_qual *prot_status_qual = | 115 | union fsf_prot_status_qual *prot_status_qual = |
131 | &qtcb->prefix.prot_status_qual; | 116 | &qtcb->prefix.prot_status_qual; |
@@ -134,33 +119,14 @@ void zfcp_hba_dbf_event_fsf_response(struct zfcp_fsf_req *fsf_req) | |||
134 | struct zfcp_port *port; | 119 | struct zfcp_port *port; |
135 | struct zfcp_unit *unit; | 120 | struct zfcp_unit *unit; |
136 | struct zfcp_send_els *send_els; | 121 | struct zfcp_send_els *send_els; |
137 | struct zfcp_hba_dbf_record *rec = &adapter->hba_dbf_buf; | 122 | struct zfcp_dbf_hba_record *rec = &dbf->hba_buf; |
138 | struct zfcp_hba_dbf_record_response *response = &rec->u.response; | 123 | struct zfcp_dbf_hba_record_response *response = &rec->u.response; |
139 | int level; | ||
140 | unsigned long flags; | 124 | unsigned long flags; |
141 | 125 | ||
142 | spin_lock_irqsave(&adapter->hba_dbf_lock, flags); | 126 | spin_lock_irqsave(&dbf->hba_lock, flags); |
143 | memset(rec, 0, sizeof(*rec)); | 127 | memset(rec, 0, sizeof(*rec)); |
144 | strncpy(rec->tag, "resp", ZFCP_DBF_TAG_SIZE); | 128 | strncpy(rec->tag, "resp", ZFCP_DBF_TAG_SIZE); |
145 | 129 | strncpy(rec->tag2, tag2, ZFCP_DBF_TAG_SIZE); | |
146 | if ((qtcb->prefix.prot_status != FSF_PROT_GOOD) && | ||
147 | (qtcb->prefix.prot_status != FSF_PROT_FSF_STATUS_PRESENTED)) { | ||
148 | strncpy(rec->tag2, "perr", ZFCP_DBF_TAG_SIZE); | ||
149 | level = 1; | ||
150 | } else if (qtcb->header.fsf_status != FSF_GOOD) { | ||
151 | strncpy(rec->tag2, "ferr", ZFCP_DBF_TAG_SIZE); | ||
152 | level = 1; | ||
153 | } else if ((fsf_req->fsf_command == FSF_QTCB_OPEN_PORT_WITH_DID) || | ||
154 | (fsf_req->fsf_command == FSF_QTCB_OPEN_LUN)) { | ||
155 | strncpy(rec->tag2, "open", ZFCP_DBF_TAG_SIZE); | ||
156 | level = 4; | ||
157 | } else if (qtcb->header.log_length) { | ||
158 | strncpy(rec->tag2, "qtcb", ZFCP_DBF_TAG_SIZE); | ||
159 | level = 5; | ||
160 | } else { | ||
161 | strncpy(rec->tag2, "norm", ZFCP_DBF_TAG_SIZE); | ||
162 | level = 6; | ||
163 | } | ||
164 | 130 | ||
165 | response->fsf_command = fsf_req->fsf_command; | 131 | response->fsf_command = fsf_req->fsf_command; |
166 | response->fsf_reqid = fsf_req->req_id; | 132 | response->fsf_reqid = fsf_req->req_id; |
@@ -173,9 +139,9 @@ void zfcp_hba_dbf_event_fsf_response(struct zfcp_fsf_req *fsf_req) | |||
173 | memcpy(response->fsf_status_qual, | 139 | memcpy(response->fsf_status_qual, |
174 | fsf_status_qual, FSF_STATUS_QUALIFIER_SIZE); | 140 | fsf_status_qual, FSF_STATUS_QUALIFIER_SIZE); |
175 | response->fsf_req_status = fsf_req->status; | 141 | response->fsf_req_status = fsf_req->status; |
176 | response->sbal_first = fsf_req->sbal_first; | 142 | response->sbal_first = fsf_req->queue_req.sbal_first; |
177 | response->sbal_last = fsf_req->sbal_last; | 143 | response->sbal_last = fsf_req->queue_req.sbal_last; |
178 | response->sbal_response = fsf_req->sbal_response; | 144 | response->sbal_response = fsf_req->queue_req.sbal_response; |
179 | response->pool = fsf_req->pool != NULL; | 145 | response->pool = fsf_req->pool != NULL; |
180 | response->erp_action = (unsigned long)fsf_req->erp_action; | 146 | response->erp_action = (unsigned long)fsf_req->erp_action; |
181 | 147 | ||
@@ -224,7 +190,7 @@ void zfcp_hba_dbf_event_fsf_response(struct zfcp_fsf_req *fsf_req) | |||
224 | break; | 190 | break; |
225 | } | 191 | } |
226 | 192 | ||
227 | debug_event(adapter->hba_dbf, level, rec, sizeof(*rec)); | 193 | debug_event(dbf->hba, level, rec, sizeof(*rec)); |
228 | 194 | ||
229 | /* have fcp channel microcode fixed to use as little as possible */ | 195 | /* have fcp channel microcode fixed to use as little as possible */ |
230 | if (fsf_req->fsf_command != FSF_QTCB_FCP_CMND) { | 196 | if (fsf_req->fsf_command != FSF_QTCB_FCP_CMND) { |
@@ -232,31 +198,25 @@ void zfcp_hba_dbf_event_fsf_response(struct zfcp_fsf_req *fsf_req) | |||
232 | char *buf = (char *)qtcb + qtcb->header.log_start; | 198 | char *buf = (char *)qtcb + qtcb->header.log_start; |
233 | int len = qtcb->header.log_length; | 199 | int len = qtcb->header.log_length; |
234 | for (; len && !buf[len - 1]; len--); | 200 | for (; len && !buf[len - 1]; len--); |
235 | zfcp_dbf_hexdump(adapter->hba_dbf, rec, sizeof(*rec), level, | 201 | zfcp_dbf_hexdump(dbf->hba, rec, sizeof(*rec), level, buf, |
236 | buf, len); | 202 | len); |
237 | } | 203 | } |
238 | 204 | ||
239 | spin_unlock_irqrestore(&adapter->hba_dbf_lock, flags); | 205 | spin_unlock_irqrestore(&dbf->hba_lock, flags); |
240 | } | 206 | } |
241 | 207 | ||
242 | /** | 208 | void _zfcp_dbf_hba_fsf_unsol(const char *tag, int level, struct zfcp_dbf *dbf, |
243 | * zfcp_hba_dbf_event_fsf_unsol - trace event for an unsolicited status buffer | 209 | struct fsf_status_read_buffer *status_buffer) |
244 | * @tag: tag indicating which kind of unsolicited status has been received | ||
245 | * @adapter: adapter that has issued the unsolicited status buffer | ||
246 | * @status_buffer: buffer containing payload of unsolicited status | ||
247 | */ | ||
248 | void zfcp_hba_dbf_event_fsf_unsol(const char *tag, struct zfcp_adapter *adapter, | ||
249 | struct fsf_status_read_buffer *status_buffer) | ||
250 | { | 210 | { |
251 | struct zfcp_hba_dbf_record *rec = &adapter->hba_dbf_buf; | 211 | struct zfcp_dbf_hba_record *rec = &dbf->hba_buf; |
252 | unsigned long flags; | 212 | unsigned long flags; |
253 | 213 | ||
254 | spin_lock_irqsave(&adapter->hba_dbf_lock, flags); | 214 | spin_lock_irqsave(&dbf->hba_lock, flags); |
255 | memset(rec, 0, sizeof(*rec)); | 215 | memset(rec, 0, sizeof(*rec)); |
256 | strncpy(rec->tag, "stat", ZFCP_DBF_TAG_SIZE); | 216 | strncpy(rec->tag, "stat", ZFCP_DBF_TAG_SIZE); |
257 | strncpy(rec->tag2, tag, ZFCP_DBF_TAG_SIZE); | 217 | strncpy(rec->tag2, tag, ZFCP_DBF_TAG_SIZE); |
258 | 218 | ||
259 | rec->u.status.failed = atomic_read(&adapter->stat_miss); | 219 | rec->u.status.failed = atomic_read(&dbf->adapter->stat_miss); |
260 | if (status_buffer != NULL) { | 220 | if (status_buffer != NULL) { |
261 | rec->u.status.status_type = status_buffer->status_type; | 221 | rec->u.status.status_type = status_buffer->status_type; |
262 | rec->u.status.status_subtype = status_buffer->status_subtype; | 222 | rec->u.status.status_subtype = status_buffer->status_subtype; |
@@ -293,63 +253,61 @@ void zfcp_hba_dbf_event_fsf_unsol(const char *tag, struct zfcp_adapter *adapter, | |||
293 | &status_buffer->payload, rec->u.status.payload_size); | 253 | &status_buffer->payload, rec->u.status.payload_size); |
294 | } | 254 | } |
295 | 255 | ||
296 | debug_event(adapter->hba_dbf, 2, rec, sizeof(*rec)); | 256 | debug_event(dbf->hba, level, rec, sizeof(*rec)); |
297 | spin_unlock_irqrestore(&adapter->hba_dbf_lock, flags); | 257 | spin_unlock_irqrestore(&dbf->hba_lock, flags); |
298 | } | 258 | } |
299 | 259 | ||
300 | /** | 260 | /** |
301 | * zfcp_hba_dbf_event_qdio - trace event for QDIO related failure | 261 | * zfcp_dbf_hba_qdio - trace event for QDIO related failure |
302 | * @adapter: adapter affected by this QDIO related event | 262 | * @qdio: qdio structure affected by this QDIO related event |
303 | * @qdio_error: as passed by qdio module | 263 | * @qdio_error: as passed by qdio module |
304 | * @sbal_index: first buffer with error condition, as passed by qdio module | 264 | * @sbal_index: first buffer with error condition, as passed by qdio module |
305 | * @sbal_count: number of buffers affected, as passed by qdio module | 265 | * @sbal_count: number of buffers affected, as passed by qdio module |
306 | */ | 266 | */ |
307 | void zfcp_hba_dbf_event_qdio(struct zfcp_adapter *adapter, | 267 | void zfcp_dbf_hba_qdio(struct zfcp_dbf *dbf, unsigned int qdio_error, |
308 | unsigned int qdio_error, int sbal_index, | 268 | int sbal_index, int sbal_count) |
309 | int sbal_count) | ||
310 | { | 269 | { |
311 | struct zfcp_hba_dbf_record *r = &adapter->hba_dbf_buf; | 270 | struct zfcp_dbf_hba_record *r = &dbf->hba_buf; |
312 | unsigned long flags; | 271 | unsigned long flags; |
313 | 272 | ||
314 | spin_lock_irqsave(&adapter->hba_dbf_lock, flags); | 273 | spin_lock_irqsave(&dbf->hba_lock, flags); |
315 | memset(r, 0, sizeof(*r)); | 274 | memset(r, 0, sizeof(*r)); |
316 | strncpy(r->tag, "qdio", ZFCP_DBF_TAG_SIZE); | 275 | strncpy(r->tag, "qdio", ZFCP_DBF_TAG_SIZE); |
317 | r->u.qdio.qdio_error = qdio_error; | 276 | r->u.qdio.qdio_error = qdio_error; |
318 | r->u.qdio.sbal_index = sbal_index; | 277 | r->u.qdio.sbal_index = sbal_index; |
319 | r->u.qdio.sbal_count = sbal_count; | 278 | r->u.qdio.sbal_count = sbal_count; |
320 | debug_event(adapter->hba_dbf, 0, r, sizeof(*r)); | 279 | debug_event(dbf->hba, 0, r, sizeof(*r)); |
321 | spin_unlock_irqrestore(&adapter->hba_dbf_lock, flags); | 280 | spin_unlock_irqrestore(&dbf->hba_lock, flags); |
322 | } | 281 | } |
323 | 282 | ||
324 | /** | 283 | /** |
325 | * zfcp_hba_dbf_event_berr - trace event for bit error threshold | 284 | * zfcp_dbf_hba_berr - trace event for bit error threshold |
326 | * @adapter: adapter affected by this QDIO related event | 285 | * @dbf: dbf structure affected by this QDIO related event |
327 | * @req: fsf request | 286 | * @req: fsf request |
328 | */ | 287 | */ |
329 | void zfcp_hba_dbf_event_berr(struct zfcp_adapter *adapter, | 288 | void zfcp_dbf_hba_berr(struct zfcp_dbf *dbf, struct zfcp_fsf_req *req) |
330 | struct zfcp_fsf_req *req) | ||
331 | { | 289 | { |
332 | struct zfcp_hba_dbf_record *r = &adapter->hba_dbf_buf; | 290 | struct zfcp_dbf_hba_record *r = &dbf->hba_buf; |
333 | struct fsf_status_read_buffer *sr_buf = req->data; | 291 | struct fsf_status_read_buffer *sr_buf = req->data; |
334 | struct fsf_bit_error_payload *err = &sr_buf->payload.bit_error; | 292 | struct fsf_bit_error_payload *err = &sr_buf->payload.bit_error; |
335 | unsigned long flags; | 293 | unsigned long flags; |
336 | 294 | ||
337 | spin_lock_irqsave(&adapter->hba_dbf_lock, flags); | 295 | spin_lock_irqsave(&dbf->hba_lock, flags); |
338 | memset(r, 0, sizeof(*r)); | 296 | memset(r, 0, sizeof(*r)); |
339 | strncpy(r->tag, "berr", ZFCP_DBF_TAG_SIZE); | 297 | strncpy(r->tag, "berr", ZFCP_DBF_TAG_SIZE); |
340 | memcpy(&r->u.berr, err, sizeof(struct fsf_bit_error_payload)); | 298 | memcpy(&r->u.berr, err, sizeof(struct fsf_bit_error_payload)); |
341 | debug_event(adapter->hba_dbf, 0, r, sizeof(*r)); | 299 | debug_event(dbf->hba, 0, r, sizeof(*r)); |
342 | spin_unlock_irqrestore(&adapter->hba_dbf_lock, flags); | 300 | spin_unlock_irqrestore(&dbf->hba_lock, flags); |
343 | } | 301 | } |
344 | static void zfcp_hba_dbf_view_response(char **p, | 302 | static void zfcp_dbf_hba_view_response(char **p, |
345 | struct zfcp_hba_dbf_record_response *r) | 303 | struct zfcp_dbf_hba_record_response *r) |
346 | { | 304 | { |
347 | struct timespec t; | 305 | struct timespec t; |
348 | 306 | ||
349 | zfcp_dbf_out(p, "fsf_command", "0x%08x", r->fsf_command); | 307 | zfcp_dbf_out(p, "fsf_command", "0x%08x", r->fsf_command); |
350 | zfcp_dbf_out(p, "fsf_reqid", "0x%0Lx", r->fsf_reqid); | 308 | zfcp_dbf_out(p, "fsf_reqid", "0x%0Lx", r->fsf_reqid); |
351 | zfcp_dbf_out(p, "fsf_seqno", "0x%08x", r->fsf_seqno); | 309 | zfcp_dbf_out(p, "fsf_seqno", "0x%08x", r->fsf_seqno); |
352 | zfcp_dbf_timestamp(r->fsf_issued, &t); | 310 | stck_to_timespec(r->fsf_issued, &t); |
353 | zfcp_dbf_out(p, "fsf_issued", "%011lu:%06lu", t.tv_sec, t.tv_nsec); | 311 | zfcp_dbf_out(p, "fsf_issued", "%011lu:%06lu", t.tv_sec, t.tv_nsec); |
354 | zfcp_dbf_out(p, "fsf_prot_status", "0x%08x", r->fsf_prot_status); | 312 | zfcp_dbf_out(p, "fsf_prot_status", "0x%08x", r->fsf_prot_status); |
355 | zfcp_dbf_out(p, "fsf_status", "0x%08x", r->fsf_status); | 313 | zfcp_dbf_out(p, "fsf_status", "0x%08x", r->fsf_status); |
@@ -403,8 +361,8 @@ static void zfcp_hba_dbf_view_response(char **p, | |||
403 | } | 361 | } |
404 | } | 362 | } |
405 | 363 | ||
406 | static void zfcp_hba_dbf_view_status(char **p, | 364 | static void zfcp_dbf_hba_view_status(char **p, |
407 | struct zfcp_hba_dbf_record_status *r) | 365 | struct zfcp_dbf_hba_record_status *r) |
408 | { | 366 | { |
409 | zfcp_dbf_out(p, "failed", "0x%02x", r->failed); | 367 | zfcp_dbf_out(p, "failed", "0x%02x", r->failed); |
410 | zfcp_dbf_out(p, "status_type", "0x%08x", r->status_type); | 368 | zfcp_dbf_out(p, "status_type", "0x%08x", r->status_type); |
@@ -416,14 +374,14 @@ static void zfcp_hba_dbf_view_status(char **p, | |||
416 | r->payload_size); | 374 | r->payload_size); |
417 | } | 375 | } |
418 | 376 | ||
419 | static void zfcp_hba_dbf_view_qdio(char **p, struct zfcp_hba_dbf_record_qdio *r) | 377 | static void zfcp_dbf_hba_view_qdio(char **p, struct zfcp_dbf_hba_record_qdio *r) |
420 | { | 378 | { |
421 | zfcp_dbf_out(p, "qdio_error", "0x%08x", r->qdio_error); | 379 | zfcp_dbf_out(p, "qdio_error", "0x%08x", r->qdio_error); |
422 | zfcp_dbf_out(p, "sbal_index", "0x%02x", r->sbal_index); | 380 | zfcp_dbf_out(p, "sbal_index", "0x%02x", r->sbal_index); |
423 | zfcp_dbf_out(p, "sbal_count", "0x%02x", r->sbal_count); | 381 | zfcp_dbf_out(p, "sbal_count", "0x%02x", r->sbal_count); |
424 | } | 382 | } |
425 | 383 | ||
426 | static void zfcp_hba_dbf_view_berr(char **p, struct fsf_bit_error_payload *r) | 384 | static void zfcp_dbf_hba_view_berr(char **p, struct fsf_bit_error_payload *r) |
427 | { | 385 | { |
428 | zfcp_dbf_out(p, "link_failures", "%d", r->link_failure_error_count); | 386 | zfcp_dbf_out(p, "link_failures", "%d", r->link_failure_error_count); |
429 | zfcp_dbf_out(p, "loss_of_sync_err", "%d", r->loss_of_sync_error_count); | 387 | zfcp_dbf_out(p, "loss_of_sync_err", "%d", r->loss_of_sync_error_count); |
@@ -447,10 +405,10 @@ static void zfcp_hba_dbf_view_berr(char **p, struct fsf_bit_error_payload *r) | |||
447 | r->current_transmit_b2b_credit); | 405 | r->current_transmit_b2b_credit); |
448 | } | 406 | } |
449 | 407 | ||
450 | static int zfcp_hba_dbf_view_format(debug_info_t *id, struct debug_view *view, | 408 | static int zfcp_dbf_hba_view_format(debug_info_t *id, struct debug_view *view, |
451 | char *out_buf, const char *in_buf) | 409 | char *out_buf, const char *in_buf) |
452 | { | 410 | { |
453 | struct zfcp_hba_dbf_record *r = (struct zfcp_hba_dbf_record *)in_buf; | 411 | struct zfcp_dbf_hba_record *r = (struct zfcp_dbf_hba_record *)in_buf; |
454 | char *p = out_buf; | 412 | char *p = out_buf; |
455 | 413 | ||
456 | if (strncmp(r->tag, "dump", ZFCP_DBF_TAG_SIZE) == 0) | 414 | if (strncmp(r->tag, "dump", ZFCP_DBF_TAG_SIZE) == 0) |
@@ -461,45 +419,42 @@ static int zfcp_hba_dbf_view_format(debug_info_t *id, struct debug_view *view, | |||
461 | zfcp_dbf_tag(&p, "tag2", r->tag2); | 419 | zfcp_dbf_tag(&p, "tag2", r->tag2); |
462 | 420 | ||
463 | if (strncmp(r->tag, "resp", ZFCP_DBF_TAG_SIZE) == 0) | 421 | if (strncmp(r->tag, "resp", ZFCP_DBF_TAG_SIZE) == 0) |
464 | zfcp_hba_dbf_view_response(&p, &r->u.response); | 422 | zfcp_dbf_hba_view_response(&p, &r->u.response); |
465 | else if (strncmp(r->tag, "stat", ZFCP_DBF_TAG_SIZE) == 0) | 423 | else if (strncmp(r->tag, "stat", ZFCP_DBF_TAG_SIZE) == 0) |
466 | zfcp_hba_dbf_view_status(&p, &r->u.status); | 424 | zfcp_dbf_hba_view_status(&p, &r->u.status); |
467 | else if (strncmp(r->tag, "qdio", ZFCP_DBF_TAG_SIZE) == 0) | 425 | else if (strncmp(r->tag, "qdio", ZFCP_DBF_TAG_SIZE) == 0) |
468 | zfcp_hba_dbf_view_qdio(&p, &r->u.qdio); | 426 | zfcp_dbf_hba_view_qdio(&p, &r->u.qdio); |
469 | else if (strncmp(r->tag, "berr", ZFCP_DBF_TAG_SIZE) == 0) | 427 | else if (strncmp(r->tag, "berr", ZFCP_DBF_TAG_SIZE) == 0) |
470 | zfcp_hba_dbf_view_berr(&p, &r->u.berr); | 428 | zfcp_dbf_hba_view_berr(&p, &r->u.berr); |
471 | 429 | ||
472 | if (strncmp(r->tag, "resp", ZFCP_DBF_TAG_SIZE) != 0) | 430 | if (strncmp(r->tag, "resp", ZFCP_DBF_TAG_SIZE) != 0) |
473 | p += sprintf(p, "\n"); | 431 | p += sprintf(p, "\n"); |
474 | return p - out_buf; | 432 | return p - out_buf; |
475 | } | 433 | } |
476 | 434 | ||
477 | static struct debug_view zfcp_hba_dbf_view = { | 435 | static struct debug_view zfcp_dbf_hba_view = { |
478 | "structured", | 436 | .name = "structured", |
479 | NULL, | 437 | .header_proc = zfcp_dbf_view_header, |
480 | &zfcp_dbf_view_header, | 438 | .format_proc = zfcp_dbf_hba_view_format, |
481 | &zfcp_hba_dbf_view_format, | ||
482 | NULL, | ||
483 | NULL | ||
484 | }; | 439 | }; |
485 | 440 | ||
486 | static const char *zfcp_rec_dbf_tags[] = { | 441 | static const char *zfcp_dbf_rec_tags[] = { |
487 | [ZFCP_REC_DBF_ID_THREAD] = "thread", | 442 | [ZFCP_REC_DBF_ID_THREAD] = "thread", |
488 | [ZFCP_REC_DBF_ID_TARGET] = "target", | 443 | [ZFCP_REC_DBF_ID_TARGET] = "target", |
489 | [ZFCP_REC_DBF_ID_TRIGGER] = "trigger", | 444 | [ZFCP_REC_DBF_ID_TRIGGER] = "trigger", |
490 | [ZFCP_REC_DBF_ID_ACTION] = "action", | 445 | [ZFCP_REC_DBF_ID_ACTION] = "action", |
491 | }; | 446 | }; |
492 | 447 | ||
493 | static int zfcp_rec_dbf_view_format(debug_info_t *id, struct debug_view *view, | 448 | static int zfcp_dbf_rec_view_format(debug_info_t *id, struct debug_view *view, |
494 | char *buf, const char *_rec) | 449 | char *buf, const char *_rec) |
495 | { | 450 | { |
496 | struct zfcp_rec_dbf_record *r = (struct zfcp_rec_dbf_record *)_rec; | 451 | struct zfcp_dbf_rec_record *r = (struct zfcp_dbf_rec_record *)_rec; |
497 | char *p = buf; | 452 | char *p = buf; |
498 | char hint[ZFCP_DBF_ID_SIZE + 1]; | 453 | char hint[ZFCP_DBF_ID_SIZE + 1]; |
499 | 454 | ||
500 | memcpy(hint, r->id2, ZFCP_DBF_ID_SIZE); | 455 | memcpy(hint, r->id2, ZFCP_DBF_ID_SIZE); |
501 | hint[ZFCP_DBF_ID_SIZE] = 0; | 456 | hint[ZFCP_DBF_ID_SIZE] = 0; |
502 | zfcp_dbf_outs(&p, "tag", zfcp_rec_dbf_tags[r->id]); | 457 | zfcp_dbf_outs(&p, "tag", zfcp_dbf_rec_tags[r->id]); |
503 | zfcp_dbf_outs(&p, "hint", hint); | 458 | zfcp_dbf_outs(&p, "hint", hint); |
504 | switch (r->id) { | 459 | switch (r->id) { |
505 | case ZFCP_REC_DBF_ID_THREAD: | 460 | case ZFCP_REC_DBF_ID_THREAD: |
@@ -537,24 +492,22 @@ static int zfcp_rec_dbf_view_format(debug_info_t *id, struct debug_view *view, | |||
537 | return p - buf; | 492 | return p - buf; |
538 | } | 493 | } |
539 | 494 | ||
540 | static struct debug_view zfcp_rec_dbf_view = { | 495 | static struct debug_view zfcp_dbf_rec_view = { |
541 | "structured", | 496 | .name = "structured", |
542 | NULL, | 497 | .header_proc = zfcp_dbf_view_header, |
543 | &zfcp_dbf_view_header, | 498 | .format_proc = zfcp_dbf_rec_view_format, |
544 | &zfcp_rec_dbf_view_format, | ||
545 | NULL, | ||
546 | NULL | ||
547 | }; | 499 | }; |
548 | 500 | ||
549 | /** | 501 | /** |
550 | * zfcp_rec_dbf_event_thread - trace event related to recovery thread operation | 502 | * zfcp_dbf_rec_thread - trace event related to recovery thread operation |
551 | * @id2: identifier for event | 503 | * @id2: identifier for event |
552 | * @adapter: adapter | 504 | * @dbf: reference to dbf structure |
553 | * This function assumes that the caller is holding erp_lock. | 505 | * This function assumes that the caller is holding erp_lock. |
554 | */ | 506 | */ |
555 | void zfcp_rec_dbf_event_thread(char *id2, struct zfcp_adapter *adapter) | 507 | void zfcp_dbf_rec_thread(char *id2, struct zfcp_dbf *dbf) |
556 | { | 508 | { |
557 | struct zfcp_rec_dbf_record *r = &adapter->rec_dbf_buf; | 509 | struct zfcp_adapter *adapter = dbf->adapter; |
510 | struct zfcp_dbf_rec_record *r = &dbf->rec_buf; | ||
558 | unsigned long flags = 0; | 511 | unsigned long flags = 0; |
559 | struct list_head *entry; | 512 | struct list_head *entry; |
560 | unsigned ready = 0, running = 0, total; | 513 | unsigned ready = 0, running = 0, total; |
@@ -565,41 +518,41 @@ void zfcp_rec_dbf_event_thread(char *id2, struct zfcp_adapter *adapter) | |||
565 | running++; | 518 | running++; |
566 | total = adapter->erp_total_count; | 519 | total = adapter->erp_total_count; |
567 | 520 | ||
568 | spin_lock_irqsave(&adapter->rec_dbf_lock, flags); | 521 | spin_lock_irqsave(&dbf->rec_lock, flags); |
569 | memset(r, 0, sizeof(*r)); | 522 | memset(r, 0, sizeof(*r)); |
570 | r->id = ZFCP_REC_DBF_ID_THREAD; | 523 | r->id = ZFCP_REC_DBF_ID_THREAD; |
571 | memcpy(r->id2, id2, ZFCP_DBF_ID_SIZE); | 524 | memcpy(r->id2, id2, ZFCP_DBF_ID_SIZE); |
572 | r->u.thread.total = total; | 525 | r->u.thread.total = total; |
573 | r->u.thread.ready = ready; | 526 | r->u.thread.ready = ready; |
574 | r->u.thread.running = running; | 527 | r->u.thread.running = running; |
575 | debug_event(adapter->rec_dbf, 6, r, sizeof(*r)); | 528 | debug_event(dbf->rec, 6, r, sizeof(*r)); |
576 | spin_unlock_irqrestore(&adapter->rec_dbf_lock, flags); | 529 | spin_unlock_irqrestore(&dbf->rec_lock, flags); |
577 | } | 530 | } |
578 | 531 | ||
579 | /** | 532 | /** |
580 | * zfcp_rec_dbf_event_thread - trace event related to recovery thread operation | 533 | * zfcp_dbf_rec_thread - trace event related to recovery thread operation |
581 | * @id2: identifier for event | 534 | * @id2: identifier for event |
582 | * @adapter: adapter | 535 | * @adapter: adapter |
583 | * This function assumes that the caller does not hold erp_lock. | 536 | * This function assumes that the caller does not hold erp_lock. |
584 | */ | 537 | */ |
585 | void zfcp_rec_dbf_event_thread_lock(char *id2, struct zfcp_adapter *adapter) | 538 | void zfcp_dbf_rec_thread_lock(char *id2, struct zfcp_dbf *dbf) |
586 | { | 539 | { |
540 | struct zfcp_adapter *adapter = dbf->adapter; | ||
587 | unsigned long flags; | 541 | unsigned long flags; |
588 | 542 | ||
589 | read_lock_irqsave(&adapter->erp_lock, flags); | 543 | read_lock_irqsave(&adapter->erp_lock, flags); |
590 | zfcp_rec_dbf_event_thread(id2, adapter); | 544 | zfcp_dbf_rec_thread(id2, dbf); |
591 | read_unlock_irqrestore(&adapter->erp_lock, flags); | 545 | read_unlock_irqrestore(&adapter->erp_lock, flags); |
592 | } | 546 | } |
593 | 547 | ||
594 | static void zfcp_rec_dbf_event_target(char *id2, void *ref, | 548 | static void zfcp_dbf_rec_target(char *id2, void *ref, struct zfcp_dbf *dbf, |
595 | struct zfcp_adapter *adapter, | 549 | atomic_t *status, atomic_t *erp_count, u64 wwpn, |
596 | atomic_t *status, atomic_t *erp_count, | 550 | u32 d_id, u64 fcp_lun) |
597 | u64 wwpn, u32 d_id, u64 fcp_lun) | ||
598 | { | 551 | { |
599 | struct zfcp_rec_dbf_record *r = &adapter->rec_dbf_buf; | 552 | struct zfcp_dbf_rec_record *r = &dbf->rec_buf; |
600 | unsigned long flags; | 553 | unsigned long flags; |
601 | 554 | ||
602 | spin_lock_irqsave(&adapter->rec_dbf_lock, flags); | 555 | spin_lock_irqsave(&dbf->rec_lock, flags); |
603 | memset(r, 0, sizeof(*r)); | 556 | memset(r, 0, sizeof(*r)); |
604 | r->id = ZFCP_REC_DBF_ID_TARGET; | 557 | r->id = ZFCP_REC_DBF_ID_TARGET; |
605 | memcpy(r->id2, id2, ZFCP_DBF_ID_SIZE); | 558 | memcpy(r->id2, id2, ZFCP_DBF_ID_SIZE); |
@@ -609,56 +562,57 @@ static void zfcp_rec_dbf_event_target(char *id2, void *ref, | |||
609 | r->u.target.d_id = d_id; | 562 | r->u.target.d_id = d_id; |
610 | r->u.target.fcp_lun = fcp_lun; | 563 | r->u.target.fcp_lun = fcp_lun; |
611 | r->u.target.erp_count = atomic_read(erp_count); | 564 | r->u.target.erp_count = atomic_read(erp_count); |
612 | debug_event(adapter->rec_dbf, 3, r, sizeof(*r)); | 565 | debug_event(dbf->rec, 3, r, sizeof(*r)); |
613 | spin_unlock_irqrestore(&adapter->rec_dbf_lock, flags); | 566 | spin_unlock_irqrestore(&dbf->rec_lock, flags); |
614 | } | 567 | } |
615 | 568 | ||
616 | /** | 569 | /** |
617 | * zfcp_rec_dbf_event_adapter - trace event for adapter state change | 570 | * zfcp_dbf_rec_adapter - trace event for adapter state change |
618 | * @id: identifier for trigger of state change | 571 | * @id: identifier for trigger of state change |
619 | * @ref: additional reference (e.g. request) | 572 | * @ref: additional reference (e.g. request) |
620 | * @adapter: adapter | 573 | * @dbf: reference to dbf structure |
621 | */ | 574 | */ |
622 | void zfcp_rec_dbf_event_adapter(char *id, void *ref, | 575 | void zfcp_dbf_rec_adapter(char *id, void *ref, struct zfcp_dbf *dbf) |
623 | struct zfcp_adapter *adapter) | ||
624 | { | 576 | { |
625 | zfcp_rec_dbf_event_target(id, ref, adapter, &adapter->status, | 577 | struct zfcp_adapter *adapter = dbf->adapter; |
578 | |||
579 | zfcp_dbf_rec_target(id, ref, dbf, &adapter->status, | ||
626 | &adapter->erp_counter, 0, 0, 0); | 580 | &adapter->erp_counter, 0, 0, 0); |
627 | } | 581 | } |
628 | 582 | ||
629 | /** | 583 | /** |
630 | * zfcp_rec_dbf_event_port - trace event for port state change | 584 | * zfcp_dbf_rec_port - trace event for port state change |
631 | * @id: identifier for trigger of state change | 585 | * @id: identifier for trigger of state change |
632 | * @ref: additional reference (e.g. request) | 586 | * @ref: additional reference (e.g. request) |
633 | * @port: port | 587 | * @port: port |
634 | */ | 588 | */ |
635 | void zfcp_rec_dbf_event_port(char *id, void *ref, struct zfcp_port *port) | 589 | void zfcp_dbf_rec_port(char *id, void *ref, struct zfcp_port *port) |
636 | { | 590 | { |
637 | struct zfcp_adapter *adapter = port->adapter; | 591 | struct zfcp_dbf *dbf = port->adapter->dbf; |
638 | 592 | ||
639 | zfcp_rec_dbf_event_target(id, ref, adapter, &port->status, | 593 | zfcp_dbf_rec_target(id, ref, dbf, &port->status, |
640 | &port->erp_counter, port->wwpn, port->d_id, | 594 | &port->erp_counter, port->wwpn, port->d_id, |
641 | 0); | 595 | 0); |
642 | } | 596 | } |
643 | 597 | ||
644 | /** | 598 | /** |
645 | * zfcp_rec_dbf_event_unit - trace event for unit state change | 599 | * zfcp_dbf_rec_unit - trace event for unit state change |
646 | * @id: identifier for trigger of state change | 600 | * @id: identifier for trigger of state change |
647 | * @ref: additional reference (e.g. request) | 601 | * @ref: additional reference (e.g. request) |
648 | * @unit: unit | 602 | * @unit: unit |
649 | */ | 603 | */ |
650 | void zfcp_rec_dbf_event_unit(char *id, void *ref, struct zfcp_unit *unit) | 604 | void zfcp_dbf_rec_unit(char *id, void *ref, struct zfcp_unit *unit) |
651 | { | 605 | { |
652 | struct zfcp_port *port = unit->port; | 606 | struct zfcp_port *port = unit->port; |
653 | struct zfcp_adapter *adapter = port->adapter; | 607 | struct zfcp_dbf *dbf = port->adapter->dbf; |
654 | 608 | ||
655 | zfcp_rec_dbf_event_target(id, ref, adapter, &unit->status, | 609 | zfcp_dbf_rec_target(id, ref, dbf, &unit->status, |
656 | &unit->erp_counter, port->wwpn, port->d_id, | 610 | &unit->erp_counter, port->wwpn, port->d_id, |
657 | unit->fcp_lun); | 611 | unit->fcp_lun); |
658 | } | 612 | } |
659 | 613 | ||
660 | /** | 614 | /** |
661 | * zfcp_rec_dbf_event_trigger - trace event for triggered error recovery | 615 | * zfcp_dbf_rec_trigger - trace event for triggered error recovery |
662 | * @id2: identifier for error recovery trigger | 616 | * @id2: identifier for error recovery trigger |
663 | * @ref: additional reference (e.g. request) | 617 | * @ref: additional reference (e.g. request) |
664 | * @want: originally requested error recovery action | 618 | * @want: originally requested error recovery action |
@@ -668,14 +622,15 @@ void zfcp_rec_dbf_event_unit(char *id, void *ref, struct zfcp_unit *unit) | |||
668 | * @port: port | 622 | * @port: port |
669 | * @unit: unit | 623 | * @unit: unit |
670 | */ | 624 | */ |
671 | void zfcp_rec_dbf_event_trigger(char *id2, void *ref, u8 want, u8 need, | 625 | void zfcp_dbf_rec_trigger(char *id2, void *ref, u8 want, u8 need, void *action, |
672 | void *action, struct zfcp_adapter *adapter, | 626 | struct zfcp_adapter *adapter, struct zfcp_port *port, |
673 | struct zfcp_port *port, struct zfcp_unit *unit) | 627 | struct zfcp_unit *unit) |
674 | { | 628 | { |
675 | struct zfcp_rec_dbf_record *r = &adapter->rec_dbf_buf; | 629 | struct zfcp_dbf *dbf = adapter->dbf; |
630 | struct zfcp_dbf_rec_record *r = &dbf->rec_buf; | ||
676 | unsigned long flags; | 631 | unsigned long flags; |
677 | 632 | ||
678 | spin_lock_irqsave(&adapter->rec_dbf_lock, flags); | 633 | spin_lock_irqsave(&dbf->rec_lock, flags); |
679 | memset(r, 0, sizeof(*r)); | 634 | memset(r, 0, sizeof(*r)); |
680 | r->id = ZFCP_REC_DBF_ID_TRIGGER; | 635 | r->id = ZFCP_REC_DBF_ID_TRIGGER; |
681 | memcpy(r->id2, id2, ZFCP_DBF_ID_SIZE); | 636 | memcpy(r->id2, id2, ZFCP_DBF_ID_SIZE); |
@@ -692,22 +647,22 @@ void zfcp_rec_dbf_event_trigger(char *id2, void *ref, u8 want, u8 need, | |||
692 | r->u.trigger.us = atomic_read(&unit->status); | 647 | r->u.trigger.us = atomic_read(&unit->status); |
693 | r->u.trigger.fcp_lun = unit->fcp_lun; | 648 | r->u.trigger.fcp_lun = unit->fcp_lun; |
694 | } | 649 | } |
695 | debug_event(adapter->rec_dbf, action ? 1 : 4, r, sizeof(*r)); | 650 | debug_event(dbf->rec, action ? 1 : 4, r, sizeof(*r)); |
696 | spin_unlock_irqrestore(&adapter->rec_dbf_lock, flags); | 651 | spin_unlock_irqrestore(&dbf->rec_lock, flags); |
697 | } | 652 | } |
698 | 653 | ||
699 | /** | 654 | /** |
700 | * zfcp_rec_dbf_event_action - trace event showing progress of recovery action | 655 | * zfcp_dbf_rec_action - trace event showing progress of recovery action |
701 | * @id2: identifier | 656 | * @id2: identifier |
702 | * @erp_action: error recovery action struct pointer | 657 | * @erp_action: error recovery action struct pointer |
703 | */ | 658 | */ |
704 | void zfcp_rec_dbf_event_action(char *id2, struct zfcp_erp_action *erp_action) | 659 | void zfcp_dbf_rec_action(char *id2, struct zfcp_erp_action *erp_action) |
705 | { | 660 | { |
706 | struct zfcp_adapter *adapter = erp_action->adapter; | 661 | struct zfcp_dbf *dbf = erp_action->adapter->dbf; |
707 | struct zfcp_rec_dbf_record *r = &adapter->rec_dbf_buf; | 662 | struct zfcp_dbf_rec_record *r = &dbf->rec_buf; |
708 | unsigned long flags; | 663 | unsigned long flags; |
709 | 664 | ||
710 | spin_lock_irqsave(&adapter->rec_dbf_lock, flags); | 665 | spin_lock_irqsave(&dbf->rec_lock, flags); |
711 | memset(r, 0, sizeof(*r)); | 666 | memset(r, 0, sizeof(*r)); |
712 | r->id = ZFCP_REC_DBF_ID_ACTION; | 667 | r->id = ZFCP_REC_DBF_ID_ACTION; |
713 | memcpy(r->id2, id2, ZFCP_DBF_ID_SIZE); | 668 | memcpy(r->id2, id2, ZFCP_DBF_ID_SIZE); |
@@ -715,26 +670,27 @@ void zfcp_rec_dbf_event_action(char *id2, struct zfcp_erp_action *erp_action) | |||
715 | r->u.action.status = erp_action->status; | 670 | r->u.action.status = erp_action->status; |
716 | r->u.action.step = erp_action->step; | 671 | r->u.action.step = erp_action->step; |
717 | r->u.action.fsf_req = (unsigned long)erp_action->fsf_req; | 672 | r->u.action.fsf_req = (unsigned long)erp_action->fsf_req; |
718 | debug_event(adapter->rec_dbf, 5, r, sizeof(*r)); | 673 | debug_event(dbf->rec, 5, r, sizeof(*r)); |
719 | spin_unlock_irqrestore(&adapter->rec_dbf_lock, flags); | 674 | spin_unlock_irqrestore(&dbf->rec_lock, flags); |
720 | } | 675 | } |
721 | 676 | ||
722 | /** | 677 | /** |
723 | * zfcp_san_dbf_event_ct_request - trace event for issued CT request | 678 | * zfcp_dbf_san_ct_request - trace event for issued CT request |
724 | * @fsf_req: request containing issued CT data | 679 | * @fsf_req: request containing issued CT data |
725 | */ | 680 | */ |
726 | void zfcp_san_dbf_event_ct_request(struct zfcp_fsf_req *fsf_req) | 681 | void zfcp_dbf_san_ct_request(struct zfcp_fsf_req *fsf_req) |
727 | { | 682 | { |
728 | struct zfcp_send_ct *ct = (struct zfcp_send_ct *)fsf_req->data; | 683 | struct zfcp_send_ct *ct = (struct zfcp_send_ct *)fsf_req->data; |
729 | struct zfcp_wka_port *wka_port = ct->wka_port; | 684 | struct zfcp_wka_port *wka_port = ct->wka_port; |
730 | struct zfcp_adapter *adapter = wka_port->adapter; | 685 | struct zfcp_adapter *adapter = wka_port->adapter; |
686 | struct zfcp_dbf *dbf = adapter->dbf; | ||
731 | struct ct_hdr *hdr = sg_virt(ct->req); | 687 | struct ct_hdr *hdr = sg_virt(ct->req); |
732 | struct zfcp_san_dbf_record *r = &adapter->san_dbf_buf; | 688 | struct zfcp_dbf_san_record *r = &dbf->san_buf; |
733 | struct zfcp_san_dbf_record_ct_request *oct = &r->u.ct_req; | 689 | struct zfcp_dbf_san_record_ct_request *oct = &r->u.ct_req; |
734 | int level = 3; | 690 | int level = 3; |
735 | unsigned long flags; | 691 | unsigned long flags; |
736 | 692 | ||
737 | spin_lock_irqsave(&adapter->san_dbf_lock, flags); | 693 | spin_lock_irqsave(&dbf->san_lock, flags); |
738 | memset(r, 0, sizeof(*r)); | 694 | memset(r, 0, sizeof(*r)); |
739 | strncpy(r->tag, "octc", ZFCP_DBF_TAG_SIZE); | 695 | strncpy(r->tag, "octc", ZFCP_DBF_TAG_SIZE); |
740 | r->fsf_reqid = fsf_req->req_id; | 696 | r->fsf_reqid = fsf_req->req_id; |
@@ -749,28 +705,29 @@ void zfcp_san_dbf_event_ct_request(struct zfcp_fsf_req *fsf_req) | |||
749 | oct->max_res_size = hdr->max_res_size; | 705 | oct->max_res_size = hdr->max_res_size; |
750 | oct->len = min((int)ct->req->length - (int)sizeof(struct ct_hdr), | 706 | oct->len = min((int)ct->req->length - (int)sizeof(struct ct_hdr), |
751 | ZFCP_DBF_SAN_MAX_PAYLOAD); | 707 | ZFCP_DBF_SAN_MAX_PAYLOAD); |
752 | debug_event(adapter->san_dbf, level, r, sizeof(*r)); | 708 | debug_event(dbf->san, level, r, sizeof(*r)); |
753 | zfcp_dbf_hexdump(adapter->san_dbf, r, sizeof(*r), level, | 709 | zfcp_dbf_hexdump(dbf->san, r, sizeof(*r), level, |
754 | (void *)hdr + sizeof(struct ct_hdr), oct->len); | 710 | (void *)hdr + sizeof(struct ct_hdr), oct->len); |
755 | spin_unlock_irqrestore(&adapter->san_dbf_lock, flags); | 711 | spin_unlock_irqrestore(&dbf->san_lock, flags); |
756 | } | 712 | } |
757 | 713 | ||
758 | /** | 714 | /** |
759 | * zfcp_san_dbf_event_ct_response - trace event for completion of CT request | 715 | * zfcp_dbf_san_ct_response - trace event for completion of CT request |
760 | * @fsf_req: request containing CT response | 716 | * @fsf_req: request containing CT response |
761 | */ | 717 | */ |
762 | void zfcp_san_dbf_event_ct_response(struct zfcp_fsf_req *fsf_req) | 718 | void zfcp_dbf_san_ct_response(struct zfcp_fsf_req *fsf_req) |
763 | { | 719 | { |
764 | struct zfcp_send_ct *ct = (struct zfcp_send_ct *)fsf_req->data; | 720 | struct zfcp_send_ct *ct = (struct zfcp_send_ct *)fsf_req->data; |
765 | struct zfcp_wka_port *wka_port = ct->wka_port; | 721 | struct zfcp_wka_port *wka_port = ct->wka_port; |
766 | struct zfcp_adapter *adapter = wka_port->adapter; | 722 | struct zfcp_adapter *adapter = wka_port->adapter; |
767 | struct ct_hdr *hdr = sg_virt(ct->resp); | 723 | struct ct_hdr *hdr = sg_virt(ct->resp); |
768 | struct zfcp_san_dbf_record *r = &adapter->san_dbf_buf; | 724 | struct zfcp_dbf *dbf = adapter->dbf; |
769 | struct zfcp_san_dbf_record_ct_response *rct = &r->u.ct_resp; | 725 | struct zfcp_dbf_san_record *r = &dbf->san_buf; |
726 | struct zfcp_dbf_san_record_ct_response *rct = &r->u.ct_resp; | ||
770 | int level = 3; | 727 | int level = 3; |
771 | unsigned long flags; | 728 | unsigned long flags; |
772 | 729 | ||
773 | spin_lock_irqsave(&adapter->san_dbf_lock, flags); | 730 | spin_lock_irqsave(&dbf->san_lock, flags); |
774 | memset(r, 0, sizeof(*r)); | 731 | memset(r, 0, sizeof(*r)); |
775 | strncpy(r->tag, "rctc", ZFCP_DBF_TAG_SIZE); | 732 | strncpy(r->tag, "rctc", ZFCP_DBF_TAG_SIZE); |
776 | r->fsf_reqid = fsf_req->req_id; | 733 | r->fsf_reqid = fsf_req->req_id; |
@@ -785,22 +742,22 @@ void zfcp_san_dbf_event_ct_response(struct zfcp_fsf_req *fsf_req) | |||
785 | rct->max_res_size = hdr->max_res_size; | 742 | rct->max_res_size = hdr->max_res_size; |
786 | rct->len = min((int)ct->resp->length - (int)sizeof(struct ct_hdr), | 743 | rct->len = min((int)ct->resp->length - (int)sizeof(struct ct_hdr), |
787 | ZFCP_DBF_SAN_MAX_PAYLOAD); | 744 | ZFCP_DBF_SAN_MAX_PAYLOAD); |
788 | debug_event(adapter->san_dbf, level, r, sizeof(*r)); | 745 | debug_event(dbf->san, level, r, sizeof(*r)); |
789 | zfcp_dbf_hexdump(adapter->san_dbf, r, sizeof(*r), level, | 746 | zfcp_dbf_hexdump(dbf->san, r, sizeof(*r), level, |
790 | (void *)hdr + sizeof(struct ct_hdr), rct->len); | 747 | (void *)hdr + sizeof(struct ct_hdr), rct->len); |
791 | spin_unlock_irqrestore(&adapter->san_dbf_lock, flags); | 748 | spin_unlock_irqrestore(&dbf->san_lock, flags); |
792 | } | 749 | } |
793 | 750 | ||
794 | static void zfcp_san_dbf_event_els(const char *tag, int level, | 751 | static void zfcp_dbf_san_els(const char *tag, int level, |
795 | struct zfcp_fsf_req *fsf_req, u32 s_id, | 752 | struct zfcp_fsf_req *fsf_req, u32 s_id, u32 d_id, |
796 | u32 d_id, u8 ls_code, void *buffer, | 753 | u8 ls_code, void *buffer, int buflen) |
797 | int buflen) | ||
798 | { | 754 | { |
799 | struct zfcp_adapter *adapter = fsf_req->adapter; | 755 | struct zfcp_adapter *adapter = fsf_req->adapter; |
800 | struct zfcp_san_dbf_record *rec = &adapter->san_dbf_buf; | 756 | struct zfcp_dbf *dbf = adapter->dbf; |
757 | struct zfcp_dbf_san_record *rec = &dbf->san_buf; | ||
801 | unsigned long flags; | 758 | unsigned long flags; |
802 | 759 | ||
803 | spin_lock_irqsave(&adapter->san_dbf_lock, flags); | 760 | spin_lock_irqsave(&dbf->san_lock, flags); |
804 | memset(rec, 0, sizeof(*rec)); | 761 | memset(rec, 0, sizeof(*rec)); |
805 | strncpy(rec->tag, tag, ZFCP_DBF_TAG_SIZE); | 762 | strncpy(rec->tag, tag, ZFCP_DBF_TAG_SIZE); |
806 | rec->fsf_reqid = fsf_req->req_id; | 763 | rec->fsf_reqid = fsf_req->req_id; |
@@ -808,45 +765,45 @@ static void zfcp_san_dbf_event_els(const char *tag, int level, | |||
808 | rec->s_id = s_id; | 765 | rec->s_id = s_id; |
809 | rec->d_id = d_id; | 766 | rec->d_id = d_id; |
810 | rec->u.els.ls_code = ls_code; | 767 | rec->u.els.ls_code = ls_code; |
811 | debug_event(adapter->san_dbf, level, rec, sizeof(*rec)); | 768 | debug_event(dbf->san, level, rec, sizeof(*rec)); |
812 | zfcp_dbf_hexdump(adapter->san_dbf, rec, sizeof(*rec), level, | 769 | zfcp_dbf_hexdump(dbf->san, rec, sizeof(*rec), level, |
813 | buffer, min(buflen, ZFCP_DBF_SAN_MAX_PAYLOAD)); | 770 | buffer, min(buflen, ZFCP_DBF_SAN_MAX_PAYLOAD)); |
814 | spin_unlock_irqrestore(&adapter->san_dbf_lock, flags); | 771 | spin_unlock_irqrestore(&dbf->san_lock, flags); |
815 | } | 772 | } |
816 | 773 | ||
817 | /** | 774 | /** |
818 | * zfcp_san_dbf_event_els_request - trace event for issued ELS | 775 | * zfcp_dbf_san_els_request - trace event for issued ELS |
819 | * @fsf_req: request containing issued ELS | 776 | * @fsf_req: request containing issued ELS |
820 | */ | 777 | */ |
821 | void zfcp_san_dbf_event_els_request(struct zfcp_fsf_req *fsf_req) | 778 | void zfcp_dbf_san_els_request(struct zfcp_fsf_req *fsf_req) |
822 | { | 779 | { |
823 | struct zfcp_send_els *els = (struct zfcp_send_els *)fsf_req->data; | 780 | struct zfcp_send_els *els = (struct zfcp_send_els *)fsf_req->data; |
824 | 781 | ||
825 | zfcp_san_dbf_event_els("oels", 2, fsf_req, | 782 | zfcp_dbf_san_els("oels", 2, fsf_req, |
826 | fc_host_port_id(els->adapter->scsi_host), | 783 | fc_host_port_id(els->adapter->scsi_host), |
827 | els->d_id, *(u8 *) sg_virt(els->req), | 784 | els->d_id, *(u8 *) sg_virt(els->req), |
828 | sg_virt(els->req), els->req->length); | 785 | sg_virt(els->req), els->req->length); |
829 | } | 786 | } |
830 | 787 | ||
831 | /** | 788 | /** |
832 | * zfcp_san_dbf_event_els_response - trace event for completed ELS | 789 | * zfcp_dbf_san_els_response - trace event for completed ELS |
833 | * @fsf_req: request containing ELS response | 790 | * @fsf_req: request containing ELS response |
834 | */ | 791 | */ |
835 | void zfcp_san_dbf_event_els_response(struct zfcp_fsf_req *fsf_req) | 792 | void zfcp_dbf_san_els_response(struct zfcp_fsf_req *fsf_req) |
836 | { | 793 | { |
837 | struct zfcp_send_els *els = (struct zfcp_send_els *)fsf_req->data; | 794 | struct zfcp_send_els *els = (struct zfcp_send_els *)fsf_req->data; |
838 | 795 | ||
839 | zfcp_san_dbf_event_els("rels", 2, fsf_req, els->d_id, | 796 | zfcp_dbf_san_els("rels", 2, fsf_req, els->d_id, |
840 | fc_host_port_id(els->adapter->scsi_host), | 797 | fc_host_port_id(els->adapter->scsi_host), |
841 | *(u8 *)sg_virt(els->req), sg_virt(els->resp), | 798 | *(u8 *)sg_virt(els->req), sg_virt(els->resp), |
842 | els->resp->length); | 799 | els->resp->length); |
843 | } | 800 | } |
844 | 801 | ||
845 | /** | 802 | /** |
846 | * zfcp_san_dbf_event_incoming_els - trace event for incomig ELS | 803 | * zfcp_dbf_san_incoming_els - trace event for incomig ELS |
847 | * @fsf_req: request containing unsolicited status buffer with incoming ELS | 804 | * @fsf_req: request containing unsolicited status buffer with incoming ELS |
848 | */ | 805 | */ |
849 | void zfcp_san_dbf_event_incoming_els(struct zfcp_fsf_req *fsf_req) | 806 | void zfcp_dbf_san_incoming_els(struct zfcp_fsf_req *fsf_req) |
850 | { | 807 | { |
851 | struct zfcp_adapter *adapter = fsf_req->adapter; | 808 | struct zfcp_adapter *adapter = fsf_req->adapter; |
852 | struct fsf_status_read_buffer *buf = | 809 | struct fsf_status_read_buffer *buf = |
@@ -854,16 +811,16 @@ void zfcp_san_dbf_event_incoming_els(struct zfcp_fsf_req *fsf_req) | |||
854 | int length = (int)buf->length - | 811 | int length = (int)buf->length - |
855 | (int)((void *)&buf->payload - (void *)buf); | 812 | (int)((void *)&buf->payload - (void *)buf); |
856 | 813 | ||
857 | zfcp_san_dbf_event_els("iels", 1, fsf_req, buf->d_id, | 814 | zfcp_dbf_san_els("iels", 1, fsf_req, buf->d_id, |
858 | fc_host_port_id(adapter->scsi_host), | 815 | fc_host_port_id(adapter->scsi_host), |
859 | buf->payload.data[0], (void *)buf->payload.data, | 816 | buf->payload.data[0], (void *)buf->payload.data, |
860 | length); | 817 | length); |
861 | } | 818 | } |
862 | 819 | ||
863 | static int zfcp_san_dbf_view_format(debug_info_t *id, struct debug_view *view, | 820 | static int zfcp_dbf_san_view_format(debug_info_t *id, struct debug_view *view, |
864 | char *out_buf, const char *in_buf) | 821 | char *out_buf, const char *in_buf) |
865 | { | 822 | { |
866 | struct zfcp_san_dbf_record *r = (struct zfcp_san_dbf_record *)in_buf; | 823 | struct zfcp_dbf_san_record *r = (struct zfcp_dbf_san_record *)in_buf; |
867 | char *p = out_buf; | 824 | char *p = out_buf; |
868 | 825 | ||
869 | if (strncmp(r->tag, "dump", ZFCP_DBF_TAG_SIZE) == 0) | 826 | if (strncmp(r->tag, "dump", ZFCP_DBF_TAG_SIZE) == 0) |
@@ -876,7 +833,7 @@ static int zfcp_san_dbf_view_format(debug_info_t *id, struct debug_view *view, | |||
876 | zfcp_dbf_out(&p, "d_id", "0x%06x", r->d_id); | 833 | zfcp_dbf_out(&p, "d_id", "0x%06x", r->d_id); |
877 | 834 | ||
878 | if (strncmp(r->tag, "octc", ZFCP_DBF_TAG_SIZE) == 0) { | 835 | if (strncmp(r->tag, "octc", ZFCP_DBF_TAG_SIZE) == 0) { |
879 | struct zfcp_san_dbf_record_ct_request *ct = &r->u.ct_req; | 836 | struct zfcp_dbf_san_record_ct_request *ct = &r->u.ct_req; |
880 | zfcp_dbf_out(&p, "cmd_req_code", "0x%04x", ct->cmd_req_code); | 837 | zfcp_dbf_out(&p, "cmd_req_code", "0x%04x", ct->cmd_req_code); |
881 | zfcp_dbf_out(&p, "revision", "0x%02x", ct->revision); | 838 | zfcp_dbf_out(&p, "revision", "0x%02x", ct->revision); |
882 | zfcp_dbf_out(&p, "gs_type", "0x%02x", ct->gs_type); | 839 | zfcp_dbf_out(&p, "gs_type", "0x%02x", ct->gs_type); |
@@ -884,7 +841,7 @@ static int zfcp_san_dbf_view_format(debug_info_t *id, struct debug_view *view, | |||
884 | zfcp_dbf_out(&p, "options", "0x%02x", ct->options); | 841 | zfcp_dbf_out(&p, "options", "0x%02x", ct->options); |
885 | zfcp_dbf_out(&p, "max_res_size", "0x%04x", ct->max_res_size); | 842 | zfcp_dbf_out(&p, "max_res_size", "0x%04x", ct->max_res_size); |
886 | } else if (strncmp(r->tag, "rctc", ZFCP_DBF_TAG_SIZE) == 0) { | 843 | } else if (strncmp(r->tag, "rctc", ZFCP_DBF_TAG_SIZE) == 0) { |
887 | struct zfcp_san_dbf_record_ct_response *ct = &r->u.ct_resp; | 844 | struct zfcp_dbf_san_record_ct_response *ct = &r->u.ct_resp; |
888 | zfcp_dbf_out(&p, "cmd_rsp_code", "0x%04x", ct->cmd_rsp_code); | 845 | zfcp_dbf_out(&p, "cmd_rsp_code", "0x%04x", ct->cmd_rsp_code); |
889 | zfcp_dbf_out(&p, "revision", "0x%02x", ct->revision); | 846 | zfcp_dbf_out(&p, "revision", "0x%02x", ct->revision); |
890 | zfcp_dbf_out(&p, "reason_code", "0x%02x", ct->reason_code); | 847 | zfcp_dbf_out(&p, "reason_code", "0x%02x", ct->reason_code); |
@@ -894,35 +851,30 @@ static int zfcp_san_dbf_view_format(debug_info_t *id, struct debug_view *view, | |||
894 | } else if (strncmp(r->tag, "oels", ZFCP_DBF_TAG_SIZE) == 0 || | 851 | } else if (strncmp(r->tag, "oels", ZFCP_DBF_TAG_SIZE) == 0 || |
895 | strncmp(r->tag, "rels", ZFCP_DBF_TAG_SIZE) == 0 || | 852 | strncmp(r->tag, "rels", ZFCP_DBF_TAG_SIZE) == 0 || |
896 | strncmp(r->tag, "iels", ZFCP_DBF_TAG_SIZE) == 0) { | 853 | strncmp(r->tag, "iels", ZFCP_DBF_TAG_SIZE) == 0) { |
897 | struct zfcp_san_dbf_record_els *els = &r->u.els; | 854 | struct zfcp_dbf_san_record_els *els = &r->u.els; |
898 | zfcp_dbf_out(&p, "ls_code", "0x%02x", els->ls_code); | 855 | zfcp_dbf_out(&p, "ls_code", "0x%02x", els->ls_code); |
899 | } | 856 | } |
900 | return p - out_buf; | 857 | return p - out_buf; |
901 | } | 858 | } |
902 | 859 | ||
903 | static struct debug_view zfcp_san_dbf_view = { | 860 | static struct debug_view zfcp_dbf_san_view = { |
904 | "structured", | 861 | .name = "structured", |
905 | NULL, | 862 | .header_proc = zfcp_dbf_view_header, |
906 | &zfcp_dbf_view_header, | 863 | .format_proc = zfcp_dbf_san_view_format, |
907 | &zfcp_san_dbf_view_format, | ||
908 | NULL, | ||
909 | NULL | ||
910 | }; | 864 | }; |
911 | 865 | ||
912 | static void zfcp_scsi_dbf_event(const char *tag, const char *tag2, int level, | 866 | void _zfcp_dbf_scsi(const char *tag, const char *tag2, int level, |
913 | struct zfcp_adapter *adapter, | 867 | struct zfcp_dbf *dbf, struct scsi_cmnd *scsi_cmnd, |
914 | struct scsi_cmnd *scsi_cmnd, | 868 | struct zfcp_fsf_req *fsf_req, unsigned long old_req_id) |
915 | struct zfcp_fsf_req *fsf_req, | ||
916 | unsigned long old_req_id) | ||
917 | { | 869 | { |
918 | struct zfcp_scsi_dbf_record *rec = &adapter->scsi_dbf_buf; | 870 | struct zfcp_dbf_scsi_record *rec = &dbf->scsi_buf; |
919 | struct zfcp_dbf_dump *dump = (struct zfcp_dbf_dump *)rec; | 871 | struct zfcp_dbf_dump *dump = (struct zfcp_dbf_dump *)rec; |
920 | unsigned long flags; | 872 | unsigned long flags; |
921 | struct fcp_rsp_iu *fcp_rsp; | 873 | struct fcp_rsp_iu *fcp_rsp; |
922 | char *fcp_rsp_info = NULL, *fcp_sns_info = NULL; | 874 | char *fcp_rsp_info = NULL, *fcp_sns_info = NULL; |
923 | int offset = 0, buflen = 0; | 875 | int offset = 0, buflen = 0; |
924 | 876 | ||
925 | spin_lock_irqsave(&adapter->scsi_dbf_lock, flags); | 877 | spin_lock_irqsave(&dbf->scsi_lock, flags); |
926 | do { | 878 | do { |
927 | memset(rec, 0, sizeof(*rec)); | 879 | memset(rec, 0, sizeof(*rec)); |
928 | if (offset == 0) { | 880 | if (offset == 0) { |
@@ -976,68 +928,20 @@ static void zfcp_scsi_dbf_event(const char *tag, const char *tag2, int level, | |||
976 | dump->offset = offset; | 928 | dump->offset = offset; |
977 | dump->size = min(buflen - offset, | 929 | dump->size = min(buflen - offset, |
978 | (int)sizeof(struct | 930 | (int)sizeof(struct |
979 | zfcp_scsi_dbf_record) - | 931 | zfcp_dbf_scsi_record) - |
980 | (int)sizeof(struct zfcp_dbf_dump)); | 932 | (int)sizeof(struct zfcp_dbf_dump)); |
981 | memcpy(dump->data, fcp_sns_info + offset, dump->size); | 933 | memcpy(dump->data, fcp_sns_info + offset, dump->size); |
982 | offset += dump->size; | 934 | offset += dump->size; |
983 | } | 935 | } |
984 | debug_event(adapter->scsi_dbf, level, rec, sizeof(*rec)); | 936 | debug_event(dbf->scsi, level, rec, sizeof(*rec)); |
985 | } while (offset < buflen); | 937 | } while (offset < buflen); |
986 | spin_unlock_irqrestore(&adapter->scsi_dbf_lock, flags); | 938 | spin_unlock_irqrestore(&dbf->scsi_lock, flags); |
987 | } | ||
988 | |||
989 | /** | ||
990 | * zfcp_scsi_dbf_event_result - trace event for SCSI command completion | ||
991 | * @tag: tag indicating success or failure of SCSI command | ||
992 | * @level: trace level applicable for this event | ||
993 | * @adapter: adapter that has been used to issue the SCSI command | ||
994 | * @scsi_cmnd: SCSI command pointer | ||
995 | * @fsf_req: request used to issue SCSI command (might be NULL) | ||
996 | */ | ||
997 | void zfcp_scsi_dbf_event_result(const char *tag, int level, | ||
998 | struct zfcp_adapter *adapter, | ||
999 | struct scsi_cmnd *scsi_cmnd, | ||
1000 | struct zfcp_fsf_req *fsf_req) | ||
1001 | { | ||
1002 | zfcp_scsi_dbf_event("rslt", tag, level, adapter, scsi_cmnd, fsf_req, 0); | ||
1003 | } | 939 | } |
1004 | 940 | ||
1005 | /** | 941 | static int zfcp_dbf_scsi_view_format(debug_info_t *id, struct debug_view *view, |
1006 | * zfcp_scsi_dbf_event_abort - trace event for SCSI command abort | ||
1007 | * @tag: tag indicating success or failure of abort operation | ||
1008 | * @adapter: adapter thas has been used to issue SCSI command to be aborted | ||
1009 | * @scsi_cmnd: SCSI command to be aborted | ||
1010 | * @new_fsf_req: request containing abort (might be NULL) | ||
1011 | * @old_req_id: identifier of request containg SCSI command to be aborted | ||
1012 | */ | ||
1013 | void zfcp_scsi_dbf_event_abort(const char *tag, struct zfcp_adapter *adapter, | ||
1014 | struct scsi_cmnd *scsi_cmnd, | ||
1015 | struct zfcp_fsf_req *new_fsf_req, | ||
1016 | unsigned long old_req_id) | ||
1017 | { | ||
1018 | zfcp_scsi_dbf_event("abrt", tag, 1, adapter, scsi_cmnd, new_fsf_req, | ||
1019 | old_req_id); | ||
1020 | } | ||
1021 | |||
1022 | /** | ||
1023 | * zfcp_scsi_dbf_event_devreset - trace event for Logical Unit or Target Reset | ||
1024 | * @tag: tag indicating success or failure of reset operation | ||
1025 | * @flag: indicates type of reset (Target Reset, Logical Unit Reset) | ||
1026 | * @unit: unit that needs reset | ||
1027 | * @scsi_cmnd: SCSI command which caused this error recovery | ||
1028 | */ | ||
1029 | void zfcp_scsi_dbf_event_devreset(const char *tag, u8 flag, | ||
1030 | struct zfcp_unit *unit, | ||
1031 | struct scsi_cmnd *scsi_cmnd) | ||
1032 | { | ||
1033 | zfcp_scsi_dbf_event(flag == FCP_TARGET_RESET ? "trst" : "lrst", tag, 1, | ||
1034 | unit->port->adapter, scsi_cmnd, NULL, 0); | ||
1035 | } | ||
1036 | |||
1037 | static int zfcp_scsi_dbf_view_format(debug_info_t *id, struct debug_view *view, | ||
1038 | char *out_buf, const char *in_buf) | 942 | char *out_buf, const char *in_buf) |
1039 | { | 943 | { |
1040 | struct zfcp_scsi_dbf_record *r = (struct zfcp_scsi_dbf_record *)in_buf; | 944 | struct zfcp_dbf_scsi_record *r = (struct zfcp_dbf_scsi_record *)in_buf; |
1041 | struct timespec t; | 945 | struct timespec t; |
1042 | char *p = out_buf; | 946 | char *p = out_buf; |
1043 | 947 | ||
@@ -1059,7 +963,7 @@ static int zfcp_scsi_dbf_view_format(debug_info_t *id, struct debug_view *view, | |||
1059 | zfcp_dbf_out(&p, "old_fsf_reqid", "0x%0Lx", r->old_fsf_reqid); | 963 | zfcp_dbf_out(&p, "old_fsf_reqid", "0x%0Lx", r->old_fsf_reqid); |
1060 | zfcp_dbf_out(&p, "fsf_reqid", "0x%0Lx", r->fsf_reqid); | 964 | zfcp_dbf_out(&p, "fsf_reqid", "0x%0Lx", r->fsf_reqid); |
1061 | zfcp_dbf_out(&p, "fsf_seqno", "0x%08x", r->fsf_seqno); | 965 | zfcp_dbf_out(&p, "fsf_seqno", "0x%08x", r->fsf_seqno); |
1062 | zfcp_dbf_timestamp(r->fsf_issued, &t); | 966 | stck_to_timespec(r->fsf_issued, &t); |
1063 | zfcp_dbf_out(&p, "fsf_issued", "%011lu:%06lu", t.tv_sec, t.tv_nsec); | 967 | zfcp_dbf_out(&p, "fsf_issued", "%011lu:%06lu", t.tv_sec, t.tv_nsec); |
1064 | 968 | ||
1065 | if (strncmp(r->tag, "rslt", ZFCP_DBF_TAG_SIZE) == 0) { | 969 | if (strncmp(r->tag, "rslt", ZFCP_DBF_TAG_SIZE) == 0) { |
@@ -1078,84 +982,96 @@ static int zfcp_scsi_dbf_view_format(debug_info_t *id, struct debug_view *view, | |||
1078 | return p - out_buf; | 982 | return p - out_buf; |
1079 | } | 983 | } |
1080 | 984 | ||
1081 | static struct debug_view zfcp_scsi_dbf_view = { | 985 | static struct debug_view zfcp_dbf_scsi_view = { |
1082 | "structured", | 986 | .name = "structured", |
1083 | NULL, | 987 | .header_proc = zfcp_dbf_view_header, |
1084 | &zfcp_dbf_view_header, | 988 | .format_proc = zfcp_dbf_scsi_view_format, |
1085 | &zfcp_scsi_dbf_view_format, | ||
1086 | NULL, | ||
1087 | NULL | ||
1088 | }; | 989 | }; |
1089 | 990 | ||
991 | static debug_info_t *zfcp_dbf_reg(const char *name, int level, | ||
992 | struct debug_view *view, int size) | ||
993 | { | ||
994 | struct debug_info *d; | ||
995 | |||
996 | d = debug_register(name, dbfsize, level, size); | ||
997 | if (!d) | ||
998 | return NULL; | ||
999 | |||
1000 | debug_register_view(d, &debug_hex_ascii_view); | ||
1001 | debug_register_view(d, view); | ||
1002 | debug_set_level(d, level); | ||
1003 | |||
1004 | return d; | ||
1005 | } | ||
1006 | |||
1090 | /** | 1007 | /** |
1091 | * zfcp_adapter_debug_register - registers debug feature for an adapter | 1008 | * zfcp_adapter_debug_register - registers debug feature for an adapter |
1092 | * @adapter: pointer to adapter for which debug features should be registered | 1009 | * @adapter: pointer to adapter for which debug features should be registered |
1093 | * return: -ENOMEM on error, 0 otherwise | 1010 | * return: -ENOMEM on error, 0 otherwise |
1094 | */ | 1011 | */ |
1095 | int zfcp_adapter_debug_register(struct zfcp_adapter *adapter) | 1012 | int zfcp_dbf_adapter_register(struct zfcp_adapter *adapter) |
1096 | { | 1013 | { |
1097 | char dbf_name[DEBUG_MAX_NAME_LEN]; | 1014 | char dbf_name[DEBUG_MAX_NAME_LEN]; |
1015 | struct zfcp_dbf *dbf; | ||
1016 | |||
1017 | dbf = kmalloc(sizeof(struct zfcp_dbf), GFP_KERNEL); | ||
1018 | if (!dbf) | ||
1019 | return -ENOMEM; | ||
1020 | |||
1021 | dbf->adapter = adapter; | ||
1022 | |||
1023 | spin_lock_init(&dbf->hba_lock); | ||
1024 | spin_lock_init(&dbf->san_lock); | ||
1025 | spin_lock_init(&dbf->scsi_lock); | ||
1026 | spin_lock_init(&dbf->rec_lock); | ||
1098 | 1027 | ||
1099 | /* debug feature area which records recovery activity */ | 1028 | /* debug feature area which records recovery activity */ |
1100 | sprintf(dbf_name, "zfcp_%s_rec", dev_name(&adapter->ccw_device->dev)); | 1029 | sprintf(dbf_name, "zfcp_%s_rec", dev_name(&adapter->ccw_device->dev)); |
1101 | adapter->rec_dbf = debug_register(dbf_name, dbfsize, 1, | 1030 | dbf->rec = zfcp_dbf_reg(dbf_name, 3, &zfcp_dbf_rec_view, |
1102 | sizeof(struct zfcp_rec_dbf_record)); | 1031 | sizeof(struct zfcp_dbf_rec_record)); |
1103 | if (!adapter->rec_dbf) | 1032 | if (!dbf->rec) |
1104 | goto failed; | 1033 | goto err_out; |
1105 | debug_register_view(adapter->rec_dbf, &debug_hex_ascii_view); | ||
1106 | debug_register_view(adapter->rec_dbf, &zfcp_rec_dbf_view); | ||
1107 | debug_set_level(adapter->rec_dbf, 3); | ||
1108 | 1034 | ||
1109 | /* debug feature area which records HBA (FSF and QDIO) conditions */ | 1035 | /* debug feature area which records HBA (FSF and QDIO) conditions */ |
1110 | sprintf(dbf_name, "zfcp_%s_hba", dev_name(&adapter->ccw_device->dev)); | 1036 | sprintf(dbf_name, "zfcp_%s_hba", dev_name(&adapter->ccw_device->dev)); |
1111 | adapter->hba_dbf = debug_register(dbf_name, dbfsize, 1, | 1037 | dbf->hba = zfcp_dbf_reg(dbf_name, 3, &zfcp_dbf_hba_view, |
1112 | sizeof(struct zfcp_hba_dbf_record)); | 1038 | sizeof(struct zfcp_dbf_hba_record)); |
1113 | if (!adapter->hba_dbf) | 1039 | if (!dbf->hba) |
1114 | goto failed; | 1040 | goto err_out; |
1115 | debug_register_view(adapter->hba_dbf, &debug_hex_ascii_view); | ||
1116 | debug_register_view(adapter->hba_dbf, &zfcp_hba_dbf_view); | ||
1117 | debug_set_level(adapter->hba_dbf, 3); | ||
1118 | 1041 | ||
1119 | /* debug feature area which records SAN command failures and recovery */ | 1042 | /* debug feature area which records SAN command failures and recovery */ |
1120 | sprintf(dbf_name, "zfcp_%s_san", dev_name(&adapter->ccw_device->dev)); | 1043 | sprintf(dbf_name, "zfcp_%s_san", dev_name(&adapter->ccw_device->dev)); |
1121 | adapter->san_dbf = debug_register(dbf_name, dbfsize, 1, | 1044 | dbf->san = zfcp_dbf_reg(dbf_name, 6, &zfcp_dbf_san_view, |
1122 | sizeof(struct zfcp_san_dbf_record)); | 1045 | sizeof(struct zfcp_dbf_san_record)); |
1123 | if (!adapter->san_dbf) | 1046 | if (!dbf->san) |
1124 | goto failed; | 1047 | goto err_out; |
1125 | debug_register_view(adapter->san_dbf, &debug_hex_ascii_view); | ||
1126 | debug_register_view(adapter->san_dbf, &zfcp_san_dbf_view); | ||
1127 | debug_set_level(adapter->san_dbf, 6); | ||
1128 | 1048 | ||
1129 | /* debug feature area which records SCSI command failures and recovery */ | 1049 | /* debug feature area which records SCSI command failures and recovery */ |
1130 | sprintf(dbf_name, "zfcp_%s_scsi", dev_name(&adapter->ccw_device->dev)); | 1050 | sprintf(dbf_name, "zfcp_%s_scsi", dev_name(&adapter->ccw_device->dev)); |
1131 | adapter->scsi_dbf = debug_register(dbf_name, dbfsize, 1, | 1051 | dbf->scsi = zfcp_dbf_reg(dbf_name, 3, &zfcp_dbf_scsi_view, |
1132 | sizeof(struct zfcp_scsi_dbf_record)); | 1052 | sizeof(struct zfcp_dbf_scsi_record)); |
1133 | if (!adapter->scsi_dbf) | 1053 | if (!dbf->scsi) |
1134 | goto failed; | 1054 | goto err_out; |
1135 | debug_register_view(adapter->scsi_dbf, &debug_hex_ascii_view); | ||
1136 | debug_register_view(adapter->scsi_dbf, &zfcp_scsi_dbf_view); | ||
1137 | debug_set_level(adapter->scsi_dbf, 3); | ||
1138 | 1055 | ||
1056 | adapter->dbf = dbf; | ||
1139 | return 0; | 1057 | return 0; |
1140 | 1058 | ||
1141 | failed: | 1059 | err_out: |
1142 | zfcp_adapter_debug_unregister(adapter); | 1060 | zfcp_dbf_adapter_unregister(dbf); |
1143 | |||
1144 | return -ENOMEM; | 1061 | return -ENOMEM; |
1145 | } | 1062 | } |
1146 | 1063 | ||
1147 | /** | 1064 | /** |
1148 | * zfcp_adapter_debug_unregister - unregisters debug feature for an adapter | 1065 | * zfcp_adapter_debug_unregister - unregisters debug feature for an adapter |
1149 | * @adapter: pointer to adapter for which debug features should be unregistered | 1066 | * @dbf: pointer to dbf for which debug features should be unregistered |
1150 | */ | 1067 | */ |
1151 | void zfcp_adapter_debug_unregister(struct zfcp_adapter *adapter) | 1068 | void zfcp_dbf_adapter_unregister(struct zfcp_dbf *dbf) |
1152 | { | 1069 | { |
1153 | debug_unregister(adapter->scsi_dbf); | 1070 | debug_unregister(dbf->scsi); |
1154 | debug_unregister(adapter->san_dbf); | 1071 | debug_unregister(dbf->san); |
1155 | debug_unregister(adapter->hba_dbf); | 1072 | debug_unregister(dbf->hba); |
1156 | debug_unregister(adapter->rec_dbf); | 1073 | debug_unregister(dbf->rec); |
1157 | adapter->scsi_dbf = NULL; | 1074 | dbf->adapter->dbf = NULL; |
1158 | adapter->san_dbf = NULL; | 1075 | kfree(dbf); |
1159 | adapter->hba_dbf = NULL; | ||
1160 | adapter->rec_dbf = NULL; | ||
1161 | } | 1076 | } |
1077 | |||
diff --git a/drivers/s390/scsi/zfcp_dbf.h b/drivers/s390/scsi/zfcp_dbf.h index a573f7344dd6..6b1461e8f847 100644 --- a/drivers/s390/scsi/zfcp_dbf.h +++ b/drivers/s390/scsi/zfcp_dbf.h | |||
@@ -2,7 +2,7 @@ | |||
2 | * This file is part of the zfcp device driver for | 2 | * This file is part of the zfcp device driver for |
3 | * FCP adapters for IBM System z9 and zSeries. | 3 | * FCP adapters for IBM System z9 and zSeries. |
4 | * | 4 | * |
5 | * Copyright IBM Corp. 2008, 2008 | 5 | * Copyright IBM Corp. 2008, 2009 |
6 | * | 6 | * |
7 | * This program is free software; you can redistribute it and/or modify | 7 | * This program is free software; you can redistribute it and/or modify |
8 | * it under the terms of the GNU General Public License as published by | 8 | * it under the terms of the GNU General Public License as published by |
@@ -22,7 +22,9 @@ | |||
22 | #ifndef ZFCP_DBF_H | 22 | #ifndef ZFCP_DBF_H |
23 | #define ZFCP_DBF_H | 23 | #define ZFCP_DBF_H |
24 | 24 | ||
25 | #include "zfcp_ext.h" | ||
25 | #include "zfcp_fsf.h" | 26 | #include "zfcp_fsf.h" |
27 | #include "zfcp_def.h" | ||
26 | 28 | ||
27 | #define ZFCP_DBF_TAG_SIZE 4 | 29 | #define ZFCP_DBF_TAG_SIZE 4 |
28 | #define ZFCP_DBF_ID_SIZE 7 | 30 | #define ZFCP_DBF_ID_SIZE 7 |
@@ -35,13 +37,13 @@ struct zfcp_dbf_dump { | |||
35 | u8 data[]; /* dump data */ | 37 | u8 data[]; /* dump data */ |
36 | } __attribute__ ((packed)); | 38 | } __attribute__ ((packed)); |
37 | 39 | ||
38 | struct zfcp_rec_dbf_record_thread { | 40 | struct zfcp_dbf_rec_record_thread { |
39 | u32 total; | 41 | u32 total; |
40 | u32 ready; | 42 | u32 ready; |
41 | u32 running; | 43 | u32 running; |
42 | }; | 44 | }; |
43 | 45 | ||
44 | struct zfcp_rec_dbf_record_target { | 46 | struct zfcp_dbf_rec_record_target { |
45 | u64 ref; | 47 | u64 ref; |
46 | u32 status; | 48 | u32 status; |
47 | u32 d_id; | 49 | u32 d_id; |
@@ -50,7 +52,7 @@ struct zfcp_rec_dbf_record_target { | |||
50 | u32 erp_count; | 52 | u32 erp_count; |
51 | }; | 53 | }; |
52 | 54 | ||
53 | struct zfcp_rec_dbf_record_trigger { | 55 | struct zfcp_dbf_rec_record_trigger { |
54 | u8 want; | 56 | u8 want; |
55 | u8 need; | 57 | u8 need; |
56 | u32 as; | 58 | u32 as; |
@@ -62,21 +64,21 @@ struct zfcp_rec_dbf_record_trigger { | |||
62 | u64 fcp_lun; | 64 | u64 fcp_lun; |
63 | }; | 65 | }; |
64 | 66 | ||
65 | struct zfcp_rec_dbf_record_action { | 67 | struct zfcp_dbf_rec_record_action { |
66 | u32 status; | 68 | u32 status; |
67 | u32 step; | 69 | u32 step; |
68 | u64 action; | 70 | u64 action; |
69 | u64 fsf_req; | 71 | u64 fsf_req; |
70 | }; | 72 | }; |
71 | 73 | ||
72 | struct zfcp_rec_dbf_record { | 74 | struct zfcp_dbf_rec_record { |
73 | u8 id; | 75 | u8 id; |
74 | char id2[7]; | 76 | char id2[7]; |
75 | union { | 77 | union { |
76 | struct zfcp_rec_dbf_record_action action; | 78 | struct zfcp_dbf_rec_record_action action; |
77 | struct zfcp_rec_dbf_record_thread thread; | 79 | struct zfcp_dbf_rec_record_thread thread; |
78 | struct zfcp_rec_dbf_record_target target; | 80 | struct zfcp_dbf_rec_record_target target; |
79 | struct zfcp_rec_dbf_record_trigger trigger; | 81 | struct zfcp_dbf_rec_record_trigger trigger; |
80 | } u; | 82 | } u; |
81 | }; | 83 | }; |
82 | 84 | ||
@@ -87,7 +89,7 @@ enum { | |||
87 | ZFCP_REC_DBF_ID_TRIGGER, | 89 | ZFCP_REC_DBF_ID_TRIGGER, |
88 | }; | 90 | }; |
89 | 91 | ||
90 | struct zfcp_hba_dbf_record_response { | 92 | struct zfcp_dbf_hba_record_response { |
91 | u32 fsf_command; | 93 | u32 fsf_command; |
92 | u64 fsf_reqid; | 94 | u64 fsf_reqid; |
93 | u32 fsf_seqno; | 95 | u32 fsf_seqno; |
@@ -125,7 +127,7 @@ struct zfcp_hba_dbf_record_response { | |||
125 | } u; | 127 | } u; |
126 | } __attribute__ ((packed)); | 128 | } __attribute__ ((packed)); |
127 | 129 | ||
128 | struct zfcp_hba_dbf_record_status { | 130 | struct zfcp_dbf_hba_record_status { |
129 | u8 failed; | 131 | u8 failed; |
130 | u32 status_type; | 132 | u32 status_type; |
131 | u32 status_subtype; | 133 | u32 status_subtype; |
@@ -139,24 +141,24 @@ struct zfcp_hba_dbf_record_status { | |||
139 | u8 payload[ZFCP_DBF_UNSOL_PAYLOAD]; | 141 | u8 payload[ZFCP_DBF_UNSOL_PAYLOAD]; |
140 | } __attribute__ ((packed)); | 142 | } __attribute__ ((packed)); |
141 | 143 | ||
142 | struct zfcp_hba_dbf_record_qdio { | 144 | struct zfcp_dbf_hba_record_qdio { |
143 | u32 qdio_error; | 145 | u32 qdio_error; |
144 | u8 sbal_index; | 146 | u8 sbal_index; |
145 | u8 sbal_count; | 147 | u8 sbal_count; |
146 | } __attribute__ ((packed)); | 148 | } __attribute__ ((packed)); |
147 | 149 | ||
148 | struct zfcp_hba_dbf_record { | 150 | struct zfcp_dbf_hba_record { |
149 | u8 tag[ZFCP_DBF_TAG_SIZE]; | 151 | u8 tag[ZFCP_DBF_TAG_SIZE]; |
150 | u8 tag2[ZFCP_DBF_TAG_SIZE]; | 152 | u8 tag2[ZFCP_DBF_TAG_SIZE]; |
151 | union { | 153 | union { |
152 | struct zfcp_hba_dbf_record_response response; | 154 | struct zfcp_dbf_hba_record_response response; |
153 | struct zfcp_hba_dbf_record_status status; | 155 | struct zfcp_dbf_hba_record_status status; |
154 | struct zfcp_hba_dbf_record_qdio qdio; | 156 | struct zfcp_dbf_hba_record_qdio qdio; |
155 | struct fsf_bit_error_payload berr; | 157 | struct fsf_bit_error_payload berr; |
156 | } u; | 158 | } u; |
157 | } __attribute__ ((packed)); | 159 | } __attribute__ ((packed)); |
158 | 160 | ||
159 | struct zfcp_san_dbf_record_ct_request { | 161 | struct zfcp_dbf_san_record_ct_request { |
160 | u16 cmd_req_code; | 162 | u16 cmd_req_code; |
161 | u8 revision; | 163 | u8 revision; |
162 | u8 gs_type; | 164 | u8 gs_type; |
@@ -166,7 +168,7 @@ struct zfcp_san_dbf_record_ct_request { | |||
166 | u32 len; | 168 | u32 len; |
167 | } __attribute__ ((packed)); | 169 | } __attribute__ ((packed)); |
168 | 170 | ||
169 | struct zfcp_san_dbf_record_ct_response { | 171 | struct zfcp_dbf_san_record_ct_response { |
170 | u16 cmd_rsp_code; | 172 | u16 cmd_rsp_code; |
171 | u8 revision; | 173 | u8 revision; |
172 | u8 reason_code; | 174 | u8 reason_code; |
@@ -176,27 +178,27 @@ struct zfcp_san_dbf_record_ct_response { | |||
176 | u32 len; | 178 | u32 len; |
177 | } __attribute__ ((packed)); | 179 | } __attribute__ ((packed)); |
178 | 180 | ||
179 | struct zfcp_san_dbf_record_els { | 181 | struct zfcp_dbf_san_record_els { |
180 | u8 ls_code; | 182 | u8 ls_code; |
181 | u32 len; | 183 | u32 len; |
182 | } __attribute__ ((packed)); | 184 | } __attribute__ ((packed)); |
183 | 185 | ||
184 | struct zfcp_san_dbf_record { | 186 | struct zfcp_dbf_san_record { |
185 | u8 tag[ZFCP_DBF_TAG_SIZE]; | 187 | u8 tag[ZFCP_DBF_TAG_SIZE]; |
186 | u64 fsf_reqid; | 188 | u64 fsf_reqid; |
187 | u32 fsf_seqno; | 189 | u32 fsf_seqno; |
188 | u32 s_id; | 190 | u32 s_id; |
189 | u32 d_id; | 191 | u32 d_id; |
190 | union { | 192 | union { |
191 | struct zfcp_san_dbf_record_ct_request ct_req; | 193 | struct zfcp_dbf_san_record_ct_request ct_req; |
192 | struct zfcp_san_dbf_record_ct_response ct_resp; | 194 | struct zfcp_dbf_san_record_ct_response ct_resp; |
193 | struct zfcp_san_dbf_record_els els; | 195 | struct zfcp_dbf_san_record_els els; |
194 | } u; | 196 | } u; |
195 | #define ZFCP_DBF_SAN_MAX_PAYLOAD 1024 | 197 | #define ZFCP_DBF_SAN_MAX_PAYLOAD 1024 |
196 | u8 payload[32]; | 198 | u8 payload[32]; |
197 | } __attribute__ ((packed)); | 199 | } __attribute__ ((packed)); |
198 | 200 | ||
199 | struct zfcp_scsi_dbf_record { | 201 | struct zfcp_dbf_scsi_record { |
200 | u8 tag[ZFCP_DBF_TAG_SIZE]; | 202 | u8 tag[ZFCP_DBF_TAG_SIZE]; |
201 | u8 tag2[ZFCP_DBF_TAG_SIZE]; | 203 | u8 tag2[ZFCP_DBF_TAG_SIZE]; |
202 | u32 scsi_id; | 204 | u32 scsi_id; |
@@ -222,4 +224,127 @@ struct zfcp_scsi_dbf_record { | |||
222 | u8 sns_info[ZFCP_DBF_SCSI_FCP_SNS_INFO]; | 224 | u8 sns_info[ZFCP_DBF_SCSI_FCP_SNS_INFO]; |
223 | } __attribute__ ((packed)); | 225 | } __attribute__ ((packed)); |
224 | 226 | ||
227 | struct zfcp_dbf { | ||
228 | debug_info_t *rec; | ||
229 | debug_info_t *hba; | ||
230 | debug_info_t *san; | ||
231 | debug_info_t *scsi; | ||
232 | spinlock_t rec_lock; | ||
233 | spinlock_t hba_lock; | ||
234 | spinlock_t san_lock; | ||
235 | spinlock_t scsi_lock; | ||
236 | struct zfcp_dbf_rec_record rec_buf; | ||
237 | struct zfcp_dbf_hba_record hba_buf; | ||
238 | struct zfcp_dbf_san_record san_buf; | ||
239 | struct zfcp_dbf_scsi_record scsi_buf; | ||
240 | struct zfcp_adapter *adapter; | ||
241 | }; | ||
242 | |||
243 | static inline | ||
244 | void zfcp_dbf_hba_fsf_resp(const char *tag2, int level, | ||
245 | struct zfcp_fsf_req *req, struct zfcp_dbf *dbf) | ||
246 | { | ||
247 | if (level <= dbf->hba->level) | ||
248 | _zfcp_dbf_hba_fsf_response(tag2, level, req, dbf); | ||
249 | } | ||
250 | |||
251 | /** | ||
252 | * zfcp_dbf_hba_fsf_response - trace event for request completion | ||
253 | * @fsf_req: request that has been completed | ||
254 | */ | ||
255 | static inline void zfcp_dbf_hba_fsf_response(struct zfcp_fsf_req *req) | ||
256 | { | ||
257 | struct zfcp_dbf *dbf = req->adapter->dbf; | ||
258 | struct fsf_qtcb *qtcb = req->qtcb; | ||
259 | |||
260 | if ((qtcb->prefix.prot_status != FSF_PROT_GOOD) && | ||
261 | (qtcb->prefix.prot_status != FSF_PROT_FSF_STATUS_PRESENTED)) { | ||
262 | zfcp_dbf_hba_fsf_resp("perr", 1, req, dbf); | ||
263 | |||
264 | } else if (qtcb->header.fsf_status != FSF_GOOD) { | ||
265 | zfcp_dbf_hba_fsf_resp("ferr", 1, req, dbf); | ||
266 | |||
267 | } else if ((req->fsf_command == FSF_QTCB_OPEN_PORT_WITH_DID) || | ||
268 | (req->fsf_command == FSF_QTCB_OPEN_LUN)) { | ||
269 | zfcp_dbf_hba_fsf_resp("open", 4, req, dbf); | ||
270 | |||
271 | } else if (qtcb->header.log_length) { | ||
272 | zfcp_dbf_hba_fsf_resp("qtcb", 5, req, dbf); | ||
273 | |||
274 | } else { | ||
275 | zfcp_dbf_hba_fsf_resp("norm", 6, req, dbf); | ||
276 | } | ||
277 | } | ||
278 | |||
279 | /** | ||
280 | * zfcp_dbf_hba_fsf_unsol - trace event for an unsolicited status buffer | ||
281 | * @tag: tag indicating which kind of unsolicited status has been received | ||
282 | * @dbf: reference to dbf structure | ||
283 | * @status_buffer: buffer containing payload of unsolicited status | ||
284 | */ | ||
285 | static inline | ||
286 | void zfcp_dbf_hba_fsf_unsol(const char *tag, struct zfcp_dbf *dbf, | ||
287 | struct fsf_status_read_buffer *buf) | ||
288 | { | ||
289 | int level = 2; | ||
290 | |||
291 | if (level <= dbf->hba->level) | ||
292 | _zfcp_dbf_hba_fsf_unsol(tag, level, dbf, buf); | ||
293 | } | ||
294 | |||
295 | static inline | ||
296 | void zfcp_dbf_scsi(const char *tag, const char *tag2, int level, | ||
297 | struct zfcp_dbf *dbf, struct scsi_cmnd *scmd, | ||
298 | struct zfcp_fsf_req *req, unsigned long old_id) | ||
299 | { | ||
300 | if (level <= dbf->scsi->level) | ||
301 | _zfcp_dbf_scsi(tag, tag2, level, dbf, scmd, req, old_id); | ||
302 | } | ||
303 | |||
304 | /** | ||
305 | * zfcp_dbf_scsi_result - trace event for SCSI command completion | ||
306 | * @tag: tag indicating success or failure of SCSI command | ||
307 | * @level: trace level applicable for this event | ||
308 | * @adapter: adapter that has been used to issue the SCSI command | ||
309 | * @scmd: SCSI command pointer | ||
310 | * @fsf_req: request used to issue SCSI command (might be NULL) | ||
311 | */ | ||
312 | static inline | ||
313 | void zfcp_dbf_scsi_result(const char *tag, int level, struct zfcp_dbf *dbf, | ||
314 | struct scsi_cmnd *scmd, struct zfcp_fsf_req *fsf_req) | ||
315 | { | ||
316 | zfcp_dbf_scsi("rslt", tag, level, dbf, scmd, fsf_req, 0); | ||
317 | } | ||
318 | |||
319 | /** | ||
320 | * zfcp_dbf_scsi_abort - trace event for SCSI command abort | ||
321 | * @tag: tag indicating success or failure of abort operation | ||
322 | * @adapter: adapter thas has been used to issue SCSI command to be aborted | ||
323 | * @scmd: SCSI command to be aborted | ||
324 | * @new_req: request containing abort (might be NULL) | ||
325 | * @old_id: identifier of request containg SCSI command to be aborted | ||
326 | */ | ||
327 | static inline | ||
328 | void zfcp_dbf_scsi_abort(const char *tag, struct zfcp_dbf *dbf, | ||
329 | struct scsi_cmnd *scmd, struct zfcp_fsf_req *new_req, | ||
330 | unsigned long old_id) | ||
331 | { | ||
332 | zfcp_dbf_scsi("abrt", tag, 1, dbf, scmd, new_req, old_id); | ||
333 | } | ||
334 | |||
335 | /** | ||
336 | * zfcp_dbf_scsi_devreset - trace event for Logical Unit or Target Reset | ||
337 | * @tag: tag indicating success or failure of reset operation | ||
338 | * @flag: indicates type of reset (Target Reset, Logical Unit Reset) | ||
339 | * @unit: unit that needs reset | ||
340 | * @scsi_cmnd: SCSI command which caused this error recovery | ||
341 | */ | ||
342 | static inline | ||
343 | void zfcp_dbf_scsi_devreset(const char *tag, u8 flag, struct zfcp_unit *unit, | ||
344 | struct scsi_cmnd *scsi_cmnd) | ||
345 | { | ||
346 | zfcp_dbf_scsi(flag == FCP_TARGET_RESET ? "trst" : "lrst", tag, 1, | ||
347 | unit->port->adapter->dbf, scsi_cmnd, NULL, 0); | ||
348 | } | ||
349 | |||
225 | #endif /* ZFCP_DBF_H */ | 350 | #endif /* ZFCP_DBF_H */ |
diff --git a/drivers/s390/scsi/zfcp_def.h b/drivers/s390/scsi/zfcp_def.h index 49d0532bca1c..7da2fad8f515 100644 --- a/drivers/s390/scsi/zfcp_def.h +++ b/drivers/s390/scsi/zfcp_def.h | |||
@@ -37,10 +37,8 @@ | |||
37 | #include <asm/debug.h> | 37 | #include <asm/debug.h> |
38 | #include <asm/ebcdic.h> | 38 | #include <asm/ebcdic.h> |
39 | #include <asm/sysinfo.h> | 39 | #include <asm/sysinfo.h> |
40 | #include "zfcp_dbf.h" | ||
41 | #include "zfcp_fsf.h" | 40 | #include "zfcp_fsf.h" |
42 | 41 | ||
43 | |||
44 | /********************* GENERAL DEFINES *********************************/ | 42 | /********************* GENERAL DEFINES *********************************/ |
45 | 43 | ||
46 | #define REQUEST_LIST_SIZE 128 | 44 | #define REQUEST_LIST_SIZE 128 |
@@ -75,9 +73,6 @@ | |||
75 | 73 | ||
76 | /*************** FIBRE CHANNEL PROTOCOL SPECIFIC DEFINES ********************/ | 74 | /*************** FIBRE CHANNEL PROTOCOL SPECIFIC DEFINES ********************/ |
77 | 75 | ||
78 | /* timeout for name-server lookup (in seconds) */ | ||
79 | #define ZFCP_NS_GID_PN_TIMEOUT 10 | ||
80 | |||
81 | /* task attribute values in FCP-2 FCP_CMND IU */ | 76 | /* task attribute values in FCP-2 FCP_CMND IU */ |
82 | #define SIMPLE_Q 0 | 77 | #define SIMPLE_Q 0 |
83 | #define HEAD_OF_Q 1 | 78 | #define HEAD_OF_Q 1 |
@@ -224,8 +219,6 @@ struct zfcp_ls_adisc { | |||
224 | #define ZFCP_STATUS_ADAPTER_QDIOUP 0x00000002 | 219 | #define ZFCP_STATUS_ADAPTER_QDIOUP 0x00000002 |
225 | #define ZFCP_STATUS_ADAPTER_XCONFIG_OK 0x00000008 | 220 | #define ZFCP_STATUS_ADAPTER_XCONFIG_OK 0x00000008 |
226 | #define ZFCP_STATUS_ADAPTER_HOST_CON_INIT 0x00000010 | 221 | #define ZFCP_STATUS_ADAPTER_HOST_CON_INIT 0x00000010 |
227 | #define ZFCP_STATUS_ADAPTER_ERP_THREAD_UP 0x00000020 | ||
228 | #define ZFCP_STATUS_ADAPTER_ERP_THREAD_KILL 0x00000080 | ||
229 | #define ZFCP_STATUS_ADAPTER_ERP_PENDING 0x00000100 | 222 | #define ZFCP_STATUS_ADAPTER_ERP_PENDING 0x00000100 |
230 | #define ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED 0x00000200 | 223 | #define ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED 0x00000200 |
231 | 224 | ||
@@ -234,6 +227,7 @@ struct zfcp_ls_adisc { | |||
234 | 227 | ||
235 | /* remote port status */ | 228 | /* remote port status */ |
236 | #define ZFCP_STATUS_PORT_PHYS_OPEN 0x00000001 | 229 | #define ZFCP_STATUS_PORT_PHYS_OPEN 0x00000001 |
230 | #define ZFCP_STATUS_PORT_LINK_TEST 0x00000002 | ||
237 | 231 | ||
238 | /* well known address (WKA) port status*/ | 232 | /* well known address (WKA) port status*/ |
239 | enum zfcp_wka_status { | 233 | enum zfcp_wka_status { |
@@ -249,7 +243,6 @@ enum zfcp_wka_status { | |||
249 | 243 | ||
250 | /* FSF request status (this does not have a common part) */ | 244 | /* FSF request status (this does not have a common part) */ |
251 | #define ZFCP_STATUS_FSFREQ_TASK_MANAGEMENT 0x00000002 | 245 | #define ZFCP_STATUS_FSFREQ_TASK_MANAGEMENT 0x00000002 |
252 | #define ZFCP_STATUS_FSFREQ_COMPLETED 0x00000004 | ||
253 | #define ZFCP_STATUS_FSFREQ_ERROR 0x00000008 | 246 | #define ZFCP_STATUS_FSFREQ_ERROR 0x00000008 |
254 | #define ZFCP_STATUS_FSFREQ_CLEANUP 0x00000010 | 247 | #define ZFCP_STATUS_FSFREQ_CLEANUP 0x00000010 |
255 | #define ZFCP_STATUS_FSFREQ_ABORTSUCCEEDED 0x00000040 | 248 | #define ZFCP_STATUS_FSFREQ_ABORTSUCCEEDED 0x00000040 |
@@ -266,12 +259,14 @@ struct zfcp_fsf_req; | |||
266 | 259 | ||
267 | /* holds various memory pools of an adapter */ | 260 | /* holds various memory pools of an adapter */ |
268 | struct zfcp_adapter_mempool { | 261 | struct zfcp_adapter_mempool { |
269 | mempool_t *fsf_req_erp; | 262 | mempool_t *erp_req; |
270 | mempool_t *fsf_req_scsi; | 263 | mempool_t *gid_pn_req; |
271 | mempool_t *fsf_req_abort; | 264 | mempool_t *scsi_req; |
272 | mempool_t *fsf_req_status_read; | 265 | mempool_t *scsi_abort; |
273 | mempool_t *data_status_read; | 266 | mempool_t *status_read_req; |
274 | mempool_t *data_gid_pn; | 267 | mempool_t *status_read_data; |
268 | mempool_t *gid_pn_data; | ||
269 | mempool_t *qtcb_pool; | ||
275 | }; | 270 | }; |
276 | 271 | ||
277 | /* | 272 | /* |
@@ -305,6 +300,15 @@ struct ct_iu_gid_pn_resp { | |||
305 | u32 d_id; | 300 | u32 d_id; |
306 | } __attribute__ ((packed)); | 301 | } __attribute__ ((packed)); |
307 | 302 | ||
303 | struct ct_iu_gpn_ft_req { | ||
304 | struct ct_hdr header; | ||
305 | u8 flags; | ||
306 | u8 domain_id_scope; | ||
307 | u8 area_id_scope; | ||
308 | u8 fc4_type; | ||
309 | } __attribute__ ((packed)); | ||
310 | |||
311 | |||
308 | /** | 312 | /** |
309 | * struct zfcp_send_ct - used to pass parameters to function zfcp_fsf_send_ct | 313 | * struct zfcp_send_ct - used to pass parameters to function zfcp_fsf_send_ct |
310 | * @wka_port: port where the request is sent to | 314 | * @wka_port: port where the request is sent to |
@@ -312,7 +316,6 @@ struct ct_iu_gid_pn_resp { | |||
312 | * @resp: scatter-gather list for response | 316 | * @resp: scatter-gather list for response |
313 | * @handler: handler function (called for response to the request) | 317 | * @handler: handler function (called for response to the request) |
314 | * @handler_data: data passed to handler function | 318 | * @handler_data: data passed to handler function |
315 | * @timeout: FSF timeout for this request | ||
316 | * @completion: completion for synchronization purposes | 319 | * @completion: completion for synchronization purposes |
317 | * @status: used to pass error status to calling function | 320 | * @status: used to pass error status to calling function |
318 | */ | 321 | */ |
@@ -322,7 +325,6 @@ struct zfcp_send_ct { | |||
322 | struct scatterlist *resp; | 325 | struct scatterlist *resp; |
323 | void (*handler)(unsigned long); | 326 | void (*handler)(unsigned long); |
324 | unsigned long handler_data; | 327 | unsigned long handler_data; |
325 | int timeout; | ||
326 | struct completion *completion; | 328 | struct completion *completion; |
327 | int status; | 329 | int status; |
328 | }; | 330 | }; |
@@ -420,6 +422,29 @@ struct zfcp_latencies { | |||
420 | spinlock_t lock; | 422 | spinlock_t lock; |
421 | }; | 423 | }; |
422 | 424 | ||
425 | /** struct zfcp_qdio - basic QDIO data structure | ||
426 | * @resp_q: response queue | ||
427 | * @req_q: request queue | ||
428 | * @stat_lock: lock to protect req_q_util and req_q_time | ||
429 | * @req_q_lock; lock to serialize access to request queue | ||
430 | * @req_q_time: time of last fill level change | ||
431 | * @req_q_util: used for accounting | ||
432 | * @req_q_full: queue full incidents | ||
433 | * @req_q_wq: used to wait for SBAL availability | ||
434 | * @adapter: adapter used in conjunction with this QDIO structure | ||
435 | */ | ||
436 | struct zfcp_qdio { | ||
437 | struct zfcp_qdio_queue resp_q; | ||
438 | struct zfcp_qdio_queue req_q; | ||
439 | spinlock_t stat_lock; | ||
440 | spinlock_t req_q_lock; | ||
441 | unsigned long long req_q_time; | ||
442 | u64 req_q_util; | ||
443 | atomic_t req_q_full; | ||
444 | wait_queue_head_t req_q_wq; | ||
445 | struct zfcp_adapter *adapter; | ||
446 | }; | ||
447 | |||
423 | struct zfcp_adapter { | 448 | struct zfcp_adapter { |
424 | atomic_t refcount; /* reference count */ | 449 | atomic_t refcount; /* reference count */ |
425 | wait_queue_head_t remove_wq; /* can be used to wait for | 450 | wait_queue_head_t remove_wq; /* can be used to wait for |
@@ -428,6 +453,7 @@ struct zfcp_adapter { | |||
428 | u64 peer_wwpn; /* P2P peer WWPN */ | 453 | u64 peer_wwpn; /* P2P peer WWPN */ |
429 | u32 peer_d_id; /* P2P peer D_ID */ | 454 | u32 peer_d_id; /* P2P peer D_ID */ |
430 | struct ccw_device *ccw_device; /* S/390 ccw device */ | 455 | struct ccw_device *ccw_device; /* S/390 ccw device */ |
456 | struct zfcp_qdio *qdio; | ||
431 | u32 hydra_version; /* Hydra version */ | 457 | u32 hydra_version; /* Hydra version */ |
432 | u32 fsf_lic_version; | 458 | u32 fsf_lic_version; |
433 | u32 adapter_features; /* FCP channel features */ | 459 | u32 adapter_features; /* FCP channel features */ |
@@ -439,15 +465,7 @@ struct zfcp_adapter { | |||
439 | unsigned long req_no; /* unique FSF req number */ | 465 | unsigned long req_no; /* unique FSF req number */ |
440 | struct list_head *req_list; /* list of pending reqs */ | 466 | struct list_head *req_list; /* list of pending reqs */ |
441 | spinlock_t req_list_lock; /* request list lock */ | 467 | spinlock_t req_list_lock; /* request list lock */ |
442 | struct zfcp_qdio_queue req_q; /* request queue */ | ||
443 | spinlock_t req_q_lock; /* for operations on queue */ | ||
444 | ktime_t req_q_time; /* time of last fill level change */ | ||
445 | u64 req_q_util; /* for accounting */ | ||
446 | spinlock_t qdio_stat_lock; | ||
447 | u32 fsf_req_seq_no; /* FSF cmnd seq number */ | 468 | u32 fsf_req_seq_no; /* FSF cmnd seq number */ |
448 | wait_queue_head_t request_wq; /* can be used to wait for | ||
449 | more avaliable SBALs */ | ||
450 | struct zfcp_qdio_queue resp_q; /* response queue */ | ||
451 | rwlock_t abort_lock; /* Protects against SCSI | 469 | rwlock_t abort_lock; /* Protects against SCSI |
452 | stack abort/command | 470 | stack abort/command |
453 | completion races */ | 471 | completion races */ |
@@ -456,10 +474,9 @@ struct zfcp_adapter { | |||
456 | atomic_t status; /* status of this adapter */ | 474 | atomic_t status; /* status of this adapter */ |
457 | struct list_head erp_ready_head; /* error recovery for this | 475 | struct list_head erp_ready_head; /* error recovery for this |
458 | adapter/devices */ | 476 | adapter/devices */ |
477 | wait_queue_head_t erp_ready_wq; | ||
459 | struct list_head erp_running_head; | 478 | struct list_head erp_running_head; |
460 | rwlock_t erp_lock; | 479 | rwlock_t erp_lock; |
461 | struct semaphore erp_ready_sem; | ||
462 | wait_queue_head_t erp_thread_wqh; | ||
463 | wait_queue_head_t erp_done_wqh; | 480 | wait_queue_head_t erp_done_wqh; |
464 | struct zfcp_erp_action erp_action; /* pending error recovery */ | 481 | struct zfcp_erp_action erp_action; /* pending error recovery */ |
465 | atomic_t erp_counter; | 482 | atomic_t erp_counter; |
@@ -467,27 +484,16 @@ struct zfcp_adapter { | |||
467 | actions */ | 484 | actions */ |
468 | u32 erp_low_mem_count; /* nr of erp actions waiting | 485 | u32 erp_low_mem_count; /* nr of erp actions waiting |
469 | for memory */ | 486 | for memory */ |
487 | struct task_struct *erp_thread; | ||
470 | struct zfcp_wka_ports *gs; /* generic services */ | 488 | struct zfcp_wka_ports *gs; /* generic services */ |
471 | debug_info_t *rec_dbf; | 489 | struct zfcp_dbf *dbf; /* debug traces */ |
472 | debug_info_t *hba_dbf; | ||
473 | debug_info_t *san_dbf; /* debug feature areas */ | ||
474 | debug_info_t *scsi_dbf; | ||
475 | spinlock_t rec_dbf_lock; | ||
476 | spinlock_t hba_dbf_lock; | ||
477 | spinlock_t san_dbf_lock; | ||
478 | spinlock_t scsi_dbf_lock; | ||
479 | struct zfcp_rec_dbf_record rec_dbf_buf; | ||
480 | struct zfcp_hba_dbf_record hba_dbf_buf; | ||
481 | struct zfcp_san_dbf_record san_dbf_buf; | ||
482 | struct zfcp_scsi_dbf_record scsi_dbf_buf; | ||
483 | struct zfcp_adapter_mempool pool; /* Adapter memory pools */ | 490 | struct zfcp_adapter_mempool pool; /* Adapter memory pools */ |
484 | struct qdio_initialize qdio_init_data; /* for qdio_establish */ | ||
485 | struct fc_host_statistics *fc_stats; | 491 | struct fc_host_statistics *fc_stats; |
486 | struct fsf_qtcb_bottom_port *stats_reset_data; | 492 | struct fsf_qtcb_bottom_port *stats_reset_data; |
487 | unsigned long stats_reset; | 493 | unsigned long stats_reset; |
488 | struct work_struct scan_work; | 494 | struct work_struct scan_work; |
489 | struct service_level service_level; | 495 | struct service_level service_level; |
490 | atomic_t qdio_outb_full; /* queue full incidents */ | 496 | struct workqueue_struct *work_queue; |
491 | }; | 497 | }; |
492 | 498 | ||
493 | struct zfcp_port { | 499 | struct zfcp_port { |
@@ -531,36 +537,64 @@ struct zfcp_unit { | |||
531 | struct work_struct scsi_work; | 537 | struct work_struct scsi_work; |
532 | }; | 538 | }; |
533 | 539 | ||
534 | /* FSF request */ | 540 | /** |
541 | * struct zfcp_queue_req - queue related values for a request | ||
542 | * @sbal_number: number of free SBALs | ||
543 | * @sbal_first: first SBAL for this request | ||
544 | * @sbal_last: last SBAL for this request | ||
545 | * @sbal_limit: last possible SBAL for this request | ||
546 | * @sbale_curr: current SBALE at creation of this request | ||
547 | * @sbal_response: SBAL used in interrupt | ||
548 | * @qdio_outb_usage: usage of outbound queue | ||
549 | * @qdio_inb_usage: usage of inbound queue | ||
550 | */ | ||
551 | struct zfcp_queue_req { | ||
552 | u8 sbal_number; | ||
553 | u8 sbal_first; | ||
554 | u8 sbal_last; | ||
555 | u8 sbal_limit; | ||
556 | u8 sbale_curr; | ||
557 | u8 sbal_response; | ||
558 | u16 qdio_outb_usage; | ||
559 | u16 qdio_inb_usage; | ||
560 | }; | ||
561 | |||
562 | /** | ||
563 | * struct zfcp_fsf_req - basic FSF request structure | ||
564 | * @list: list of FSF requests | ||
565 | * @req_id: unique request ID | ||
566 | * @adapter: adapter this request belongs to | ||
567 | * @queue_req: queue related values | ||
568 | * @completion: used to signal the completion of the request | ||
569 | * @status: status of the request | ||
570 | * @fsf_command: FSF command issued | ||
571 | * @qtcb: associated QTCB | ||
572 | * @seq_no: sequence number of this request | ||
573 | * @data: private data | ||
574 | * @timer: timer data of this request | ||
575 | * @erp_action: reference to erp action if request issued on behalf of ERP | ||
576 | * @pool: reference to memory pool if used for this request | ||
577 | * @issued: time when request was send (STCK) | ||
578 | * @unit: reference to unit if this request is a SCSI request | ||
579 | * @handler: handler which should be called to process response | ||
580 | */ | ||
535 | struct zfcp_fsf_req { | 581 | struct zfcp_fsf_req { |
536 | struct list_head list; /* list of FSF requests */ | 582 | struct list_head list; |
537 | unsigned long req_id; /* unique request ID */ | 583 | unsigned long req_id; |
538 | struct zfcp_adapter *adapter; /* adapter request belongs to */ | 584 | struct zfcp_adapter *adapter; |
539 | u8 sbal_number; /* nr of SBALs free for use */ | 585 | struct zfcp_queue_req queue_req; |
540 | u8 sbal_first; /* first SBAL for this request */ | 586 | struct completion completion; |
541 | u8 sbal_last; /* last SBAL for this request */ | 587 | u32 status; |
542 | u8 sbal_limit; /* last possible SBAL for | 588 | u32 fsf_command; |
543 | this reuest */ | 589 | struct fsf_qtcb *qtcb; |
544 | u8 sbale_curr; /* current SBALE during creation | 590 | u32 seq_no; |
545 | of request */ | 591 | void *data; |
546 | u8 sbal_response; /* SBAL used in interrupt */ | 592 | struct timer_list timer; |
547 | wait_queue_head_t completion_wq; /* can be used by a routine | 593 | struct zfcp_erp_action *erp_action; |
548 | to wait for completion */ | 594 | mempool_t *pool; |
549 | u32 status; /* status of this request */ | 595 | unsigned long long issued; |
550 | u32 fsf_command; /* FSF Command copy */ | 596 | struct zfcp_unit *unit; |
551 | struct fsf_qtcb *qtcb; /* address of associated QTCB */ | ||
552 | u32 seq_no; /* Sequence number of request */ | ||
553 | void *data; /* private data of request */ | ||
554 | struct timer_list timer; /* used for erp or scsi er */ | ||
555 | struct zfcp_erp_action *erp_action; /* used if this request is | ||
556 | issued on behalf of erp */ | ||
557 | mempool_t *pool; /* used if request was alloacted | ||
558 | from emergency pool */ | ||
559 | unsigned long long issued; /* request sent time (STCK) */ | ||
560 | struct zfcp_unit *unit; | ||
561 | void (*handler)(struct zfcp_fsf_req *); | 597 | void (*handler)(struct zfcp_fsf_req *); |
562 | u16 qdio_outb_usage;/* usage of outbound queue */ | ||
563 | u16 qdio_inb_usage; /* usage of inbound queue */ | ||
564 | }; | 598 | }; |
565 | 599 | ||
566 | /* driver data */ | 600 | /* driver data */ |
@@ -570,18 +604,11 @@ struct zfcp_data { | |||
570 | rwlock_t config_lock; /* serialises changes | 604 | rwlock_t config_lock; /* serialises changes |
571 | to adapter/port/unit | 605 | to adapter/port/unit |
572 | lists */ | 606 | lists */ |
573 | struct semaphore config_sema; /* serialises configuration | 607 | struct mutex config_mutex; |
574 | changes */ | 608 | struct kmem_cache *gpn_ft_cache; |
575 | struct kmem_cache *fsf_req_qtcb_cache; | 609 | struct kmem_cache *qtcb_cache; |
576 | struct kmem_cache *sr_buffer_cache; | 610 | struct kmem_cache *sr_buffer_cache; |
577 | struct kmem_cache *gid_pn_cache; | 611 | struct kmem_cache *gid_pn_cache; |
578 | struct workqueue_struct *work_queue; | ||
579 | }; | ||
580 | |||
581 | /* struct used by memory pools for fsf_requests */ | ||
582 | struct zfcp_fsf_req_qtcb { | ||
583 | struct zfcp_fsf_req fsf_req; | ||
584 | struct fsf_qtcb qtcb; | ||
585 | }; | 612 | }; |
586 | 613 | ||
587 | /********************** ZFCP SPECIFIC DEFINES ********************************/ | 614 | /********************** ZFCP SPECIFIC DEFINES ********************************/ |
diff --git a/drivers/s390/scsi/zfcp_erp.c b/drivers/s390/scsi/zfcp_erp.c index c75d6f35cb5f..73d366ba31e5 100644 --- a/drivers/s390/scsi/zfcp_erp.c +++ b/drivers/s390/scsi/zfcp_erp.c | |||
@@ -9,6 +9,7 @@ | |||
9 | #define KMSG_COMPONENT "zfcp" | 9 | #define KMSG_COMPONENT "zfcp" |
10 | #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt | 10 | #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt |
11 | 11 | ||
12 | #include <linux/kthread.h> | ||
12 | #include "zfcp_ext.h" | 13 | #include "zfcp_ext.h" |
13 | 14 | ||
14 | #define ZFCP_MAX_ERPS 3 | 15 | #define ZFCP_MAX_ERPS 3 |
@@ -26,7 +27,6 @@ enum zfcp_erp_steps { | |||
26 | ZFCP_ERP_STEP_FSF_XCONFIG = 0x0001, | 27 | ZFCP_ERP_STEP_FSF_XCONFIG = 0x0001, |
27 | ZFCP_ERP_STEP_PHYS_PORT_CLOSING = 0x0010, | 28 | ZFCP_ERP_STEP_PHYS_PORT_CLOSING = 0x0010, |
28 | ZFCP_ERP_STEP_PORT_CLOSING = 0x0100, | 29 | ZFCP_ERP_STEP_PORT_CLOSING = 0x0100, |
29 | ZFCP_ERP_STEP_NAMESERVER_LOOKUP = 0x0400, | ||
30 | ZFCP_ERP_STEP_PORT_OPENING = 0x0800, | 30 | ZFCP_ERP_STEP_PORT_OPENING = 0x0800, |
31 | ZFCP_ERP_STEP_UNIT_CLOSING = 0x1000, | 31 | ZFCP_ERP_STEP_UNIT_CLOSING = 0x1000, |
32 | ZFCP_ERP_STEP_UNIT_OPENING = 0x2000, | 32 | ZFCP_ERP_STEP_UNIT_OPENING = 0x2000, |
@@ -75,9 +75,9 @@ static void zfcp_erp_action_ready(struct zfcp_erp_action *act) | |||
75 | struct zfcp_adapter *adapter = act->adapter; | 75 | struct zfcp_adapter *adapter = act->adapter; |
76 | 76 | ||
77 | list_move(&act->list, &act->adapter->erp_ready_head); | 77 | list_move(&act->list, &act->adapter->erp_ready_head); |
78 | zfcp_rec_dbf_event_action("erardy1", act); | 78 | zfcp_dbf_rec_action("erardy1", act); |
79 | up(&adapter->erp_ready_sem); | 79 | wake_up(&adapter->erp_ready_wq); |
80 | zfcp_rec_dbf_event_thread("erardy2", adapter); | 80 | zfcp_dbf_rec_thread("erardy2", adapter->dbf); |
81 | } | 81 | } |
82 | 82 | ||
83 | static void zfcp_erp_action_dismiss(struct zfcp_erp_action *act) | 83 | static void zfcp_erp_action_dismiss(struct zfcp_erp_action *act) |
@@ -150,6 +150,9 @@ static int zfcp_erp_required_act(int want, struct zfcp_adapter *adapter, | |||
150 | a_status = atomic_read(&adapter->status); | 150 | a_status = atomic_read(&adapter->status); |
151 | if (a_status & ZFCP_STATUS_COMMON_ERP_INUSE) | 151 | if (a_status & ZFCP_STATUS_COMMON_ERP_INUSE) |
152 | return 0; | 152 | return 0; |
153 | if (!(a_status & ZFCP_STATUS_COMMON_RUNNING) && | ||
154 | !(a_status & ZFCP_STATUS_COMMON_OPEN)) | ||
155 | return 0; /* shutdown requested for closed adapter */ | ||
153 | } | 156 | } |
154 | 157 | ||
155 | return need; | 158 | return need; |
@@ -213,8 +216,7 @@ static int zfcp_erp_action_enqueue(int want, struct zfcp_adapter *adapter, | |||
213 | int retval = 1, need; | 216 | int retval = 1, need; |
214 | struct zfcp_erp_action *act = NULL; | 217 | struct zfcp_erp_action *act = NULL; |
215 | 218 | ||
216 | if (!(atomic_read(&adapter->status) & | 219 | if (!adapter->erp_thread) |
217 | ZFCP_STATUS_ADAPTER_ERP_THREAD_UP)) | ||
218 | return -EIO; | 220 | return -EIO; |
219 | 221 | ||
220 | need = zfcp_erp_required_act(want, adapter, port, unit); | 222 | need = zfcp_erp_required_act(want, adapter, port, unit); |
@@ -227,12 +229,11 @@ static int zfcp_erp_action_enqueue(int want, struct zfcp_adapter *adapter, | |||
227 | goto out; | 229 | goto out; |
228 | ++adapter->erp_total_count; | 230 | ++adapter->erp_total_count; |
229 | list_add_tail(&act->list, &adapter->erp_ready_head); | 231 | list_add_tail(&act->list, &adapter->erp_ready_head); |
230 | up(&adapter->erp_ready_sem); | 232 | wake_up(&adapter->erp_ready_wq); |
231 | zfcp_rec_dbf_event_thread("eracte1", adapter); | 233 | zfcp_dbf_rec_thread("eracte1", adapter->dbf); |
232 | retval = 0; | 234 | retval = 0; |
233 | out: | 235 | out: |
234 | zfcp_rec_dbf_event_trigger(id, ref, want, need, act, | 236 | zfcp_dbf_rec_trigger(id, ref, want, need, act, adapter, port, unit); |
235 | adapter, port, unit); | ||
236 | return retval; | 237 | return retval; |
237 | } | 238 | } |
238 | 239 | ||
@@ -443,28 +444,28 @@ static int status_change_clear(unsigned long mask, atomic_t *status) | |||
443 | static void zfcp_erp_adapter_unblock(struct zfcp_adapter *adapter) | 444 | static void zfcp_erp_adapter_unblock(struct zfcp_adapter *adapter) |
444 | { | 445 | { |
445 | if (status_change_set(ZFCP_STATUS_COMMON_UNBLOCKED, &adapter->status)) | 446 | if (status_change_set(ZFCP_STATUS_COMMON_UNBLOCKED, &adapter->status)) |
446 | zfcp_rec_dbf_event_adapter("eraubl1", NULL, adapter); | 447 | zfcp_dbf_rec_adapter("eraubl1", NULL, adapter->dbf); |
447 | atomic_set_mask(ZFCP_STATUS_COMMON_UNBLOCKED, &adapter->status); | 448 | atomic_set_mask(ZFCP_STATUS_COMMON_UNBLOCKED, &adapter->status); |
448 | } | 449 | } |
449 | 450 | ||
450 | static void zfcp_erp_port_unblock(struct zfcp_port *port) | 451 | static void zfcp_erp_port_unblock(struct zfcp_port *port) |
451 | { | 452 | { |
452 | if (status_change_set(ZFCP_STATUS_COMMON_UNBLOCKED, &port->status)) | 453 | if (status_change_set(ZFCP_STATUS_COMMON_UNBLOCKED, &port->status)) |
453 | zfcp_rec_dbf_event_port("erpubl1", NULL, port); | 454 | zfcp_dbf_rec_port("erpubl1", NULL, port); |
454 | atomic_set_mask(ZFCP_STATUS_COMMON_UNBLOCKED, &port->status); | 455 | atomic_set_mask(ZFCP_STATUS_COMMON_UNBLOCKED, &port->status); |
455 | } | 456 | } |
456 | 457 | ||
457 | static void zfcp_erp_unit_unblock(struct zfcp_unit *unit) | 458 | static void zfcp_erp_unit_unblock(struct zfcp_unit *unit) |
458 | { | 459 | { |
459 | if (status_change_set(ZFCP_STATUS_COMMON_UNBLOCKED, &unit->status)) | 460 | if (status_change_set(ZFCP_STATUS_COMMON_UNBLOCKED, &unit->status)) |
460 | zfcp_rec_dbf_event_unit("eruubl1", NULL, unit); | 461 | zfcp_dbf_rec_unit("eruubl1", NULL, unit); |
461 | atomic_set_mask(ZFCP_STATUS_COMMON_UNBLOCKED, &unit->status); | 462 | atomic_set_mask(ZFCP_STATUS_COMMON_UNBLOCKED, &unit->status); |
462 | } | 463 | } |
463 | 464 | ||
464 | static void zfcp_erp_action_to_running(struct zfcp_erp_action *erp_action) | 465 | static void zfcp_erp_action_to_running(struct zfcp_erp_action *erp_action) |
465 | { | 466 | { |
466 | list_move(&erp_action->list, &erp_action->adapter->erp_running_head); | 467 | list_move(&erp_action->list, &erp_action->adapter->erp_running_head); |
467 | zfcp_rec_dbf_event_action("erator1", erp_action); | 468 | zfcp_dbf_rec_action("erator1", erp_action); |
468 | } | 469 | } |
469 | 470 | ||
470 | static void zfcp_erp_strategy_check_fsfreq(struct zfcp_erp_action *act) | 471 | static void zfcp_erp_strategy_check_fsfreq(struct zfcp_erp_action *act) |
@@ -480,13 +481,12 @@ static void zfcp_erp_strategy_check_fsfreq(struct zfcp_erp_action *act) | |||
480 | if (act->status & (ZFCP_STATUS_ERP_DISMISSED | | 481 | if (act->status & (ZFCP_STATUS_ERP_DISMISSED | |
481 | ZFCP_STATUS_ERP_TIMEDOUT)) { | 482 | ZFCP_STATUS_ERP_TIMEDOUT)) { |
482 | act->fsf_req->status |= ZFCP_STATUS_FSFREQ_DISMISSED; | 483 | act->fsf_req->status |= ZFCP_STATUS_FSFREQ_DISMISSED; |
483 | zfcp_rec_dbf_event_action("erscf_1", act); | 484 | zfcp_dbf_rec_action("erscf_1", act); |
484 | act->fsf_req->erp_action = NULL; | 485 | act->fsf_req->erp_action = NULL; |
485 | } | 486 | } |
486 | if (act->status & ZFCP_STATUS_ERP_TIMEDOUT) | 487 | if (act->status & ZFCP_STATUS_ERP_TIMEDOUT) |
487 | zfcp_rec_dbf_event_action("erscf_2", act); | 488 | zfcp_dbf_rec_action("erscf_2", act); |
488 | if (act->fsf_req->status & (ZFCP_STATUS_FSFREQ_COMPLETED | | 489 | if (act->fsf_req->status & ZFCP_STATUS_FSFREQ_DISMISSED) |
489 | ZFCP_STATUS_FSFREQ_DISMISSED)) | ||
490 | act->fsf_req = NULL; | 490 | act->fsf_req = NULL; |
491 | } else | 491 | } else |
492 | act->fsf_req = NULL; | 492 | act->fsf_req = NULL; |
@@ -604,9 +604,11 @@ static void zfcp_erp_wakeup(struct zfcp_adapter *adapter) | |||
604 | 604 | ||
605 | static int zfcp_erp_adapter_strategy_open_qdio(struct zfcp_erp_action *act) | 605 | static int zfcp_erp_adapter_strategy_open_qdio(struct zfcp_erp_action *act) |
606 | { | 606 | { |
607 | if (zfcp_qdio_open(act->adapter)) | 607 | struct zfcp_qdio *qdio = act->adapter->qdio; |
608 | |||
609 | if (zfcp_qdio_open(qdio)) | ||
608 | return ZFCP_ERP_FAILED; | 610 | return ZFCP_ERP_FAILED; |
609 | init_waitqueue_head(&act->adapter->request_wq); | 611 | init_waitqueue_head(&qdio->req_q_wq); |
610 | atomic_set_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &act->adapter->status); | 612 | atomic_set_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &act->adapter->status); |
611 | return ZFCP_ERP_SUCCEEDED; | 613 | return ZFCP_ERP_SUCCEEDED; |
612 | } | 614 | } |
@@ -641,9 +643,10 @@ static int zfcp_erp_adapter_strat_fsf_xconf(struct zfcp_erp_action *erp_action) | |||
641 | return ZFCP_ERP_FAILED; | 643 | return ZFCP_ERP_FAILED; |
642 | } | 644 | } |
643 | 645 | ||
644 | zfcp_rec_dbf_event_thread_lock("erasfx1", adapter); | 646 | zfcp_dbf_rec_thread_lock("erasfx1", adapter->dbf); |
645 | down(&adapter->erp_ready_sem); | 647 | wait_event(adapter->erp_ready_wq, |
646 | zfcp_rec_dbf_event_thread_lock("erasfx2", adapter); | 648 | !list_empty(&adapter->erp_ready_head)); |
649 | zfcp_dbf_rec_thread_lock("erasfx2", adapter->dbf); | ||
647 | if (erp_action->status & ZFCP_STATUS_ERP_TIMEDOUT) | 650 | if (erp_action->status & ZFCP_STATUS_ERP_TIMEDOUT) |
648 | break; | 651 | break; |
649 | 652 | ||
@@ -682,9 +685,10 @@ static int zfcp_erp_adapter_strategy_open_fsf_xport(struct zfcp_erp_action *act) | |||
682 | if (ret) | 685 | if (ret) |
683 | return ZFCP_ERP_FAILED; | 686 | return ZFCP_ERP_FAILED; |
684 | 687 | ||
685 | zfcp_rec_dbf_event_thread_lock("erasox1", adapter); | 688 | zfcp_dbf_rec_thread_lock("erasox1", adapter->dbf); |
686 | down(&adapter->erp_ready_sem); | 689 | wait_event(adapter->erp_ready_wq, |
687 | zfcp_rec_dbf_event_thread_lock("erasox2", adapter); | 690 | !list_empty(&adapter->erp_ready_head)); |
691 | zfcp_dbf_rec_thread_lock("erasox2", adapter->dbf); | ||
688 | if (act->status & ZFCP_STATUS_ERP_TIMEDOUT) | 692 | if (act->status & ZFCP_STATUS_ERP_TIMEDOUT) |
689 | return ZFCP_ERP_FAILED; | 693 | return ZFCP_ERP_FAILED; |
690 | 694 | ||
@@ -711,10 +715,10 @@ static void zfcp_erp_adapter_strategy_close(struct zfcp_erp_action *act) | |||
711 | struct zfcp_adapter *adapter = act->adapter; | 715 | struct zfcp_adapter *adapter = act->adapter; |
712 | 716 | ||
713 | /* close queues to ensure that buffers are not accessed by adapter */ | 717 | /* close queues to ensure that buffers are not accessed by adapter */ |
714 | zfcp_qdio_close(adapter); | 718 | zfcp_qdio_close(adapter->qdio); |
715 | zfcp_fsf_req_dismiss_all(adapter); | 719 | zfcp_fsf_req_dismiss_all(adapter); |
716 | adapter->fsf_req_seq_no = 0; | 720 | adapter->fsf_req_seq_no = 0; |
717 | zfcp_fc_wka_port_force_offline(&adapter->gs->ds); | 721 | zfcp_fc_wka_ports_force_offline(adapter->gs); |
718 | /* all ports and units are closed */ | 722 | /* all ports and units are closed */ |
719 | zfcp_erp_modify_adapter_status(adapter, "erascl1", NULL, | 723 | zfcp_erp_modify_adapter_status(adapter, "erascl1", NULL, |
720 | ZFCP_STATUS_COMMON_OPEN, ZFCP_CLEAR); | 724 | ZFCP_STATUS_COMMON_OPEN, ZFCP_CLEAR); |
@@ -841,27 +845,6 @@ static int zfcp_erp_open_ptp_port(struct zfcp_erp_action *act) | |||
841 | return zfcp_erp_port_strategy_open_port(act); | 845 | return zfcp_erp_port_strategy_open_port(act); |
842 | } | 846 | } |
843 | 847 | ||
844 | void zfcp_erp_port_strategy_open_lookup(struct work_struct *work) | ||
845 | { | ||
846 | int retval; | ||
847 | struct zfcp_port *port = container_of(work, struct zfcp_port, | ||
848 | gid_pn_work); | ||
849 | |||
850 | retval = zfcp_fc_ns_gid_pn(&port->erp_action); | ||
851 | if (!retval) { | ||
852 | port->erp_action.step = ZFCP_ERP_STEP_NAMESERVER_LOOKUP; | ||
853 | goto out; | ||
854 | } | ||
855 | if (retval == -ENOMEM) { | ||
856 | zfcp_erp_notify(&port->erp_action, ZFCP_STATUS_ERP_LOWMEM); | ||
857 | goto out; | ||
858 | } | ||
859 | /* all other error condtions */ | ||
860 | zfcp_erp_notify(&port->erp_action, 0); | ||
861 | out: | ||
862 | zfcp_port_put(port); | ||
863 | } | ||
864 | |||
865 | static int zfcp_erp_port_strategy_open_common(struct zfcp_erp_action *act) | 848 | static int zfcp_erp_port_strategy_open_common(struct zfcp_erp_action *act) |
866 | { | 849 | { |
867 | struct zfcp_adapter *adapter = act->adapter; | 850 | struct zfcp_adapter *adapter = act->adapter; |
@@ -876,15 +859,11 @@ static int zfcp_erp_port_strategy_open_common(struct zfcp_erp_action *act) | |||
876 | return zfcp_erp_open_ptp_port(act); | 859 | return zfcp_erp_open_ptp_port(act); |
877 | if (!port->d_id) { | 860 | if (!port->d_id) { |
878 | zfcp_port_get(port); | 861 | zfcp_port_get(port); |
879 | if (!queue_work(zfcp_data.work_queue, | 862 | if (!queue_work(adapter->work_queue, |
880 | &port->gid_pn_work)) | 863 | &port->gid_pn_work)) |
881 | zfcp_port_put(port); | 864 | zfcp_port_put(port); |
882 | return ZFCP_ERP_CONTINUES; | 865 | return ZFCP_ERP_EXIT; |
883 | } | 866 | } |
884 | /* fall through */ | ||
885 | case ZFCP_ERP_STEP_NAMESERVER_LOOKUP: | ||
886 | if (!port->d_id) | ||
887 | return ZFCP_ERP_FAILED; | ||
888 | return zfcp_erp_port_strategy_open_port(act); | 867 | return zfcp_erp_port_strategy_open_port(act); |
889 | 868 | ||
890 | case ZFCP_ERP_STEP_PORT_OPENING: | 869 | case ZFCP_ERP_STEP_PORT_OPENING: |
@@ -1163,7 +1142,7 @@ static void zfcp_erp_action_dequeue(struct zfcp_erp_action *erp_action) | |||
1163 | } | 1142 | } |
1164 | 1143 | ||
1165 | list_del(&erp_action->list); | 1144 | list_del(&erp_action->list); |
1166 | zfcp_rec_dbf_event_action("eractd1", erp_action); | 1145 | zfcp_dbf_rec_action("eractd1", erp_action); |
1167 | 1146 | ||
1168 | switch (erp_action->action) { | 1147 | switch (erp_action->action) { |
1169 | case ZFCP_ERP_ACTION_REOPEN_UNIT: | 1148 | case ZFCP_ERP_ACTION_REOPEN_UNIT: |
@@ -1311,20 +1290,16 @@ static int zfcp_erp_thread(void *data) | |||
1311 | struct list_head *next; | 1290 | struct list_head *next; |
1312 | struct zfcp_erp_action *act; | 1291 | struct zfcp_erp_action *act; |
1313 | unsigned long flags; | 1292 | unsigned long flags; |
1314 | int ignore; | ||
1315 | |||
1316 | daemonize("zfcperp%s", dev_name(&adapter->ccw_device->dev)); | ||
1317 | /* Block all signals */ | ||
1318 | siginitsetinv(¤t->blocked, 0); | ||
1319 | atomic_set_mask(ZFCP_STATUS_ADAPTER_ERP_THREAD_UP, &adapter->status); | ||
1320 | wake_up(&adapter->erp_thread_wqh); | ||
1321 | 1293 | ||
1322 | while (!(atomic_read(&adapter->status) & | 1294 | for (;;) { |
1323 | ZFCP_STATUS_ADAPTER_ERP_THREAD_KILL)) { | 1295 | zfcp_dbf_rec_thread_lock("erthrd1", adapter->dbf); |
1296 | wait_event_interruptible(adapter->erp_ready_wq, | ||
1297 | !list_empty(&adapter->erp_ready_head) || | ||
1298 | kthread_should_stop()); | ||
1299 | zfcp_dbf_rec_thread_lock("erthrd2", adapter->dbf); | ||
1324 | 1300 | ||
1325 | zfcp_rec_dbf_event_thread_lock("erthrd1", adapter); | 1301 | if (kthread_should_stop()) |
1326 | ignore = down_interruptible(&adapter->erp_ready_sem); | 1302 | break; |
1327 | zfcp_rec_dbf_event_thread_lock("erthrd2", adapter); | ||
1328 | 1303 | ||
1329 | write_lock_irqsave(&adapter->erp_lock, flags); | 1304 | write_lock_irqsave(&adapter->erp_lock, flags); |
1330 | next = adapter->erp_ready_head.next; | 1305 | next = adapter->erp_ready_head.next; |
@@ -1339,9 +1314,6 @@ static int zfcp_erp_thread(void *data) | |||
1339 | } | 1314 | } |
1340 | } | 1315 | } |
1341 | 1316 | ||
1342 | atomic_clear_mask(ZFCP_STATUS_ADAPTER_ERP_THREAD_UP, &adapter->status); | ||
1343 | wake_up(&adapter->erp_thread_wqh); | ||
1344 | |||
1345 | return 0; | 1317 | return 0; |
1346 | } | 1318 | } |
1347 | 1319 | ||
@@ -1353,18 +1325,17 @@ static int zfcp_erp_thread(void *data) | |||
1353 | */ | 1325 | */ |
1354 | int zfcp_erp_thread_setup(struct zfcp_adapter *adapter) | 1326 | int zfcp_erp_thread_setup(struct zfcp_adapter *adapter) |
1355 | { | 1327 | { |
1356 | int retval; | 1328 | struct task_struct *thread; |
1357 | 1329 | ||
1358 | atomic_clear_mask(ZFCP_STATUS_ADAPTER_ERP_THREAD_UP, &adapter->status); | 1330 | thread = kthread_run(zfcp_erp_thread, adapter, "zfcperp%s", |
1359 | retval = kernel_thread(zfcp_erp_thread, adapter, SIGCHLD); | 1331 | dev_name(&adapter->ccw_device->dev)); |
1360 | if (retval < 0) { | 1332 | if (IS_ERR(thread)) { |
1361 | dev_err(&adapter->ccw_device->dev, | 1333 | dev_err(&adapter->ccw_device->dev, |
1362 | "Creating an ERP thread for the FCP device failed.\n"); | 1334 | "Creating an ERP thread for the FCP device failed.\n"); |
1363 | return retval; | 1335 | return PTR_ERR(thread); |
1364 | } | 1336 | } |
1365 | wait_event(adapter->erp_thread_wqh, | 1337 | |
1366 | atomic_read(&adapter->status) & | 1338 | adapter->erp_thread = thread; |
1367 | ZFCP_STATUS_ADAPTER_ERP_THREAD_UP); | ||
1368 | return 0; | 1339 | return 0; |
1369 | } | 1340 | } |
1370 | 1341 | ||
@@ -1379,16 +1350,10 @@ int zfcp_erp_thread_setup(struct zfcp_adapter *adapter) | |||
1379 | */ | 1350 | */ |
1380 | void zfcp_erp_thread_kill(struct zfcp_adapter *adapter) | 1351 | void zfcp_erp_thread_kill(struct zfcp_adapter *adapter) |
1381 | { | 1352 | { |
1382 | atomic_set_mask(ZFCP_STATUS_ADAPTER_ERP_THREAD_KILL, &adapter->status); | 1353 | kthread_stop(adapter->erp_thread); |
1383 | up(&adapter->erp_ready_sem); | 1354 | adapter->erp_thread = NULL; |
1384 | zfcp_rec_dbf_event_thread_lock("erthrk1", adapter); | 1355 | WARN_ON(!list_empty(&adapter->erp_ready_head)); |
1385 | 1356 | WARN_ON(!list_empty(&adapter->erp_running_head)); | |
1386 | wait_event(adapter->erp_thread_wqh, | ||
1387 | !(atomic_read(&adapter->status) & | ||
1388 | ZFCP_STATUS_ADAPTER_ERP_THREAD_UP)); | ||
1389 | |||
1390 | atomic_clear_mask(ZFCP_STATUS_ADAPTER_ERP_THREAD_KILL, | ||
1391 | &adapter->status); | ||
1392 | } | 1357 | } |
1393 | 1358 | ||
1394 | /** | 1359 | /** |
@@ -1456,11 +1421,11 @@ void zfcp_erp_modify_adapter_status(struct zfcp_adapter *adapter, char *id, | |||
1456 | 1421 | ||
1457 | if (set_or_clear == ZFCP_SET) { | 1422 | if (set_or_clear == ZFCP_SET) { |
1458 | if (status_change_set(mask, &adapter->status)) | 1423 | if (status_change_set(mask, &adapter->status)) |
1459 | zfcp_rec_dbf_event_adapter(id, ref, adapter); | 1424 | zfcp_dbf_rec_adapter(id, ref, adapter->dbf); |
1460 | atomic_set_mask(mask, &adapter->status); | 1425 | atomic_set_mask(mask, &adapter->status); |
1461 | } else { | 1426 | } else { |
1462 | if (status_change_clear(mask, &adapter->status)) | 1427 | if (status_change_clear(mask, &adapter->status)) |
1463 | zfcp_rec_dbf_event_adapter(id, ref, adapter); | 1428 | zfcp_dbf_rec_adapter(id, ref, adapter->dbf); |
1464 | atomic_clear_mask(mask, &adapter->status); | 1429 | atomic_clear_mask(mask, &adapter->status); |
1465 | if (mask & ZFCP_STATUS_COMMON_ERP_FAILED) | 1430 | if (mask & ZFCP_STATUS_COMMON_ERP_FAILED) |
1466 | atomic_set(&adapter->erp_counter, 0); | 1431 | atomic_set(&adapter->erp_counter, 0); |
@@ -1490,11 +1455,11 @@ void zfcp_erp_modify_port_status(struct zfcp_port *port, char *id, void *ref, | |||
1490 | 1455 | ||
1491 | if (set_or_clear == ZFCP_SET) { | 1456 | if (set_or_clear == ZFCP_SET) { |
1492 | if (status_change_set(mask, &port->status)) | 1457 | if (status_change_set(mask, &port->status)) |
1493 | zfcp_rec_dbf_event_port(id, ref, port); | 1458 | zfcp_dbf_rec_port(id, ref, port); |
1494 | atomic_set_mask(mask, &port->status); | 1459 | atomic_set_mask(mask, &port->status); |
1495 | } else { | 1460 | } else { |
1496 | if (status_change_clear(mask, &port->status)) | 1461 | if (status_change_clear(mask, &port->status)) |
1497 | zfcp_rec_dbf_event_port(id, ref, port); | 1462 | zfcp_dbf_rec_port(id, ref, port); |
1498 | atomic_clear_mask(mask, &port->status); | 1463 | atomic_clear_mask(mask, &port->status); |
1499 | if (mask & ZFCP_STATUS_COMMON_ERP_FAILED) | 1464 | if (mask & ZFCP_STATUS_COMMON_ERP_FAILED) |
1500 | atomic_set(&port->erp_counter, 0); | 1465 | atomic_set(&port->erp_counter, 0); |
@@ -1519,11 +1484,11 @@ void zfcp_erp_modify_unit_status(struct zfcp_unit *unit, char *id, void *ref, | |||
1519 | { | 1484 | { |
1520 | if (set_or_clear == ZFCP_SET) { | 1485 | if (set_or_clear == ZFCP_SET) { |
1521 | if (status_change_set(mask, &unit->status)) | 1486 | if (status_change_set(mask, &unit->status)) |
1522 | zfcp_rec_dbf_event_unit(id, ref, unit); | 1487 | zfcp_dbf_rec_unit(id, ref, unit); |
1523 | atomic_set_mask(mask, &unit->status); | 1488 | atomic_set_mask(mask, &unit->status); |
1524 | } else { | 1489 | } else { |
1525 | if (status_change_clear(mask, &unit->status)) | 1490 | if (status_change_clear(mask, &unit->status)) |
1526 | zfcp_rec_dbf_event_unit(id, ref, unit); | 1491 | zfcp_dbf_rec_unit(id, ref, unit); |
1527 | atomic_clear_mask(mask, &unit->status); | 1492 | atomic_clear_mask(mask, &unit->status); |
1528 | if (mask & ZFCP_STATUS_COMMON_ERP_FAILED) { | 1493 | if (mask & ZFCP_STATUS_COMMON_ERP_FAILED) { |
1529 | atomic_set(&unit->erp_counter, 0); | 1494 | atomic_set(&unit->erp_counter, 0); |
diff --git a/drivers/s390/scsi/zfcp_ext.h b/drivers/s390/scsi/zfcp_ext.h index 3044c6010306..36935bc0818f 100644 --- a/drivers/s390/scsi/zfcp_ext.h +++ b/drivers/s390/scsi/zfcp_ext.h | |||
@@ -34,37 +34,31 @@ extern struct zfcp_adapter *zfcp_get_adapter_by_busid(char *); | |||
34 | extern struct miscdevice zfcp_cfdc_misc; | 34 | extern struct miscdevice zfcp_cfdc_misc; |
35 | 35 | ||
36 | /* zfcp_dbf.c */ | 36 | /* zfcp_dbf.c */ |
37 | extern int zfcp_adapter_debug_register(struct zfcp_adapter *); | 37 | extern int zfcp_dbf_adapter_register(struct zfcp_adapter *); |
38 | extern void zfcp_adapter_debug_unregister(struct zfcp_adapter *); | 38 | extern void zfcp_dbf_adapter_unregister(struct zfcp_dbf *); |
39 | extern void zfcp_rec_dbf_event_thread(char *, struct zfcp_adapter *); | 39 | extern void zfcp_dbf_rec_thread(char *, struct zfcp_dbf *); |
40 | extern void zfcp_rec_dbf_event_thread_lock(char *, struct zfcp_adapter *); | 40 | extern void zfcp_dbf_rec_thread_lock(char *, struct zfcp_dbf *); |
41 | extern void zfcp_rec_dbf_event_adapter(char *, void *, struct zfcp_adapter *); | 41 | extern void zfcp_dbf_rec_adapter(char *, void *, struct zfcp_dbf *); |
42 | extern void zfcp_rec_dbf_event_port(char *, void *, struct zfcp_port *); | 42 | extern void zfcp_dbf_rec_port(char *, void *, struct zfcp_port *); |
43 | extern void zfcp_rec_dbf_event_unit(char *, void *, struct zfcp_unit *); | 43 | extern void zfcp_dbf_rec_unit(char *, void *, struct zfcp_unit *); |
44 | extern void zfcp_rec_dbf_event_trigger(char *, void *, u8, u8, void *, | 44 | extern void zfcp_dbf_rec_trigger(char *, void *, u8, u8, void *, |
45 | struct zfcp_adapter *, | 45 | struct zfcp_adapter *, struct zfcp_port *, |
46 | struct zfcp_port *, struct zfcp_unit *); | 46 | struct zfcp_unit *); |
47 | extern void zfcp_rec_dbf_event_action(char *, struct zfcp_erp_action *); | 47 | extern void zfcp_dbf_rec_action(char *, struct zfcp_erp_action *); |
48 | extern void zfcp_hba_dbf_event_fsf_response(struct zfcp_fsf_req *); | 48 | extern void _zfcp_dbf_hba_fsf_response(const char *, int, struct zfcp_fsf_req *, |
49 | extern void zfcp_hba_dbf_event_fsf_unsol(const char *, struct zfcp_adapter *, | 49 | struct zfcp_dbf *); |
50 | struct fsf_status_read_buffer *); | 50 | extern void _zfcp_dbf_hba_fsf_unsol(const char *, int level, struct zfcp_dbf *, |
51 | extern void zfcp_hba_dbf_event_qdio(struct zfcp_adapter *, unsigned int, int, | 51 | struct fsf_status_read_buffer *); |
52 | int); | 52 | extern void zfcp_dbf_hba_qdio(struct zfcp_dbf *, unsigned int, int, int); |
53 | extern void zfcp_hba_dbf_event_berr(struct zfcp_adapter *, | 53 | extern void zfcp_dbf_hba_berr(struct zfcp_dbf *, struct zfcp_fsf_req *); |
54 | struct zfcp_fsf_req *); | 54 | extern void zfcp_dbf_san_ct_request(struct zfcp_fsf_req *); |
55 | extern void zfcp_san_dbf_event_ct_request(struct zfcp_fsf_req *); | 55 | extern void zfcp_dbf_san_ct_response(struct zfcp_fsf_req *); |
56 | extern void zfcp_san_dbf_event_ct_response(struct zfcp_fsf_req *); | 56 | extern void zfcp_dbf_san_els_request(struct zfcp_fsf_req *); |
57 | extern void zfcp_san_dbf_event_els_request(struct zfcp_fsf_req *); | 57 | extern void zfcp_dbf_san_els_response(struct zfcp_fsf_req *); |
58 | extern void zfcp_san_dbf_event_els_response(struct zfcp_fsf_req *); | 58 | extern void zfcp_dbf_san_incoming_els(struct zfcp_fsf_req *); |
59 | extern void zfcp_san_dbf_event_incoming_els(struct zfcp_fsf_req *); | 59 | extern void _zfcp_dbf_scsi(const char *, const char *, int, struct zfcp_dbf *, |
60 | extern void zfcp_scsi_dbf_event_result(const char *, int, struct zfcp_adapter *, | 60 | struct scsi_cmnd *, struct zfcp_fsf_req *, |
61 | struct scsi_cmnd *, | 61 | unsigned long); |
62 | struct zfcp_fsf_req *); | ||
63 | extern void zfcp_scsi_dbf_event_abort(const char *, struct zfcp_adapter *, | ||
64 | struct scsi_cmnd *, struct zfcp_fsf_req *, | ||
65 | unsigned long); | ||
66 | extern void zfcp_scsi_dbf_event_devreset(const char *, u8, struct zfcp_unit *, | ||
67 | struct scsi_cmnd *); | ||
68 | 62 | ||
69 | /* zfcp_erp.c */ | 63 | /* zfcp_erp.c */ |
70 | extern void zfcp_erp_modify_adapter_status(struct zfcp_adapter *, char *, | 64 | extern void zfcp_erp_modify_adapter_status(struct zfcp_adapter *, char *, |
@@ -96,22 +90,20 @@ extern void zfcp_erp_unit_access_denied(struct zfcp_unit *, char *, void *); | |||
96 | extern void zfcp_erp_adapter_access_changed(struct zfcp_adapter *, char *, | 90 | extern void zfcp_erp_adapter_access_changed(struct zfcp_adapter *, char *, |
97 | void *); | 91 | void *); |
98 | extern void zfcp_erp_timeout_handler(unsigned long); | 92 | extern void zfcp_erp_timeout_handler(unsigned long); |
99 | extern void zfcp_erp_port_strategy_open_lookup(struct work_struct *); | ||
100 | 93 | ||
101 | /* zfcp_fc.c */ | 94 | /* zfcp_fc.c */ |
102 | extern int zfcp_scan_ports(struct zfcp_adapter *); | 95 | extern int zfcp_fc_scan_ports(struct zfcp_adapter *); |
103 | extern void _zfcp_scan_ports_later(struct work_struct *); | 96 | extern void _zfcp_fc_scan_ports_later(struct work_struct *); |
104 | extern void zfcp_fc_incoming_els(struct zfcp_fsf_req *); | 97 | extern void zfcp_fc_incoming_els(struct zfcp_fsf_req *); |
105 | extern int zfcp_fc_ns_gid_pn(struct zfcp_erp_action *); | 98 | extern void zfcp_fc_port_did_lookup(struct work_struct *); |
106 | extern void zfcp_fc_plogi_evaluate(struct zfcp_port *, struct fsf_plogi *); | 99 | extern void zfcp_fc_plogi_evaluate(struct zfcp_port *, struct fsf_plogi *); |
107 | extern void zfcp_test_link(struct zfcp_port *); | 100 | extern void zfcp_fc_test_link(struct zfcp_port *); |
108 | extern void zfcp_fc_link_test_work(struct work_struct *); | 101 | extern void zfcp_fc_link_test_work(struct work_struct *); |
109 | extern void zfcp_fc_wka_port_force_offline(struct zfcp_wka_port *); | 102 | extern void zfcp_fc_wka_ports_force_offline(struct zfcp_wka_ports *); |
110 | extern void zfcp_fc_wka_ports_init(struct zfcp_adapter *); | 103 | extern int zfcp_fc_gs_setup(struct zfcp_adapter *); |
104 | extern void zfcp_fc_gs_destroy(struct zfcp_adapter *); | ||
111 | extern int zfcp_fc_execute_els_fc_job(struct fc_bsg_job *); | 105 | extern int zfcp_fc_execute_els_fc_job(struct fc_bsg_job *); |
112 | extern int zfcp_fc_execute_ct_fc_job(struct fc_bsg_job *); | 106 | extern int zfcp_fc_execute_ct_fc_job(struct fc_bsg_job *); |
113 | extern void zfcp_fc_wka_port_force_offline(struct zfcp_wka_port *); | ||
114 | |||
115 | 107 | ||
116 | /* zfcp_fsf.c */ | 108 | /* zfcp_fsf.c */ |
117 | extern int zfcp_fsf_open_port(struct zfcp_erp_action *); | 109 | extern int zfcp_fsf_open_port(struct zfcp_erp_action *); |
@@ -122,37 +114,39 @@ extern int zfcp_fsf_close_physical_port(struct zfcp_erp_action *); | |||
122 | extern int zfcp_fsf_open_unit(struct zfcp_erp_action *); | 114 | extern int zfcp_fsf_open_unit(struct zfcp_erp_action *); |
123 | extern int zfcp_fsf_close_unit(struct zfcp_erp_action *); | 115 | extern int zfcp_fsf_close_unit(struct zfcp_erp_action *); |
124 | extern int zfcp_fsf_exchange_config_data(struct zfcp_erp_action *); | 116 | extern int zfcp_fsf_exchange_config_data(struct zfcp_erp_action *); |
125 | extern int zfcp_fsf_exchange_config_data_sync(struct zfcp_adapter *, | 117 | extern int zfcp_fsf_exchange_config_data_sync(struct zfcp_qdio *, |
126 | struct fsf_qtcb_bottom_config *); | 118 | struct fsf_qtcb_bottom_config *); |
127 | extern int zfcp_fsf_exchange_port_data(struct zfcp_erp_action *); | 119 | extern int zfcp_fsf_exchange_port_data(struct zfcp_erp_action *); |
128 | extern int zfcp_fsf_exchange_port_data_sync(struct zfcp_adapter *, | 120 | extern int zfcp_fsf_exchange_port_data_sync(struct zfcp_qdio *, |
129 | struct fsf_qtcb_bottom_port *); | 121 | struct fsf_qtcb_bottom_port *); |
130 | extern struct zfcp_fsf_req *zfcp_fsf_control_file(struct zfcp_adapter *, | 122 | extern struct zfcp_fsf_req *zfcp_fsf_control_file(struct zfcp_adapter *, |
131 | struct zfcp_fsf_cfdc *); | 123 | struct zfcp_fsf_cfdc *); |
132 | extern void zfcp_fsf_req_dismiss_all(struct zfcp_adapter *); | 124 | extern void zfcp_fsf_req_dismiss_all(struct zfcp_adapter *); |
133 | extern int zfcp_fsf_status_read(struct zfcp_adapter *); | 125 | extern int zfcp_fsf_status_read(struct zfcp_qdio *); |
134 | extern int zfcp_status_read_refill(struct zfcp_adapter *adapter); | 126 | extern int zfcp_status_read_refill(struct zfcp_adapter *adapter); |
135 | extern int zfcp_fsf_send_ct(struct zfcp_send_ct *, mempool_t *, | 127 | extern int zfcp_fsf_send_ct(struct zfcp_send_ct *, mempool_t *); |
136 | struct zfcp_erp_action *); | ||
137 | extern int zfcp_fsf_send_els(struct zfcp_send_els *); | 128 | extern int zfcp_fsf_send_els(struct zfcp_send_els *); |
138 | extern int zfcp_fsf_send_fcp_command_task(struct zfcp_unit *, | 129 | extern int zfcp_fsf_send_fcp_command_task(struct zfcp_unit *, |
139 | struct scsi_cmnd *); | 130 | struct scsi_cmnd *); |
140 | extern void zfcp_fsf_req_complete(struct zfcp_fsf_req *); | ||
141 | extern void zfcp_fsf_req_free(struct zfcp_fsf_req *); | 131 | extern void zfcp_fsf_req_free(struct zfcp_fsf_req *); |
142 | extern struct zfcp_fsf_req *zfcp_fsf_send_fcp_ctm(struct zfcp_unit *, u8); | 132 | extern struct zfcp_fsf_req *zfcp_fsf_send_fcp_ctm(struct zfcp_unit *, u8); |
143 | extern struct zfcp_fsf_req *zfcp_fsf_abort_fcp_command(unsigned long, | 133 | extern struct zfcp_fsf_req *zfcp_fsf_abort_fcp_command(unsigned long, |
144 | struct zfcp_unit *); | 134 | struct zfcp_unit *); |
135 | extern void zfcp_fsf_reqid_check(struct zfcp_qdio *, int); | ||
145 | 136 | ||
146 | /* zfcp_qdio.c */ | 137 | /* zfcp_qdio.c */ |
147 | extern int zfcp_qdio_allocate(struct zfcp_adapter *); | 138 | extern int zfcp_qdio_setup(struct zfcp_adapter *); |
148 | extern void zfcp_qdio_free(struct zfcp_adapter *); | 139 | extern void zfcp_qdio_destroy(struct zfcp_qdio *); |
149 | extern int zfcp_qdio_send(struct zfcp_fsf_req *); | 140 | extern int zfcp_qdio_send(struct zfcp_qdio *, struct zfcp_queue_req *); |
150 | extern struct qdio_buffer_element *zfcp_qdio_sbale_req(struct zfcp_fsf_req *); | 141 | extern struct qdio_buffer_element |
151 | extern struct qdio_buffer_element *zfcp_qdio_sbale_curr(struct zfcp_fsf_req *); | 142 | *zfcp_qdio_sbale_req(struct zfcp_qdio *, struct zfcp_queue_req *); |
152 | extern int zfcp_qdio_sbals_from_sg(struct zfcp_fsf_req *, unsigned long, | 143 | extern struct qdio_buffer_element |
144 | *zfcp_qdio_sbale_curr(struct zfcp_qdio *, struct zfcp_queue_req *); | ||
145 | extern int zfcp_qdio_sbals_from_sg(struct zfcp_qdio *, | ||
146 | struct zfcp_queue_req *, unsigned long, | ||
153 | struct scatterlist *, int); | 147 | struct scatterlist *, int); |
154 | extern int zfcp_qdio_open(struct zfcp_adapter *); | 148 | extern int zfcp_qdio_open(struct zfcp_qdio *); |
155 | extern void zfcp_qdio_close(struct zfcp_adapter *); | 149 | extern void zfcp_qdio_close(struct zfcp_qdio *); |
156 | 150 | ||
157 | /* zfcp_scsi.c */ | 151 | /* zfcp_scsi.c */ |
158 | extern struct zfcp_data zfcp_data; | 152 | extern struct zfcp_data zfcp_data; |
diff --git a/drivers/s390/scsi/zfcp_fc.c b/drivers/s390/scsi/zfcp_fc.c index 47daebfa7e59..722f22de8753 100644 --- a/drivers/s390/scsi/zfcp_fc.c +++ b/drivers/s390/scsi/zfcp_fc.c | |||
@@ -25,14 +25,6 @@ static u32 rscn_range_mask[] = { | |||
25 | [RSCN_FABRIC_ADDRESS] = 0x000000, | 25 | [RSCN_FABRIC_ADDRESS] = 0x000000, |
26 | }; | 26 | }; |
27 | 27 | ||
28 | struct ct_iu_gpn_ft_req { | ||
29 | struct ct_hdr header; | ||
30 | u8 flags; | ||
31 | u8 domain_id_scope; | ||
32 | u8 area_id_scope; | ||
33 | u8 fc4_type; | ||
34 | } __attribute__ ((packed)); | ||
35 | |||
36 | struct gpn_ft_resp_acc { | 28 | struct gpn_ft_resp_acc { |
37 | u8 control; | 29 | u8 control; |
38 | u8 port_id[3]; | 30 | u8 port_id[3]; |
@@ -65,7 +57,7 @@ struct zfcp_fc_ns_handler_data { | |||
65 | unsigned long handler_data; | 57 | unsigned long handler_data; |
66 | }; | 58 | }; |
67 | 59 | ||
68 | static int zfcp_wka_port_get(struct zfcp_wka_port *wka_port) | 60 | static int zfcp_fc_wka_port_get(struct zfcp_wka_port *wka_port) |
69 | { | 61 | { |
70 | if (mutex_lock_interruptible(&wka_port->mutex)) | 62 | if (mutex_lock_interruptible(&wka_port->mutex)) |
71 | return -ERESTARTSYS; | 63 | return -ERESTARTSYS; |
@@ -90,7 +82,7 @@ static int zfcp_wka_port_get(struct zfcp_wka_port *wka_port) | |||
90 | return -EIO; | 82 | return -EIO; |
91 | } | 83 | } |
92 | 84 | ||
93 | static void zfcp_wka_port_offline(struct work_struct *work) | 85 | static void zfcp_fc_wka_port_offline(struct work_struct *work) |
94 | { | 86 | { |
95 | struct delayed_work *dw = to_delayed_work(work); | 87 | struct delayed_work *dw = to_delayed_work(work); |
96 | struct zfcp_wka_port *wka_port = | 88 | struct zfcp_wka_port *wka_port = |
@@ -110,7 +102,7 @@ out: | |||
110 | mutex_unlock(&wka_port->mutex); | 102 | mutex_unlock(&wka_port->mutex); |
111 | } | 103 | } |
112 | 104 | ||
113 | static void zfcp_wka_port_put(struct zfcp_wka_port *wka_port) | 105 | static void zfcp_fc_wka_port_put(struct zfcp_wka_port *wka_port) |
114 | { | 106 | { |
115 | if (atomic_dec_return(&wka_port->refcount) != 0) | 107 | if (atomic_dec_return(&wka_port->refcount) != 0) |
116 | return; | 108 | return; |
@@ -129,10 +121,10 @@ static void zfcp_fc_wka_port_init(struct zfcp_wka_port *wka_port, u32 d_id, | |||
129 | wka_port->status = ZFCP_WKA_PORT_OFFLINE; | 121 | wka_port->status = ZFCP_WKA_PORT_OFFLINE; |
130 | atomic_set(&wka_port->refcount, 0); | 122 | atomic_set(&wka_port->refcount, 0); |
131 | mutex_init(&wka_port->mutex); | 123 | mutex_init(&wka_port->mutex); |
132 | INIT_DELAYED_WORK(&wka_port->work, zfcp_wka_port_offline); | 124 | INIT_DELAYED_WORK(&wka_port->work, zfcp_fc_wka_port_offline); |
133 | } | 125 | } |
134 | 126 | ||
135 | void zfcp_fc_wka_port_force_offline(struct zfcp_wka_port *wka) | 127 | static void zfcp_fc_wka_port_force_offline(struct zfcp_wka_port *wka) |
136 | { | 128 | { |
137 | cancel_delayed_work_sync(&wka->work); | 129 | cancel_delayed_work_sync(&wka->work); |
138 | mutex_lock(&wka->mutex); | 130 | mutex_lock(&wka->mutex); |
@@ -140,15 +132,13 @@ void zfcp_fc_wka_port_force_offline(struct zfcp_wka_port *wka) | |||
140 | mutex_unlock(&wka->mutex); | 132 | mutex_unlock(&wka->mutex); |
141 | } | 133 | } |
142 | 134 | ||
143 | void zfcp_fc_wka_ports_init(struct zfcp_adapter *adapter) | 135 | void zfcp_fc_wka_ports_force_offline(struct zfcp_wka_ports *gs) |
144 | { | 136 | { |
145 | struct zfcp_wka_ports *gs = adapter->gs; | 137 | zfcp_fc_wka_port_force_offline(&gs->ms); |
146 | 138 | zfcp_fc_wka_port_force_offline(&gs->ts); | |
147 | zfcp_fc_wka_port_init(&gs->ms, FC_FID_MGMT_SERV, adapter); | 139 | zfcp_fc_wka_port_force_offline(&gs->ds); |
148 | zfcp_fc_wka_port_init(&gs->ts, FC_FID_TIME_SERV, adapter); | 140 | zfcp_fc_wka_port_force_offline(&gs->as); |
149 | zfcp_fc_wka_port_init(&gs->ds, FC_FID_DIR_SERV, adapter); | 141 | zfcp_fc_wka_port_force_offline(&gs->ks); |
150 | zfcp_fc_wka_port_init(&gs->as, FC_FID_ALIASES, adapter); | ||
151 | zfcp_fc_wka_port_init(&gs->ks, FC_FID_SEC_KEY, adapter); | ||
152 | } | 142 | } |
153 | 143 | ||
154 | static void _zfcp_fc_incoming_rscn(struct zfcp_fsf_req *fsf_req, u32 range, | 144 | static void _zfcp_fc_incoming_rscn(struct zfcp_fsf_req *fsf_req, u32 range, |
@@ -160,7 +150,7 @@ static void _zfcp_fc_incoming_rscn(struct zfcp_fsf_req *fsf_req, u32 range, | |||
160 | read_lock_irqsave(&zfcp_data.config_lock, flags); | 150 | read_lock_irqsave(&zfcp_data.config_lock, flags); |
161 | list_for_each_entry(port, &fsf_req->adapter->port_list_head, list) { | 151 | list_for_each_entry(port, &fsf_req->adapter->port_list_head, list) { |
162 | if ((port->d_id & range) == (elem->nport_did & range)) | 152 | if ((port->d_id & range) == (elem->nport_did & range)) |
163 | zfcp_test_link(port); | 153 | zfcp_fc_test_link(port); |
164 | if (!port->d_id) | 154 | if (!port->d_id) |
165 | zfcp_erp_port_reopen(port, | 155 | zfcp_erp_port_reopen(port, |
166 | ZFCP_STATUS_COMMON_ERP_FAILED, | 156 | ZFCP_STATUS_COMMON_ERP_FAILED, |
@@ -241,7 +231,7 @@ void zfcp_fc_incoming_els(struct zfcp_fsf_req *fsf_req) | |||
241 | (struct fsf_status_read_buffer *) fsf_req->data; | 231 | (struct fsf_status_read_buffer *) fsf_req->data; |
242 | unsigned int els_type = status_buffer->payload.data[0]; | 232 | unsigned int els_type = status_buffer->payload.data[0]; |
243 | 233 | ||
244 | zfcp_san_dbf_event_incoming_els(fsf_req); | 234 | zfcp_dbf_san_incoming_els(fsf_req); |
245 | if (els_type == LS_PLOGI) | 235 | if (els_type == LS_PLOGI) |
246 | zfcp_fc_incoming_plogi(fsf_req); | 236 | zfcp_fc_incoming_plogi(fsf_req); |
247 | else if (els_type == LS_LOGO) | 237 | else if (els_type == LS_LOGO) |
@@ -281,19 +271,18 @@ static void zfcp_fc_ns_gid_pn_eval(unsigned long data) | |||
281 | port->d_id = ct_iu_resp->d_id & ZFCP_DID_MASK; | 271 | port->d_id = ct_iu_resp->d_id & ZFCP_DID_MASK; |
282 | } | 272 | } |
283 | 273 | ||
284 | int static zfcp_fc_ns_gid_pn_request(struct zfcp_erp_action *erp_action, | 274 | static int zfcp_fc_ns_gid_pn_request(struct zfcp_port *port, |
285 | struct zfcp_gid_pn_data *gid_pn) | 275 | struct zfcp_gid_pn_data *gid_pn) |
286 | { | 276 | { |
287 | struct zfcp_adapter *adapter = erp_action->adapter; | 277 | struct zfcp_adapter *adapter = port->adapter; |
288 | struct zfcp_fc_ns_handler_data compl_rec; | 278 | struct zfcp_fc_ns_handler_data compl_rec; |
289 | int ret; | 279 | int ret; |
290 | 280 | ||
291 | /* setup parameters for send generic command */ | 281 | /* setup parameters for send generic command */ |
292 | gid_pn->port = erp_action->port; | 282 | gid_pn->port = port; |
293 | gid_pn->ct.wka_port = &adapter->gs->ds; | 283 | gid_pn->ct.wka_port = &adapter->gs->ds; |
294 | gid_pn->ct.handler = zfcp_fc_ns_handler; | 284 | gid_pn->ct.handler = zfcp_fc_ns_handler; |
295 | gid_pn->ct.handler_data = (unsigned long) &compl_rec; | 285 | gid_pn->ct.handler_data = (unsigned long) &compl_rec; |
296 | gid_pn->ct.timeout = ZFCP_NS_GID_PN_TIMEOUT; | ||
297 | gid_pn->ct.req = &gid_pn->req; | 286 | gid_pn->ct.req = &gid_pn->req; |
298 | gid_pn->ct.resp = &gid_pn->resp; | 287 | gid_pn->ct.resp = &gid_pn->resp; |
299 | sg_init_one(&gid_pn->req, &gid_pn->ct_iu_req, | 288 | sg_init_one(&gid_pn->req, &gid_pn->ct_iu_req, |
@@ -308,13 +297,12 @@ int static zfcp_fc_ns_gid_pn_request(struct zfcp_erp_action *erp_action, | |||
308 | gid_pn->ct_iu_req.header.options = ZFCP_CT_SYNCHRONOUS; | 297 | gid_pn->ct_iu_req.header.options = ZFCP_CT_SYNCHRONOUS; |
309 | gid_pn->ct_iu_req.header.cmd_rsp_code = ZFCP_CT_GID_PN; | 298 | gid_pn->ct_iu_req.header.cmd_rsp_code = ZFCP_CT_GID_PN; |
310 | gid_pn->ct_iu_req.header.max_res_size = ZFCP_CT_SIZE_ONE_PAGE / 4; | 299 | gid_pn->ct_iu_req.header.max_res_size = ZFCP_CT_SIZE_ONE_PAGE / 4; |
311 | gid_pn->ct_iu_req.wwpn = erp_action->port->wwpn; | 300 | gid_pn->ct_iu_req.wwpn = port->wwpn; |
312 | 301 | ||
313 | init_completion(&compl_rec.done); | 302 | init_completion(&compl_rec.done); |
314 | compl_rec.handler = zfcp_fc_ns_gid_pn_eval; | 303 | compl_rec.handler = zfcp_fc_ns_gid_pn_eval; |
315 | compl_rec.handler_data = (unsigned long) gid_pn; | 304 | compl_rec.handler_data = (unsigned long) gid_pn; |
316 | ret = zfcp_fsf_send_ct(&gid_pn->ct, adapter->pool.fsf_req_erp, | 305 | ret = zfcp_fsf_send_ct(&gid_pn->ct, adapter->pool.gid_pn_req); |
317 | erp_action); | ||
318 | if (!ret) | 306 | if (!ret) |
319 | wait_for_completion(&compl_rec.done); | 307 | wait_for_completion(&compl_rec.done); |
320 | return ret; | 308 | return ret; |
@@ -322,33 +310,56 @@ int static zfcp_fc_ns_gid_pn_request(struct zfcp_erp_action *erp_action, | |||
322 | 310 | ||
323 | /** | 311 | /** |
324 | * zfcp_fc_ns_gid_pn_request - initiate GID_PN nameserver request | 312 | * zfcp_fc_ns_gid_pn_request - initiate GID_PN nameserver request |
325 | * @erp_action: pointer to zfcp_erp_action where GID_PN request is needed | 313 | * @port: port where GID_PN request is needed |
326 | * return: -ENOMEM on error, 0 otherwise | 314 | * return: -ENOMEM on error, 0 otherwise |
327 | */ | 315 | */ |
328 | int zfcp_fc_ns_gid_pn(struct zfcp_erp_action *erp_action) | 316 | static int zfcp_fc_ns_gid_pn(struct zfcp_port *port) |
329 | { | 317 | { |
330 | int ret; | 318 | int ret; |
331 | struct zfcp_gid_pn_data *gid_pn; | 319 | struct zfcp_gid_pn_data *gid_pn; |
332 | struct zfcp_adapter *adapter = erp_action->adapter; | 320 | struct zfcp_adapter *adapter = port->adapter; |
333 | 321 | ||
334 | gid_pn = mempool_alloc(adapter->pool.data_gid_pn, GFP_ATOMIC); | 322 | gid_pn = mempool_alloc(adapter->pool.gid_pn_data, GFP_ATOMIC); |
335 | if (!gid_pn) | 323 | if (!gid_pn) |
336 | return -ENOMEM; | 324 | return -ENOMEM; |
337 | 325 | ||
338 | memset(gid_pn, 0, sizeof(*gid_pn)); | 326 | memset(gid_pn, 0, sizeof(*gid_pn)); |
339 | 327 | ||
340 | ret = zfcp_wka_port_get(&adapter->gs->ds); | 328 | ret = zfcp_fc_wka_port_get(&adapter->gs->ds); |
341 | if (ret) | 329 | if (ret) |
342 | goto out; | 330 | goto out; |
343 | 331 | ||
344 | ret = zfcp_fc_ns_gid_pn_request(erp_action, gid_pn); | 332 | ret = zfcp_fc_ns_gid_pn_request(port, gid_pn); |
345 | 333 | ||
346 | zfcp_wka_port_put(&adapter->gs->ds); | 334 | zfcp_fc_wka_port_put(&adapter->gs->ds); |
347 | out: | 335 | out: |
348 | mempool_free(gid_pn, adapter->pool.data_gid_pn); | 336 | mempool_free(gid_pn, adapter->pool.gid_pn_data); |
349 | return ret; | 337 | return ret; |
350 | } | 338 | } |
351 | 339 | ||
340 | void zfcp_fc_port_did_lookup(struct work_struct *work) | ||
341 | { | ||
342 | int ret; | ||
343 | struct zfcp_port *port = container_of(work, struct zfcp_port, | ||
344 | gid_pn_work); | ||
345 | |||
346 | ret = zfcp_fc_ns_gid_pn(port); | ||
347 | if (ret) { | ||
348 | /* could not issue gid_pn for some reason */ | ||
349 | zfcp_erp_adapter_reopen(port->adapter, 0, "fcgpn_1", NULL); | ||
350 | goto out; | ||
351 | } | ||
352 | |||
353 | if (!port->d_id) { | ||
354 | zfcp_erp_port_failed(port, "fcgpn_2", NULL); | ||
355 | goto out; | ||
356 | } | ||
357 | |||
358 | zfcp_erp_port_reopen(port, 0, "fcgpn_3", NULL); | ||
359 | out: | ||
360 | zfcp_port_put(port); | ||
361 | } | ||
362 | |||
352 | /** | 363 | /** |
353 | * zfcp_fc_plogi_evaluate - evaluate PLOGI playload | 364 | * zfcp_fc_plogi_evaluate - evaluate PLOGI playload |
354 | * @port: zfcp_port structure | 365 | * @port: zfcp_port structure |
@@ -404,6 +415,7 @@ static void zfcp_fc_adisc_handler(unsigned long data) | |||
404 | /* port is good, unblock rport without going through erp */ | 415 | /* port is good, unblock rport without going through erp */ |
405 | zfcp_scsi_schedule_rport_register(port); | 416 | zfcp_scsi_schedule_rport_register(port); |
406 | out: | 417 | out: |
418 | atomic_clear_mask(ZFCP_STATUS_PORT_LINK_TEST, &port->status); | ||
407 | zfcp_port_put(port); | 419 | zfcp_port_put(port); |
408 | kfree(adisc); | 420 | kfree(adisc); |
409 | } | 421 | } |
@@ -450,28 +462,36 @@ void zfcp_fc_link_test_work(struct work_struct *work) | |||
450 | port->rport_task = RPORT_DEL; | 462 | port->rport_task = RPORT_DEL; |
451 | zfcp_scsi_rport_work(&port->rport_work); | 463 | zfcp_scsi_rport_work(&port->rport_work); |
452 | 464 | ||
465 | /* only issue one test command at one time per port */ | ||
466 | if (atomic_read(&port->status) & ZFCP_STATUS_PORT_LINK_TEST) | ||
467 | goto out; | ||
468 | |||
469 | atomic_set_mask(ZFCP_STATUS_PORT_LINK_TEST, &port->status); | ||
470 | |||
453 | retval = zfcp_fc_adisc(port); | 471 | retval = zfcp_fc_adisc(port); |
454 | if (retval == 0) | 472 | if (retval == 0) |
455 | return; | 473 | return; |
456 | 474 | ||
457 | /* send of ADISC was not possible */ | 475 | /* send of ADISC was not possible */ |
476 | atomic_clear_mask(ZFCP_STATUS_PORT_LINK_TEST, &port->status); | ||
458 | zfcp_erp_port_forced_reopen(port, 0, "fcltwk1", NULL); | 477 | zfcp_erp_port_forced_reopen(port, 0, "fcltwk1", NULL); |
459 | 478 | ||
479 | out: | ||
460 | zfcp_port_put(port); | 480 | zfcp_port_put(port); |
461 | } | 481 | } |
462 | 482 | ||
463 | /** | 483 | /** |
464 | * zfcp_test_link - lightweight link test procedure | 484 | * zfcp_fc_test_link - lightweight link test procedure |
465 | * @port: port to be tested | 485 | * @port: port to be tested |
466 | * | 486 | * |
467 | * Test status of a link to a remote port using the ELS command ADISC. | 487 | * Test status of a link to a remote port using the ELS command ADISC. |
468 | * If there is a problem with the remote port, error recovery steps | 488 | * If there is a problem with the remote port, error recovery steps |
469 | * will be triggered. | 489 | * will be triggered. |
470 | */ | 490 | */ |
471 | void zfcp_test_link(struct zfcp_port *port) | 491 | void zfcp_fc_test_link(struct zfcp_port *port) |
472 | { | 492 | { |
473 | zfcp_port_get(port); | 493 | zfcp_port_get(port); |
474 | if (!queue_work(zfcp_data.work_queue, &port->test_link_work)) | 494 | if (!queue_work(port->adapter->work_queue, &port->test_link_work)) |
475 | zfcp_port_put(port); | 495 | zfcp_port_put(port); |
476 | } | 496 | } |
477 | 497 | ||
@@ -479,7 +499,7 @@ static void zfcp_free_sg_env(struct zfcp_gpn_ft *gpn_ft, int buf_num) | |||
479 | { | 499 | { |
480 | struct scatterlist *sg = &gpn_ft->sg_req; | 500 | struct scatterlist *sg = &gpn_ft->sg_req; |
481 | 501 | ||
482 | kfree(sg_virt(sg)); /* free request buffer */ | 502 | kmem_cache_free(zfcp_data.gpn_ft_cache, sg_virt(sg)); |
483 | zfcp_sg_free_table(gpn_ft->sg_resp, buf_num); | 503 | zfcp_sg_free_table(gpn_ft->sg_resp, buf_num); |
484 | 504 | ||
485 | kfree(gpn_ft); | 505 | kfree(gpn_ft); |
@@ -494,7 +514,7 @@ static struct zfcp_gpn_ft *zfcp_alloc_sg_env(int buf_num) | |||
494 | if (!gpn_ft) | 514 | if (!gpn_ft) |
495 | return NULL; | 515 | return NULL; |
496 | 516 | ||
497 | req = kzalloc(sizeof(struct ct_iu_gpn_ft_req), GFP_KERNEL); | 517 | req = kmem_cache_alloc(zfcp_data.gpn_ft_cache, GFP_KERNEL); |
498 | if (!req) { | 518 | if (!req) { |
499 | kfree(gpn_ft); | 519 | kfree(gpn_ft); |
500 | gpn_ft = NULL; | 520 | gpn_ft = NULL; |
@@ -511,9 +531,8 @@ out: | |||
511 | } | 531 | } |
512 | 532 | ||
513 | 533 | ||
514 | static int zfcp_scan_issue_gpn_ft(struct zfcp_gpn_ft *gpn_ft, | 534 | static int zfcp_fc_send_gpn_ft(struct zfcp_gpn_ft *gpn_ft, |
515 | struct zfcp_adapter *adapter, | 535 | struct zfcp_adapter *adapter, int max_bytes) |
516 | int max_bytes) | ||
517 | { | 536 | { |
518 | struct zfcp_send_ct *ct = &gpn_ft->ct; | 537 | struct zfcp_send_ct *ct = &gpn_ft->ct; |
519 | struct ct_iu_gpn_ft_req *req = sg_virt(&gpn_ft->sg_req); | 538 | struct ct_iu_gpn_ft_req *req = sg_virt(&gpn_ft->sg_req); |
@@ -536,19 +555,18 @@ static int zfcp_scan_issue_gpn_ft(struct zfcp_gpn_ft *gpn_ft, | |||
536 | ct->wka_port = &adapter->gs->ds; | 555 | ct->wka_port = &adapter->gs->ds; |
537 | ct->handler = zfcp_fc_ns_handler; | 556 | ct->handler = zfcp_fc_ns_handler; |
538 | ct->handler_data = (unsigned long)&compl_rec; | 557 | ct->handler_data = (unsigned long)&compl_rec; |
539 | ct->timeout = 10; | ||
540 | ct->req = &gpn_ft->sg_req; | 558 | ct->req = &gpn_ft->sg_req; |
541 | ct->resp = gpn_ft->sg_resp; | 559 | ct->resp = gpn_ft->sg_resp; |
542 | 560 | ||
543 | init_completion(&compl_rec.done); | 561 | init_completion(&compl_rec.done); |
544 | compl_rec.handler = NULL; | 562 | compl_rec.handler = NULL; |
545 | ret = zfcp_fsf_send_ct(ct, NULL, NULL); | 563 | ret = zfcp_fsf_send_ct(ct, NULL); |
546 | if (!ret) | 564 | if (!ret) |
547 | wait_for_completion(&compl_rec.done); | 565 | wait_for_completion(&compl_rec.done); |
548 | return ret; | 566 | return ret; |
549 | } | 567 | } |
550 | 568 | ||
551 | static void zfcp_validate_port(struct zfcp_port *port) | 569 | static void zfcp_fc_validate_port(struct zfcp_port *port) |
552 | { | 570 | { |
553 | struct zfcp_adapter *adapter = port->adapter; | 571 | struct zfcp_adapter *adapter = port->adapter; |
554 | 572 | ||
@@ -568,7 +586,7 @@ static void zfcp_validate_port(struct zfcp_port *port) | |||
568 | zfcp_port_dequeue(port); | 586 | zfcp_port_dequeue(port); |
569 | } | 587 | } |
570 | 588 | ||
571 | static int zfcp_scan_eval_gpn_ft(struct zfcp_gpn_ft *gpn_ft, int max_entries) | 589 | static int zfcp_fc_eval_gpn_ft(struct zfcp_gpn_ft *gpn_ft, int max_entries) |
572 | { | 590 | { |
573 | struct zfcp_send_ct *ct = &gpn_ft->ct; | 591 | struct zfcp_send_ct *ct = &gpn_ft->ct; |
574 | struct scatterlist *sg = gpn_ft->sg_resp; | 592 | struct scatterlist *sg = gpn_ft->sg_resp; |
@@ -595,7 +613,7 @@ static int zfcp_scan_eval_gpn_ft(struct zfcp_gpn_ft *gpn_ft, int max_entries) | |||
595 | return -E2BIG; | 613 | return -E2BIG; |
596 | } | 614 | } |
597 | 615 | ||
598 | down(&zfcp_data.config_sema); | 616 | mutex_lock(&zfcp_data.config_mutex); |
599 | 617 | ||
600 | /* first entry is the header */ | 618 | /* first entry is the header */ |
601 | for (x = 1; x < max_entries && !last; x++) { | 619 | for (x = 1; x < max_entries && !last; x++) { |
@@ -628,16 +646,16 @@ static int zfcp_scan_eval_gpn_ft(struct zfcp_gpn_ft *gpn_ft, int max_entries) | |||
628 | 646 | ||
629 | zfcp_erp_wait(adapter); | 647 | zfcp_erp_wait(adapter); |
630 | list_for_each_entry_safe(port, tmp, &adapter->port_list_head, list) | 648 | list_for_each_entry_safe(port, tmp, &adapter->port_list_head, list) |
631 | zfcp_validate_port(port); | 649 | zfcp_fc_validate_port(port); |
632 | up(&zfcp_data.config_sema); | 650 | mutex_unlock(&zfcp_data.config_mutex); |
633 | return ret; | 651 | return ret; |
634 | } | 652 | } |
635 | 653 | ||
636 | /** | 654 | /** |
637 | * zfcp_scan_ports - scan remote ports and attach new ports | 655 | * zfcp_fc_scan_ports - scan remote ports and attach new ports |
638 | * @adapter: pointer to struct zfcp_adapter | 656 | * @adapter: pointer to struct zfcp_adapter |
639 | */ | 657 | */ |
640 | int zfcp_scan_ports(struct zfcp_adapter *adapter) | 658 | int zfcp_fc_scan_ports(struct zfcp_adapter *adapter) |
641 | { | 659 | { |
642 | int ret, i; | 660 | int ret, i; |
643 | struct zfcp_gpn_ft *gpn_ft; | 661 | struct zfcp_gpn_ft *gpn_ft; |
@@ -652,7 +670,7 @@ int zfcp_scan_ports(struct zfcp_adapter *adapter) | |||
652 | fc_host_port_type(adapter->scsi_host) != FC_PORTTYPE_NPIV) | 670 | fc_host_port_type(adapter->scsi_host) != FC_PORTTYPE_NPIV) |
653 | return 0; | 671 | return 0; |
654 | 672 | ||
655 | ret = zfcp_wka_port_get(&adapter->gs->ds); | 673 | ret = zfcp_fc_wka_port_get(&adapter->gs->ds); |
656 | if (ret) | 674 | if (ret) |
657 | return ret; | 675 | return ret; |
658 | 676 | ||
@@ -663,9 +681,9 @@ int zfcp_scan_ports(struct zfcp_adapter *adapter) | |||
663 | } | 681 | } |
664 | 682 | ||
665 | for (i = 0; i < 3; i++) { | 683 | for (i = 0; i < 3; i++) { |
666 | ret = zfcp_scan_issue_gpn_ft(gpn_ft, adapter, max_bytes); | 684 | ret = zfcp_fc_send_gpn_ft(gpn_ft, adapter, max_bytes); |
667 | if (!ret) { | 685 | if (!ret) { |
668 | ret = zfcp_scan_eval_gpn_ft(gpn_ft, max_entries); | 686 | ret = zfcp_fc_eval_gpn_ft(gpn_ft, max_entries); |
669 | if (ret == -EAGAIN) | 687 | if (ret == -EAGAIN) |
670 | ssleep(1); | 688 | ssleep(1); |
671 | else | 689 | else |
@@ -674,14 +692,14 @@ int zfcp_scan_ports(struct zfcp_adapter *adapter) | |||
674 | } | 692 | } |
675 | zfcp_free_sg_env(gpn_ft, buf_num); | 693 | zfcp_free_sg_env(gpn_ft, buf_num); |
676 | out: | 694 | out: |
677 | zfcp_wka_port_put(&adapter->gs->ds); | 695 | zfcp_fc_wka_port_put(&adapter->gs->ds); |
678 | return ret; | 696 | return ret; |
679 | } | 697 | } |
680 | 698 | ||
681 | 699 | ||
682 | void _zfcp_scan_ports_later(struct work_struct *work) | 700 | void _zfcp_fc_scan_ports_later(struct work_struct *work) |
683 | { | 701 | { |
684 | zfcp_scan_ports(container_of(work, struct zfcp_adapter, scan_work)); | 702 | zfcp_fc_scan_ports(container_of(work, struct zfcp_adapter, scan_work)); |
685 | } | 703 | } |
686 | 704 | ||
687 | struct zfcp_els_fc_job { | 705 | struct zfcp_els_fc_job { |
@@ -732,7 +750,7 @@ int zfcp_fc_execute_els_fc_job(struct fc_bsg_job *job) | |||
732 | els_fc_job->els.adapter = adapter; | 750 | els_fc_job->els.adapter = adapter; |
733 | if (rport) { | 751 | if (rport) { |
734 | read_lock_irq(&zfcp_data.config_lock); | 752 | read_lock_irq(&zfcp_data.config_lock); |
735 | port = rport->dd_data; | 753 | port = zfcp_get_port_by_wwpn(adapter, rport->port_name); |
736 | if (port) | 754 | if (port) |
737 | els_fc_job->els.d_id = port->d_id; | 755 | els_fc_job->els.d_id = port->d_id; |
738 | read_unlock_irq(&zfcp_data.config_lock); | 756 | read_unlock_irq(&zfcp_data.config_lock); |
@@ -771,7 +789,7 @@ static void zfcp_fc_generic_ct_handler(unsigned long data) | |||
771 | job->state_flags = FC_RQST_STATE_DONE; | 789 | job->state_flags = FC_RQST_STATE_DONE; |
772 | job->job_done(job); | 790 | job->job_done(job); |
773 | 791 | ||
774 | zfcp_wka_port_put(ct_fc_job->ct.wka_port); | 792 | zfcp_fc_wka_port_put(ct_fc_job->ct.wka_port); |
775 | 793 | ||
776 | kfree(ct_fc_job); | 794 | kfree(ct_fc_job); |
777 | } | 795 | } |
@@ -817,7 +835,7 @@ int zfcp_fc_execute_ct_fc_job(struct fc_bsg_job *job) | |||
817 | return -EINVAL; /* no such service */ | 835 | return -EINVAL; /* no such service */ |
818 | } | 836 | } |
819 | 837 | ||
820 | ret = zfcp_wka_port_get(ct_fc_job->ct.wka_port); | 838 | ret = zfcp_fc_wka_port_get(ct_fc_job->ct.wka_port); |
821 | if (ret) { | 839 | if (ret) { |
822 | kfree(ct_fc_job); | 840 | kfree(ct_fc_job); |
823 | return ret; | 841 | return ret; |
@@ -825,16 +843,40 @@ int zfcp_fc_execute_ct_fc_job(struct fc_bsg_job *job) | |||
825 | 843 | ||
826 | ct_fc_job->ct.req = job->request_payload.sg_list; | 844 | ct_fc_job->ct.req = job->request_payload.sg_list; |
827 | ct_fc_job->ct.resp = job->reply_payload.sg_list; | 845 | ct_fc_job->ct.resp = job->reply_payload.sg_list; |
828 | ct_fc_job->ct.timeout = ZFCP_FSF_REQUEST_TIMEOUT; | ||
829 | ct_fc_job->ct.handler = zfcp_fc_generic_ct_handler; | 846 | ct_fc_job->ct.handler = zfcp_fc_generic_ct_handler; |
830 | ct_fc_job->ct.handler_data = (unsigned long) ct_fc_job; | 847 | ct_fc_job->ct.handler_data = (unsigned long) ct_fc_job; |
831 | ct_fc_job->ct.completion = NULL; | 848 | ct_fc_job->ct.completion = NULL; |
832 | ct_fc_job->job = job; | 849 | ct_fc_job->job = job; |
833 | 850 | ||
834 | ret = zfcp_fsf_send_ct(&ct_fc_job->ct, NULL, NULL); | 851 | ret = zfcp_fsf_send_ct(&ct_fc_job->ct, NULL); |
835 | if (ret) { | 852 | if (ret) { |
836 | kfree(ct_fc_job); | 853 | kfree(ct_fc_job); |
837 | zfcp_wka_port_put(ct_fc_job->ct.wka_port); | 854 | zfcp_fc_wka_port_put(ct_fc_job->ct.wka_port); |
838 | } | 855 | } |
839 | return ret; | 856 | return ret; |
840 | } | 857 | } |
858 | |||
859 | int zfcp_fc_gs_setup(struct zfcp_adapter *adapter) | ||
860 | { | ||
861 | struct zfcp_wka_ports *wka_ports; | ||
862 | |||
863 | wka_ports = kzalloc(sizeof(struct zfcp_wka_ports), GFP_KERNEL); | ||
864 | if (!wka_ports) | ||
865 | return -ENOMEM; | ||
866 | |||
867 | adapter->gs = wka_ports; | ||
868 | zfcp_fc_wka_port_init(&wka_ports->ms, FC_FID_MGMT_SERV, adapter); | ||
869 | zfcp_fc_wka_port_init(&wka_ports->ts, FC_FID_TIME_SERV, adapter); | ||
870 | zfcp_fc_wka_port_init(&wka_ports->ds, FC_FID_DIR_SERV, adapter); | ||
871 | zfcp_fc_wka_port_init(&wka_ports->as, FC_FID_ALIASES, adapter); | ||
872 | zfcp_fc_wka_port_init(&wka_ports->ks, FC_FID_SEC_KEY, adapter); | ||
873 | |||
874 | return 0; | ||
875 | } | ||
876 | |||
877 | void zfcp_fc_gs_destroy(struct zfcp_adapter *adapter) | ||
878 | { | ||
879 | kfree(adapter->gs); | ||
880 | adapter->gs = NULL; | ||
881 | } | ||
882 | |||
diff --git a/drivers/s390/scsi/zfcp_fsf.c b/drivers/s390/scsi/zfcp_fsf.c index 47795fbf081f..f09c863dc6bd 100644 --- a/drivers/s390/scsi/zfcp_fsf.c +++ b/drivers/s390/scsi/zfcp_fsf.c | |||
@@ -11,9 +11,7 @@ | |||
11 | 11 | ||
12 | #include <linux/blktrace_api.h> | 12 | #include <linux/blktrace_api.h> |
13 | #include "zfcp_ext.h" | 13 | #include "zfcp_ext.h" |
14 | 14 | #include "zfcp_dbf.h" | |
15 | #define ZFCP_REQ_AUTO_CLEANUP 0x00000002 | ||
16 | #define ZFCP_REQ_NO_QTCB 0x00000008 | ||
17 | 15 | ||
18 | static void zfcp_fsf_request_timeout_handler(unsigned long data) | 16 | static void zfcp_fsf_request_timeout_handler(unsigned long data) |
19 | { | 17 | { |
@@ -111,43 +109,15 @@ static void zfcp_fsf_class_not_supp(struct zfcp_fsf_req *req) | |||
111 | void zfcp_fsf_req_free(struct zfcp_fsf_req *req) | 109 | void zfcp_fsf_req_free(struct zfcp_fsf_req *req) |
112 | { | 110 | { |
113 | if (likely(req->pool)) { | 111 | if (likely(req->pool)) { |
112 | if (likely(req->qtcb)) | ||
113 | mempool_free(req->qtcb, req->adapter->pool.qtcb_pool); | ||
114 | mempool_free(req, req->pool); | 114 | mempool_free(req, req->pool); |
115 | return; | 115 | return; |
116 | } | 116 | } |
117 | 117 | ||
118 | if (req->qtcb) { | 118 | if (likely(req->qtcb)) |
119 | kmem_cache_free(zfcp_data.fsf_req_qtcb_cache, req); | 119 | kmem_cache_free(zfcp_data.qtcb_cache, req->qtcb); |
120 | return; | 120 | kfree(req); |
121 | } | ||
122 | } | ||
123 | |||
124 | /** | ||
125 | * zfcp_fsf_req_dismiss_all - dismiss all fsf requests | ||
126 | * @adapter: pointer to struct zfcp_adapter | ||
127 | * | ||
128 | * Never ever call this without shutting down the adapter first. | ||
129 | * Otherwise the adapter would continue using and corrupting s390 storage. | ||
130 | * Included BUG_ON() call to ensure this is done. | ||
131 | * ERP is supposed to be the only user of this function. | ||
132 | */ | ||
133 | void zfcp_fsf_req_dismiss_all(struct zfcp_adapter *adapter) | ||
134 | { | ||
135 | struct zfcp_fsf_req *req, *tmp; | ||
136 | unsigned long flags; | ||
137 | LIST_HEAD(remove_queue); | ||
138 | unsigned int i; | ||
139 | |||
140 | BUG_ON(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP); | ||
141 | spin_lock_irqsave(&adapter->req_list_lock, flags); | ||
142 | for (i = 0; i < REQUEST_LIST_SIZE; i++) | ||
143 | list_splice_init(&adapter->req_list[i], &remove_queue); | ||
144 | spin_unlock_irqrestore(&adapter->req_list_lock, flags); | ||
145 | |||
146 | list_for_each_entry_safe(req, tmp, &remove_queue, list) { | ||
147 | list_del(&req->list); | ||
148 | req->status |= ZFCP_STATUS_FSFREQ_DISMISSED; | ||
149 | zfcp_fsf_req_complete(req); | ||
150 | } | ||
151 | } | 121 | } |
152 | 122 | ||
153 | static void zfcp_fsf_status_read_port_closed(struct zfcp_fsf_req *req) | 123 | static void zfcp_fsf_status_read_port_closed(struct zfcp_fsf_req *req) |
@@ -278,13 +248,13 @@ static void zfcp_fsf_status_read_handler(struct zfcp_fsf_req *req) | |||
278 | struct fsf_status_read_buffer *sr_buf = req->data; | 248 | struct fsf_status_read_buffer *sr_buf = req->data; |
279 | 249 | ||
280 | if (req->status & ZFCP_STATUS_FSFREQ_DISMISSED) { | 250 | if (req->status & ZFCP_STATUS_FSFREQ_DISMISSED) { |
281 | zfcp_hba_dbf_event_fsf_unsol("dism", adapter, sr_buf); | 251 | zfcp_dbf_hba_fsf_unsol("dism", adapter->dbf, sr_buf); |
282 | mempool_free(sr_buf, adapter->pool.data_status_read); | 252 | mempool_free(sr_buf, adapter->pool.status_read_data); |
283 | zfcp_fsf_req_free(req); | 253 | zfcp_fsf_req_free(req); |
284 | return; | 254 | return; |
285 | } | 255 | } |
286 | 256 | ||
287 | zfcp_hba_dbf_event_fsf_unsol("read", adapter, sr_buf); | 257 | zfcp_dbf_hba_fsf_unsol("read", adapter->dbf, sr_buf); |
288 | 258 | ||
289 | switch (sr_buf->status_type) { | 259 | switch (sr_buf->status_type) { |
290 | case FSF_STATUS_READ_PORT_CLOSED: | 260 | case FSF_STATUS_READ_PORT_CLOSED: |
@@ -299,7 +269,7 @@ static void zfcp_fsf_status_read_handler(struct zfcp_fsf_req *req) | |||
299 | dev_warn(&adapter->ccw_device->dev, | 269 | dev_warn(&adapter->ccw_device->dev, |
300 | "The error threshold for checksum statistics " | 270 | "The error threshold for checksum statistics " |
301 | "has been exceeded\n"); | 271 | "has been exceeded\n"); |
302 | zfcp_hba_dbf_event_berr(adapter, req); | 272 | zfcp_dbf_hba_berr(adapter->dbf, req); |
303 | break; | 273 | break; |
304 | case FSF_STATUS_READ_LINK_DOWN: | 274 | case FSF_STATUS_READ_LINK_DOWN: |
305 | zfcp_fsf_status_read_link_down(req); | 275 | zfcp_fsf_status_read_link_down(req); |
@@ -331,11 +301,11 @@ static void zfcp_fsf_status_read_handler(struct zfcp_fsf_req *req) | |||
331 | break; | 301 | break; |
332 | } | 302 | } |
333 | 303 | ||
334 | mempool_free(sr_buf, adapter->pool.data_status_read); | 304 | mempool_free(sr_buf, adapter->pool.status_read_data); |
335 | zfcp_fsf_req_free(req); | 305 | zfcp_fsf_req_free(req); |
336 | 306 | ||
337 | atomic_inc(&adapter->stat_miss); | 307 | atomic_inc(&adapter->stat_miss); |
338 | queue_work(zfcp_data.work_queue, &adapter->stat_work); | 308 | queue_work(adapter->work_queue, &adapter->stat_work); |
339 | } | 309 | } |
340 | 310 | ||
341 | static void zfcp_fsf_fsfstatus_qual_eval(struct zfcp_fsf_req *req) | 311 | static void zfcp_fsf_fsfstatus_qual_eval(struct zfcp_fsf_req *req) |
@@ -385,7 +355,7 @@ static void zfcp_fsf_protstatus_eval(struct zfcp_fsf_req *req) | |||
385 | struct fsf_qtcb *qtcb = req->qtcb; | 355 | struct fsf_qtcb *qtcb = req->qtcb; |
386 | union fsf_prot_status_qual *psq = &qtcb->prefix.prot_status_qual; | 356 | union fsf_prot_status_qual *psq = &qtcb->prefix.prot_status_qual; |
387 | 357 | ||
388 | zfcp_hba_dbf_event_fsf_response(req); | 358 | zfcp_dbf_hba_fsf_response(req); |
389 | 359 | ||
390 | if (req->status & ZFCP_STATUS_FSFREQ_DISMISSED) { | 360 | if (req->status & ZFCP_STATUS_FSFREQ_DISMISSED) { |
391 | req->status |= ZFCP_STATUS_FSFREQ_ERROR | | 361 | req->status |= ZFCP_STATUS_FSFREQ_ERROR | |
@@ -458,7 +428,7 @@ static void zfcp_fsf_protstatus_eval(struct zfcp_fsf_req *req) | |||
458 | * is called to process the completion status and trigger further | 428 | * is called to process the completion status and trigger further |
459 | * events related to the FSF request. | 429 | * events related to the FSF request. |
460 | */ | 430 | */ |
461 | void zfcp_fsf_req_complete(struct zfcp_fsf_req *req) | 431 | static void zfcp_fsf_req_complete(struct zfcp_fsf_req *req) |
462 | { | 432 | { |
463 | if (unlikely(req->fsf_command == FSF_QTCB_UNSOLICITED_STATUS)) { | 433 | if (unlikely(req->fsf_command == FSF_QTCB_UNSOLICITED_STATUS)) { |
464 | zfcp_fsf_status_read_handler(req); | 434 | zfcp_fsf_status_read_handler(req); |
@@ -472,23 +442,40 @@ void zfcp_fsf_req_complete(struct zfcp_fsf_req *req) | |||
472 | 442 | ||
473 | if (req->erp_action) | 443 | if (req->erp_action) |
474 | zfcp_erp_notify(req->erp_action, 0); | 444 | zfcp_erp_notify(req->erp_action, 0); |
475 | req->status |= ZFCP_STATUS_FSFREQ_COMPLETED; | ||
476 | 445 | ||
477 | if (likely(req->status & ZFCP_STATUS_FSFREQ_CLEANUP)) | 446 | if (likely(req->status & ZFCP_STATUS_FSFREQ_CLEANUP)) |
478 | zfcp_fsf_req_free(req); | 447 | zfcp_fsf_req_free(req); |
479 | else | 448 | else |
480 | /* notify initiator waiting for the requests completion */ | 449 | complete(&req->completion); |
481 | /* | 450 | } |
482 | * FIXME: Race! We must not access fsf_req here as it might have been | 451 | |
483 | * cleaned up already due to the set ZFCP_STATUS_FSFREQ_COMPLETED | 452 | /** |
484 | * flag. It's an improbable case. But, we have the same paranoia for | 453 | * zfcp_fsf_req_dismiss_all - dismiss all fsf requests |
485 | * the cleanup flag already. | 454 | * @adapter: pointer to struct zfcp_adapter |
486 | * Might better be handled using complete()? | 455 | * |
487 | * (setting the flag and doing wakeup ought to be atomic | 456 | * Never ever call this without shutting down the adapter first. |
488 | * with regard to checking the flag as long as waitqueue is | 457 | * Otherwise the adapter would continue using and corrupting s390 storage. |
489 | * part of the to be released structure) | 458 | * Included BUG_ON() call to ensure this is done. |
490 | */ | 459 | * ERP is supposed to be the only user of this function. |
491 | wake_up(&req->completion_wq); | 460 | */ |
461 | void zfcp_fsf_req_dismiss_all(struct zfcp_adapter *adapter) | ||
462 | { | ||
463 | struct zfcp_fsf_req *req, *tmp; | ||
464 | unsigned long flags; | ||
465 | LIST_HEAD(remove_queue); | ||
466 | unsigned int i; | ||
467 | |||
468 | BUG_ON(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP); | ||
469 | spin_lock_irqsave(&adapter->req_list_lock, flags); | ||
470 | for (i = 0; i < REQUEST_LIST_SIZE; i++) | ||
471 | list_splice_init(&adapter->req_list[i], &remove_queue); | ||
472 | spin_unlock_irqrestore(&adapter->req_list_lock, flags); | ||
473 | |||
474 | list_for_each_entry_safe(req, tmp, &remove_queue, list) { | ||
475 | list_del(&req->list); | ||
476 | req->status |= ZFCP_STATUS_FSFREQ_DISMISSED; | ||
477 | zfcp_fsf_req_complete(req); | ||
478 | } | ||
492 | } | 479 | } |
493 | 480 | ||
494 | static int zfcp_fsf_exchange_config_evaluate(struct zfcp_fsf_req *req) | 481 | static int zfcp_fsf_exchange_config_evaluate(struct zfcp_fsf_req *req) |
@@ -650,79 +637,77 @@ static void zfcp_fsf_exchange_port_data_handler(struct zfcp_fsf_req *req) | |||
650 | } | 637 | } |
651 | } | 638 | } |
652 | 639 | ||
653 | static int zfcp_fsf_sbal_check(struct zfcp_adapter *adapter) | 640 | static int zfcp_fsf_sbal_check(struct zfcp_qdio *qdio) |
654 | { | 641 | { |
655 | struct zfcp_qdio_queue *req_q = &adapter->req_q; | 642 | struct zfcp_qdio_queue *req_q = &qdio->req_q; |
656 | 643 | ||
657 | spin_lock_bh(&adapter->req_q_lock); | 644 | spin_lock_bh(&qdio->req_q_lock); |
658 | if (atomic_read(&req_q->count)) | 645 | if (atomic_read(&req_q->count)) |
659 | return 1; | 646 | return 1; |
660 | spin_unlock_bh(&adapter->req_q_lock); | 647 | spin_unlock_bh(&qdio->req_q_lock); |
661 | return 0; | 648 | return 0; |
662 | } | 649 | } |
663 | 650 | ||
664 | static int zfcp_fsf_req_sbal_get(struct zfcp_adapter *adapter) | 651 | static int zfcp_fsf_req_sbal_get(struct zfcp_qdio *qdio) |
665 | { | 652 | { |
653 | struct zfcp_adapter *adapter = qdio->adapter; | ||
666 | long ret; | 654 | long ret; |
667 | 655 | ||
668 | spin_unlock_bh(&adapter->req_q_lock); | 656 | spin_unlock_bh(&qdio->req_q_lock); |
669 | ret = wait_event_interruptible_timeout(adapter->request_wq, | 657 | ret = wait_event_interruptible_timeout(qdio->req_q_wq, |
670 | zfcp_fsf_sbal_check(adapter), 5 * HZ); | 658 | zfcp_fsf_sbal_check(qdio), 5 * HZ); |
671 | if (ret > 0) | 659 | if (ret > 0) |
672 | return 0; | 660 | return 0; |
673 | if (!ret) { | 661 | if (!ret) { |
674 | atomic_inc(&adapter->qdio_outb_full); | 662 | atomic_inc(&qdio->req_q_full); |
675 | /* assume hanging outbound queue, try queue recovery */ | 663 | /* assume hanging outbound queue, try queue recovery */ |
676 | zfcp_erp_adapter_reopen(adapter, 0, "fsrsg_1", NULL); | 664 | zfcp_erp_adapter_reopen(adapter, 0, "fsrsg_1", NULL); |
677 | } | 665 | } |
678 | 666 | ||
679 | spin_lock_bh(&adapter->req_q_lock); | 667 | spin_lock_bh(&qdio->req_q_lock); |
680 | return -EIO; | 668 | return -EIO; |
681 | } | 669 | } |
682 | 670 | ||
683 | static struct zfcp_fsf_req *zfcp_fsf_alloc_noqtcb(mempool_t *pool) | 671 | static struct zfcp_fsf_req *zfcp_fsf_alloc(mempool_t *pool) |
684 | { | 672 | { |
685 | struct zfcp_fsf_req *req; | 673 | struct zfcp_fsf_req *req; |
686 | req = mempool_alloc(pool, GFP_ATOMIC); | 674 | |
687 | if (!req) | 675 | if (likely(pool)) |
676 | req = mempool_alloc(pool, GFP_ATOMIC); | ||
677 | else | ||
678 | req = kmalloc(sizeof(*req), GFP_ATOMIC); | ||
679 | |||
680 | if (unlikely(!req)) | ||
688 | return NULL; | 681 | return NULL; |
682 | |||
689 | memset(req, 0, sizeof(*req)); | 683 | memset(req, 0, sizeof(*req)); |
690 | req->pool = pool; | 684 | req->pool = pool; |
691 | return req; | 685 | return req; |
692 | } | 686 | } |
693 | 687 | ||
694 | static struct zfcp_fsf_req *zfcp_fsf_alloc_qtcb(mempool_t *pool) | 688 | static struct fsf_qtcb *zfcp_qtcb_alloc(mempool_t *pool) |
695 | { | 689 | { |
696 | struct zfcp_fsf_req_qtcb *qtcb; | 690 | struct fsf_qtcb *qtcb; |
697 | 691 | ||
698 | if (likely(pool)) | 692 | if (likely(pool)) |
699 | qtcb = mempool_alloc(pool, GFP_ATOMIC); | 693 | qtcb = mempool_alloc(pool, GFP_ATOMIC); |
700 | else | 694 | else |
701 | qtcb = kmem_cache_alloc(zfcp_data.fsf_req_qtcb_cache, | 695 | qtcb = kmem_cache_alloc(zfcp_data.qtcb_cache, GFP_ATOMIC); |
702 | GFP_ATOMIC); | 696 | |
703 | if (unlikely(!qtcb)) | 697 | if (unlikely(!qtcb)) |
704 | return NULL; | 698 | return NULL; |
705 | 699 | ||
706 | memset(qtcb, 0, sizeof(*qtcb)); | 700 | memset(qtcb, 0, sizeof(*qtcb)); |
707 | qtcb->fsf_req.qtcb = &qtcb->qtcb; | 701 | return qtcb; |
708 | qtcb->fsf_req.pool = pool; | ||
709 | |||
710 | return &qtcb->fsf_req; | ||
711 | } | 702 | } |
712 | 703 | ||
713 | static struct zfcp_fsf_req *zfcp_fsf_req_create(struct zfcp_adapter *adapter, | 704 | static struct zfcp_fsf_req *zfcp_fsf_req_create(struct zfcp_qdio *qdio, |
714 | u32 fsf_cmd, int req_flags, | 705 | u32 fsf_cmd, mempool_t *pool) |
715 | mempool_t *pool) | ||
716 | { | 706 | { |
717 | struct qdio_buffer_element *sbale; | 707 | struct qdio_buffer_element *sbale; |
718 | 708 | struct zfcp_qdio_queue *req_q = &qdio->req_q; | |
719 | struct zfcp_fsf_req *req; | 709 | struct zfcp_adapter *adapter = qdio->adapter; |
720 | struct zfcp_qdio_queue *req_q = &adapter->req_q; | 710 | struct zfcp_fsf_req *req = zfcp_fsf_alloc(pool); |
721 | |||
722 | if (req_flags & ZFCP_REQ_NO_QTCB) | ||
723 | req = zfcp_fsf_alloc_noqtcb(pool); | ||
724 | else | ||
725 | req = zfcp_fsf_alloc_qtcb(pool); | ||
726 | 711 | ||
727 | if (unlikely(!req)) | 712 | if (unlikely(!req)) |
728 | return ERR_PTR(-ENOMEM); | 713 | return ERR_PTR(-ENOMEM); |
@@ -732,22 +717,32 @@ static struct zfcp_fsf_req *zfcp_fsf_req_create(struct zfcp_adapter *adapter, | |||
732 | 717 | ||
733 | INIT_LIST_HEAD(&req->list); | 718 | INIT_LIST_HEAD(&req->list); |
734 | init_timer(&req->timer); | 719 | init_timer(&req->timer); |
735 | init_waitqueue_head(&req->completion_wq); | 720 | init_completion(&req->completion); |
736 | 721 | ||
737 | req->adapter = adapter; | 722 | req->adapter = adapter; |
738 | req->fsf_command = fsf_cmd; | 723 | req->fsf_command = fsf_cmd; |
739 | req->req_id = adapter->req_no; | 724 | req->req_id = adapter->req_no; |
740 | req->sbal_number = 1; | 725 | req->queue_req.sbal_number = 1; |
741 | req->sbal_first = req_q->first; | 726 | req->queue_req.sbal_first = req_q->first; |
742 | req->sbal_last = req_q->first; | 727 | req->queue_req.sbal_last = req_q->first; |
743 | req->sbale_curr = 1; | 728 | req->queue_req.sbale_curr = 1; |
744 | 729 | ||
745 | sbale = zfcp_qdio_sbale_req(req); | 730 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); |
746 | sbale[0].addr = (void *) req->req_id; | 731 | sbale[0].addr = (void *) req->req_id; |
747 | sbale[0].flags |= SBAL_FLAGS0_COMMAND; | 732 | sbale[0].flags |= SBAL_FLAGS0_COMMAND; |
748 | 733 | ||
749 | if (likely(req->qtcb)) { | 734 | if (likely(fsf_cmd != FSF_QTCB_UNSOLICITED_STATUS)) { |
750 | req->qtcb->prefix.req_seq_no = req->adapter->fsf_req_seq_no; | 735 | if (likely(pool)) |
736 | req->qtcb = zfcp_qtcb_alloc(adapter->pool.qtcb_pool); | ||
737 | else | ||
738 | req->qtcb = zfcp_qtcb_alloc(NULL); | ||
739 | |||
740 | if (unlikely(!req->qtcb)) { | ||
741 | zfcp_fsf_req_free(req); | ||
742 | return ERR_PTR(-ENOMEM); | ||
743 | } | ||
744 | |||
745 | req->qtcb->prefix.req_seq_no = adapter->fsf_req_seq_no; | ||
751 | req->qtcb->prefix.req_id = req->req_id; | 746 | req->qtcb->prefix.req_id = req->req_id; |
752 | req->qtcb->prefix.ulp_info = 26; | 747 | req->qtcb->prefix.ulp_info = 26; |
753 | req->qtcb->prefix.qtcb_type = fsf_qtcb_type[req->fsf_command]; | 748 | req->qtcb->prefix.qtcb_type = fsf_qtcb_type[req->fsf_command]; |
@@ -765,15 +760,13 @@ static struct zfcp_fsf_req *zfcp_fsf_req_create(struct zfcp_adapter *adapter, | |||
765 | return ERR_PTR(-EIO); | 760 | return ERR_PTR(-EIO); |
766 | } | 761 | } |
767 | 762 | ||
768 | if (likely(req_flags & ZFCP_REQ_AUTO_CLEANUP)) | ||
769 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; | ||
770 | |||
771 | return req; | 763 | return req; |
772 | } | 764 | } |
773 | 765 | ||
774 | static int zfcp_fsf_req_send(struct zfcp_fsf_req *req) | 766 | static int zfcp_fsf_req_send(struct zfcp_fsf_req *req) |
775 | { | 767 | { |
776 | struct zfcp_adapter *adapter = req->adapter; | 768 | struct zfcp_adapter *adapter = req->adapter; |
769 | struct zfcp_qdio *qdio = adapter->qdio; | ||
777 | unsigned long flags; | 770 | unsigned long flags; |
778 | int idx; | 771 | int idx; |
779 | int with_qtcb = (req->qtcb != NULL); | 772 | int with_qtcb = (req->qtcb != NULL); |
@@ -784,9 +777,9 @@ static int zfcp_fsf_req_send(struct zfcp_fsf_req *req) | |||
784 | list_add_tail(&req->list, &adapter->req_list[idx]); | 777 | list_add_tail(&req->list, &adapter->req_list[idx]); |
785 | spin_unlock_irqrestore(&adapter->req_list_lock, flags); | 778 | spin_unlock_irqrestore(&adapter->req_list_lock, flags); |
786 | 779 | ||
787 | req->qdio_outb_usage = atomic_read(&adapter->req_q.count); | 780 | req->queue_req.qdio_outb_usage = atomic_read(&qdio->req_q.count); |
788 | req->issued = get_clock(); | 781 | req->issued = get_clock(); |
789 | if (zfcp_qdio_send(req)) { | 782 | if (zfcp_qdio_send(qdio, &req->queue_req)) { |
790 | del_timer(&req->timer); | 783 | del_timer(&req->timer); |
791 | spin_lock_irqsave(&adapter->req_list_lock, flags); | 784 | spin_lock_irqsave(&adapter->req_list_lock, flags); |
792 | /* lookup request again, list might have changed */ | 785 | /* lookup request again, list might have changed */ |
@@ -811,38 +804,37 @@ static int zfcp_fsf_req_send(struct zfcp_fsf_req *req) | |||
811 | * @req_flags: request flags | 804 | * @req_flags: request flags |
812 | * Returns: 0 on success, ERROR otherwise | 805 | * Returns: 0 on success, ERROR otherwise |
813 | */ | 806 | */ |
814 | int zfcp_fsf_status_read(struct zfcp_adapter *adapter) | 807 | int zfcp_fsf_status_read(struct zfcp_qdio *qdio) |
815 | { | 808 | { |
809 | struct zfcp_adapter *adapter = qdio->adapter; | ||
816 | struct zfcp_fsf_req *req; | 810 | struct zfcp_fsf_req *req; |
817 | struct fsf_status_read_buffer *sr_buf; | 811 | struct fsf_status_read_buffer *sr_buf; |
818 | struct qdio_buffer_element *sbale; | 812 | struct qdio_buffer_element *sbale; |
819 | int retval = -EIO; | 813 | int retval = -EIO; |
820 | 814 | ||
821 | spin_lock_bh(&adapter->req_q_lock); | 815 | spin_lock_bh(&qdio->req_q_lock); |
822 | if (zfcp_fsf_req_sbal_get(adapter)) | 816 | if (zfcp_fsf_req_sbal_get(qdio)) |
823 | goto out; | 817 | goto out; |
824 | 818 | ||
825 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_UNSOLICITED_STATUS, | 819 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_UNSOLICITED_STATUS, |
826 | ZFCP_REQ_NO_QTCB, | 820 | adapter->pool.status_read_req); |
827 | adapter->pool.fsf_req_status_read); | ||
828 | if (IS_ERR(req)) { | 821 | if (IS_ERR(req)) { |
829 | retval = PTR_ERR(req); | 822 | retval = PTR_ERR(req); |
830 | goto out; | 823 | goto out; |
831 | } | 824 | } |
832 | 825 | ||
833 | sbale = zfcp_qdio_sbale_req(req); | 826 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); |
834 | sbale[0].flags |= SBAL_FLAGS0_TYPE_STATUS; | ||
835 | sbale[2].flags |= SBAL_FLAGS_LAST_ENTRY; | 827 | sbale[2].flags |= SBAL_FLAGS_LAST_ENTRY; |
836 | req->sbale_curr = 2; | 828 | req->queue_req.sbale_curr = 2; |
837 | 829 | ||
838 | sr_buf = mempool_alloc(adapter->pool.data_status_read, GFP_ATOMIC); | 830 | sr_buf = mempool_alloc(adapter->pool.status_read_data, GFP_ATOMIC); |
839 | if (!sr_buf) { | 831 | if (!sr_buf) { |
840 | retval = -ENOMEM; | 832 | retval = -ENOMEM; |
841 | goto failed_buf; | 833 | goto failed_buf; |
842 | } | 834 | } |
843 | memset(sr_buf, 0, sizeof(*sr_buf)); | 835 | memset(sr_buf, 0, sizeof(*sr_buf)); |
844 | req->data = sr_buf; | 836 | req->data = sr_buf; |
845 | sbale = zfcp_qdio_sbale_curr(req); | 837 | sbale = zfcp_qdio_sbale_curr(qdio, &req->queue_req); |
846 | sbale->addr = (void *) sr_buf; | 838 | sbale->addr = (void *) sr_buf; |
847 | sbale->length = sizeof(*sr_buf); | 839 | sbale->length = sizeof(*sr_buf); |
848 | 840 | ||
@@ -853,12 +845,12 @@ int zfcp_fsf_status_read(struct zfcp_adapter *adapter) | |||
853 | goto out; | 845 | goto out; |
854 | 846 | ||
855 | failed_req_send: | 847 | failed_req_send: |
856 | mempool_free(sr_buf, adapter->pool.data_status_read); | 848 | mempool_free(sr_buf, adapter->pool.status_read_data); |
857 | failed_buf: | 849 | failed_buf: |
858 | zfcp_fsf_req_free(req); | 850 | zfcp_fsf_req_free(req); |
859 | zfcp_hba_dbf_event_fsf_unsol("fail", adapter, NULL); | 851 | zfcp_dbf_hba_fsf_unsol("fail", adapter->dbf, NULL); |
860 | out: | 852 | out: |
861 | spin_unlock_bh(&adapter->req_q_lock); | 853 | spin_unlock_bh(&qdio->req_q_lock); |
862 | return retval; | 854 | return retval; |
863 | } | 855 | } |
864 | 856 | ||
@@ -900,7 +892,7 @@ static void zfcp_fsf_abort_fcp_command_handler(struct zfcp_fsf_req *req) | |||
900 | case FSF_ADAPTER_STATUS_AVAILABLE: | 892 | case FSF_ADAPTER_STATUS_AVAILABLE: |
901 | switch (fsq->word[0]) { | 893 | switch (fsq->word[0]) { |
902 | case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: | 894 | case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: |
903 | zfcp_test_link(unit->port); | 895 | zfcp_fc_test_link(unit->port); |
904 | /* fall through */ | 896 | /* fall through */ |
905 | case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: | 897 | case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: |
906 | req->status |= ZFCP_STATUS_FSFREQ_ERROR; | 898 | req->status |= ZFCP_STATUS_FSFREQ_ERROR; |
@@ -925,13 +917,13 @@ struct zfcp_fsf_req *zfcp_fsf_abort_fcp_command(unsigned long old_req_id, | |||
925 | { | 917 | { |
926 | struct qdio_buffer_element *sbale; | 918 | struct qdio_buffer_element *sbale; |
927 | struct zfcp_fsf_req *req = NULL; | 919 | struct zfcp_fsf_req *req = NULL; |
928 | struct zfcp_adapter *adapter = unit->port->adapter; | 920 | struct zfcp_qdio *qdio = unit->port->adapter->qdio; |
929 | 921 | ||
930 | spin_lock_bh(&adapter->req_q_lock); | 922 | spin_lock_bh(&qdio->req_q_lock); |
931 | if (zfcp_fsf_req_sbal_get(adapter)) | 923 | if (zfcp_fsf_req_sbal_get(qdio)) |
932 | goto out; | 924 | goto out; |
933 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_ABORT_FCP_CMND, | 925 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_ABORT_FCP_CMND, |
934 | 0, adapter->pool.fsf_req_abort); | 926 | qdio->adapter->pool.scsi_abort); |
935 | if (IS_ERR(req)) { | 927 | if (IS_ERR(req)) { |
936 | req = NULL; | 928 | req = NULL; |
937 | goto out; | 929 | goto out; |
@@ -941,7 +933,7 @@ struct zfcp_fsf_req *zfcp_fsf_abort_fcp_command(unsigned long old_req_id, | |||
941 | ZFCP_STATUS_COMMON_UNBLOCKED))) | 933 | ZFCP_STATUS_COMMON_UNBLOCKED))) |
942 | goto out_error_free; | 934 | goto out_error_free; |
943 | 935 | ||
944 | sbale = zfcp_qdio_sbale_req(req); | 936 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); |
945 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; | 937 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; |
946 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 938 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
947 | 939 | ||
@@ -959,7 +951,7 @@ out_error_free: | |||
959 | zfcp_fsf_req_free(req); | 951 | zfcp_fsf_req_free(req); |
960 | req = NULL; | 952 | req = NULL; |
961 | out: | 953 | out: |
962 | spin_unlock_bh(&adapter->req_q_lock); | 954 | spin_unlock_bh(&qdio->req_q_lock); |
963 | return req; | 955 | return req; |
964 | } | 956 | } |
965 | 957 | ||
@@ -976,7 +968,7 @@ static void zfcp_fsf_send_ct_handler(struct zfcp_fsf_req *req) | |||
976 | 968 | ||
977 | switch (header->fsf_status) { | 969 | switch (header->fsf_status) { |
978 | case FSF_GOOD: | 970 | case FSF_GOOD: |
979 | zfcp_san_dbf_event_ct_response(req); | 971 | zfcp_dbf_san_ct_response(req); |
980 | send_ct->status = 0; | 972 | send_ct->status = 0; |
981 | break; | 973 | break; |
982 | case FSF_SERVICE_CLASS_NOT_SUPPORTED: | 974 | case FSF_SERVICE_CLASS_NOT_SUPPORTED: |
@@ -1035,8 +1027,10 @@ static int zfcp_fsf_setup_ct_els_sbals(struct zfcp_fsf_req *req, | |||
1035 | struct scatterlist *sg_resp, | 1027 | struct scatterlist *sg_resp, |
1036 | int max_sbals) | 1028 | int max_sbals) |
1037 | { | 1029 | { |
1038 | struct qdio_buffer_element *sbale = zfcp_qdio_sbale_req(req); | 1030 | struct zfcp_adapter *adapter = req->adapter; |
1039 | u32 feat = req->adapter->adapter_features; | 1031 | struct qdio_buffer_element *sbale = zfcp_qdio_sbale_req(adapter->qdio, |
1032 | &req->queue_req); | ||
1033 | u32 feat = adapter->adapter_features; | ||
1040 | int bytes; | 1034 | int bytes; |
1041 | 1035 | ||
1042 | if (!(feat & FSF_FEATURE_ELS_CT_CHAINED_SBALS)) { | 1036 | if (!(feat & FSF_FEATURE_ELS_CT_CHAINED_SBALS)) { |
@@ -1053,18 +1047,25 @@ static int zfcp_fsf_setup_ct_els_sbals(struct zfcp_fsf_req *req, | |||
1053 | return 0; | 1047 | return 0; |
1054 | } | 1048 | } |
1055 | 1049 | ||
1056 | bytes = zfcp_qdio_sbals_from_sg(req, SBAL_FLAGS0_TYPE_WRITE_READ, | 1050 | bytes = zfcp_qdio_sbals_from_sg(adapter->qdio, &req->queue_req, |
1051 | SBAL_FLAGS0_TYPE_WRITE_READ, | ||
1057 | sg_req, max_sbals); | 1052 | sg_req, max_sbals); |
1058 | if (bytes <= 0) | 1053 | if (bytes <= 0) |
1059 | return -EIO; | 1054 | return -EIO; |
1060 | req->qtcb->bottom.support.req_buf_length = bytes; | 1055 | req->qtcb->bottom.support.req_buf_length = bytes; |
1061 | req->sbale_curr = ZFCP_LAST_SBALE_PER_SBAL; | 1056 | req->queue_req.sbale_curr = ZFCP_LAST_SBALE_PER_SBAL; |
1062 | 1057 | ||
1063 | bytes = zfcp_qdio_sbals_from_sg(req, SBAL_FLAGS0_TYPE_WRITE_READ, | 1058 | bytes = zfcp_qdio_sbals_from_sg(adapter->qdio, &req->queue_req, |
1059 | SBAL_FLAGS0_TYPE_WRITE_READ, | ||
1064 | sg_resp, max_sbals); | 1060 | sg_resp, max_sbals); |
1065 | if (bytes <= 0) | 1061 | if (bytes <= 0) |
1066 | return -EIO; | 1062 | return -EIO; |
1063 | |||
1064 | /* common settings for ct/gs and els requests */ | ||
1067 | req->qtcb->bottom.support.resp_buf_length = bytes; | 1065 | req->qtcb->bottom.support.resp_buf_length = bytes; |
1066 | req->qtcb->bottom.support.service_class = FSF_CLASS_3; | ||
1067 | req->qtcb->bottom.support.timeout = 2 * R_A_TOV; | ||
1068 | zfcp_fsf_start_timer(req, 2 * R_A_TOV + 10); | ||
1068 | 1069 | ||
1069 | return 0; | 1070 | return 0; |
1070 | } | 1071 | } |
@@ -1073,27 +1074,26 @@ static int zfcp_fsf_setup_ct_els_sbals(struct zfcp_fsf_req *req, | |||
1073 | * zfcp_fsf_send_ct - initiate a Generic Service request (FC-GS) | 1074 | * zfcp_fsf_send_ct - initiate a Generic Service request (FC-GS) |
1074 | * @ct: pointer to struct zfcp_send_ct with data for request | 1075 | * @ct: pointer to struct zfcp_send_ct with data for request |
1075 | * @pool: if non-null this mempool is used to allocate struct zfcp_fsf_req | 1076 | * @pool: if non-null this mempool is used to allocate struct zfcp_fsf_req |
1076 | * @erp_action: if non-null the Generic Service request sent within ERP | ||
1077 | */ | 1077 | */ |
1078 | int zfcp_fsf_send_ct(struct zfcp_send_ct *ct, mempool_t *pool, | 1078 | int zfcp_fsf_send_ct(struct zfcp_send_ct *ct, mempool_t *pool) |
1079 | struct zfcp_erp_action *erp_action) | ||
1080 | { | 1079 | { |
1081 | struct zfcp_wka_port *wka_port = ct->wka_port; | 1080 | struct zfcp_wka_port *wka_port = ct->wka_port; |
1082 | struct zfcp_adapter *adapter = wka_port->adapter; | 1081 | struct zfcp_qdio *qdio = wka_port->adapter->qdio; |
1083 | struct zfcp_fsf_req *req; | 1082 | struct zfcp_fsf_req *req; |
1084 | int ret = -EIO; | 1083 | int ret = -EIO; |
1085 | 1084 | ||
1086 | spin_lock_bh(&adapter->req_q_lock); | 1085 | spin_lock_bh(&qdio->req_q_lock); |
1087 | if (zfcp_fsf_req_sbal_get(adapter)) | 1086 | if (zfcp_fsf_req_sbal_get(qdio)) |
1088 | goto out; | 1087 | goto out; |
1089 | 1088 | ||
1090 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_SEND_GENERIC, | 1089 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_SEND_GENERIC, pool); |
1091 | ZFCP_REQ_AUTO_CLEANUP, pool); | 1090 | |
1092 | if (IS_ERR(req)) { | 1091 | if (IS_ERR(req)) { |
1093 | ret = PTR_ERR(req); | 1092 | ret = PTR_ERR(req); |
1094 | goto out; | 1093 | goto out; |
1095 | } | 1094 | } |
1096 | 1095 | ||
1096 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; | ||
1097 | ret = zfcp_fsf_setup_ct_els_sbals(req, ct->req, ct->resp, | 1097 | ret = zfcp_fsf_setup_ct_els_sbals(req, ct->req, ct->resp, |
1098 | FSF_MAX_SBALS_PER_REQ); | 1098 | FSF_MAX_SBALS_PER_REQ); |
1099 | if (ret) | 1099 | if (ret) |
@@ -1101,18 +1101,9 @@ int zfcp_fsf_send_ct(struct zfcp_send_ct *ct, mempool_t *pool, | |||
1101 | 1101 | ||
1102 | req->handler = zfcp_fsf_send_ct_handler; | 1102 | req->handler = zfcp_fsf_send_ct_handler; |
1103 | req->qtcb->header.port_handle = wka_port->handle; | 1103 | req->qtcb->header.port_handle = wka_port->handle; |
1104 | req->qtcb->bottom.support.service_class = FSF_CLASS_3; | ||
1105 | req->qtcb->bottom.support.timeout = ct->timeout; | ||
1106 | req->data = ct; | 1104 | req->data = ct; |
1107 | 1105 | ||
1108 | zfcp_san_dbf_event_ct_request(req); | 1106 | zfcp_dbf_san_ct_request(req); |
1109 | |||
1110 | if (erp_action) { | ||
1111 | erp_action->fsf_req = req; | ||
1112 | req->erp_action = erp_action; | ||
1113 | zfcp_fsf_start_erp_timer(req); | ||
1114 | } else | ||
1115 | zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); | ||
1116 | 1107 | ||
1117 | ret = zfcp_fsf_req_send(req); | 1108 | ret = zfcp_fsf_req_send(req); |
1118 | if (ret) | 1109 | if (ret) |
@@ -1122,10 +1113,8 @@ int zfcp_fsf_send_ct(struct zfcp_send_ct *ct, mempool_t *pool, | |||
1122 | 1113 | ||
1123 | failed_send: | 1114 | failed_send: |
1124 | zfcp_fsf_req_free(req); | 1115 | zfcp_fsf_req_free(req); |
1125 | if (erp_action) | ||
1126 | erp_action->fsf_req = NULL; | ||
1127 | out: | 1116 | out: |
1128 | spin_unlock_bh(&adapter->req_q_lock); | 1117 | spin_unlock_bh(&qdio->req_q_lock); |
1129 | return ret; | 1118 | return ret; |
1130 | } | 1119 | } |
1131 | 1120 | ||
@@ -1142,7 +1131,7 @@ static void zfcp_fsf_send_els_handler(struct zfcp_fsf_req *req) | |||
1142 | 1131 | ||
1143 | switch (header->fsf_status) { | 1132 | switch (header->fsf_status) { |
1144 | case FSF_GOOD: | 1133 | case FSF_GOOD: |
1145 | zfcp_san_dbf_event_els_response(req); | 1134 | zfcp_dbf_san_els_response(req); |
1146 | send_els->status = 0; | 1135 | send_els->status = 0; |
1147 | break; | 1136 | break; |
1148 | case FSF_SERVICE_CLASS_NOT_SUPPORTED: | 1137 | case FSF_SERVICE_CLASS_NOT_SUPPORTED: |
@@ -1152,7 +1141,7 @@ static void zfcp_fsf_send_els_handler(struct zfcp_fsf_req *req) | |||
1152 | switch (header->fsf_status_qual.word[0]){ | 1141 | switch (header->fsf_status_qual.word[0]){ |
1153 | case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: | 1142 | case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: |
1154 | if (port && (send_els->ls_code != ZFCP_LS_ADISC)) | 1143 | if (port && (send_els->ls_code != ZFCP_LS_ADISC)) |
1155 | zfcp_test_link(port); | 1144 | zfcp_fc_test_link(port); |
1156 | /*fall through */ | 1145 | /*fall through */ |
1157 | case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: | 1146 | case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: |
1158 | case FSF_SQ_RETRY_IF_POSSIBLE: | 1147 | case FSF_SQ_RETRY_IF_POSSIBLE: |
@@ -1188,35 +1177,32 @@ skip_fsfstatus: | |||
1188 | int zfcp_fsf_send_els(struct zfcp_send_els *els) | 1177 | int zfcp_fsf_send_els(struct zfcp_send_els *els) |
1189 | { | 1178 | { |
1190 | struct zfcp_fsf_req *req; | 1179 | struct zfcp_fsf_req *req; |
1191 | struct zfcp_adapter *adapter = els->adapter; | 1180 | struct zfcp_qdio *qdio = els->adapter->qdio; |
1192 | struct fsf_qtcb_bottom_support *bottom; | ||
1193 | int ret = -EIO; | 1181 | int ret = -EIO; |
1194 | 1182 | ||
1195 | spin_lock_bh(&adapter->req_q_lock); | 1183 | spin_lock_bh(&qdio->req_q_lock); |
1196 | if (zfcp_fsf_req_sbal_get(adapter)) | 1184 | if (zfcp_fsf_req_sbal_get(qdio)) |
1197 | goto out; | 1185 | goto out; |
1198 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_SEND_ELS, | 1186 | |
1199 | ZFCP_REQ_AUTO_CLEANUP, NULL); | 1187 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_SEND_ELS, NULL); |
1188 | |||
1200 | if (IS_ERR(req)) { | 1189 | if (IS_ERR(req)) { |
1201 | ret = PTR_ERR(req); | 1190 | ret = PTR_ERR(req); |
1202 | goto out; | 1191 | goto out; |
1203 | } | 1192 | } |
1204 | 1193 | ||
1194 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; | ||
1205 | ret = zfcp_fsf_setup_ct_els_sbals(req, els->req, els->resp, 2); | 1195 | ret = zfcp_fsf_setup_ct_els_sbals(req, els->req, els->resp, 2); |
1206 | 1196 | ||
1207 | if (ret) | 1197 | if (ret) |
1208 | goto failed_send; | 1198 | goto failed_send; |
1209 | 1199 | ||
1210 | bottom = &req->qtcb->bottom.support; | 1200 | req->qtcb->bottom.support.d_id = els->d_id; |
1211 | req->handler = zfcp_fsf_send_els_handler; | 1201 | req->handler = zfcp_fsf_send_els_handler; |
1212 | bottom->d_id = els->d_id; | ||
1213 | bottom->service_class = FSF_CLASS_3; | ||
1214 | bottom->timeout = 2 * R_A_TOV; | ||
1215 | req->data = els; | 1202 | req->data = els; |
1216 | 1203 | ||
1217 | zfcp_san_dbf_event_els_request(req); | 1204 | zfcp_dbf_san_els_request(req); |
1218 | 1205 | ||
1219 | zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); | ||
1220 | ret = zfcp_fsf_req_send(req); | 1206 | ret = zfcp_fsf_req_send(req); |
1221 | if (ret) | 1207 | if (ret) |
1222 | goto failed_send; | 1208 | goto failed_send; |
@@ -1226,7 +1212,7 @@ int zfcp_fsf_send_els(struct zfcp_send_els *els) | |||
1226 | failed_send: | 1212 | failed_send: |
1227 | zfcp_fsf_req_free(req); | 1213 | zfcp_fsf_req_free(req); |
1228 | out: | 1214 | out: |
1229 | spin_unlock_bh(&adapter->req_q_lock); | 1215 | spin_unlock_bh(&qdio->req_q_lock); |
1230 | return ret; | 1216 | return ret; |
1231 | } | 1217 | } |
1232 | 1218 | ||
@@ -1234,22 +1220,23 @@ int zfcp_fsf_exchange_config_data(struct zfcp_erp_action *erp_action) | |||
1234 | { | 1220 | { |
1235 | struct qdio_buffer_element *sbale; | 1221 | struct qdio_buffer_element *sbale; |
1236 | struct zfcp_fsf_req *req; | 1222 | struct zfcp_fsf_req *req; |
1237 | struct zfcp_adapter *adapter = erp_action->adapter; | 1223 | struct zfcp_qdio *qdio = erp_action->adapter->qdio; |
1238 | int retval = -EIO; | 1224 | int retval = -EIO; |
1239 | 1225 | ||
1240 | spin_lock_bh(&adapter->req_q_lock); | 1226 | spin_lock_bh(&qdio->req_q_lock); |
1241 | if (zfcp_fsf_req_sbal_get(adapter)) | 1227 | if (zfcp_fsf_req_sbal_get(qdio)) |
1242 | goto out; | 1228 | goto out; |
1243 | req = zfcp_fsf_req_create(adapter, | 1229 | |
1244 | FSF_QTCB_EXCHANGE_CONFIG_DATA, | 1230 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_CONFIG_DATA, |
1245 | ZFCP_REQ_AUTO_CLEANUP, | 1231 | qdio->adapter->pool.erp_req); |
1246 | adapter->pool.fsf_req_erp); | 1232 | |
1247 | if (IS_ERR(req)) { | 1233 | if (IS_ERR(req)) { |
1248 | retval = PTR_ERR(req); | 1234 | retval = PTR_ERR(req); |
1249 | goto out; | 1235 | goto out; |
1250 | } | 1236 | } |
1251 | 1237 | ||
1252 | sbale = zfcp_qdio_sbale_req(req); | 1238 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; |
1239 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); | ||
1253 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; | 1240 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; |
1254 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 1241 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
1255 | 1242 | ||
@@ -1269,29 +1256,29 @@ int zfcp_fsf_exchange_config_data(struct zfcp_erp_action *erp_action) | |||
1269 | erp_action->fsf_req = NULL; | 1256 | erp_action->fsf_req = NULL; |
1270 | } | 1257 | } |
1271 | out: | 1258 | out: |
1272 | spin_unlock_bh(&adapter->req_q_lock); | 1259 | spin_unlock_bh(&qdio->req_q_lock); |
1273 | return retval; | 1260 | return retval; |
1274 | } | 1261 | } |
1275 | 1262 | ||
1276 | int zfcp_fsf_exchange_config_data_sync(struct zfcp_adapter *adapter, | 1263 | int zfcp_fsf_exchange_config_data_sync(struct zfcp_qdio *qdio, |
1277 | struct fsf_qtcb_bottom_config *data) | 1264 | struct fsf_qtcb_bottom_config *data) |
1278 | { | 1265 | { |
1279 | struct qdio_buffer_element *sbale; | 1266 | struct qdio_buffer_element *sbale; |
1280 | struct zfcp_fsf_req *req = NULL; | 1267 | struct zfcp_fsf_req *req = NULL; |
1281 | int retval = -EIO; | 1268 | int retval = -EIO; |
1282 | 1269 | ||
1283 | spin_lock_bh(&adapter->req_q_lock); | 1270 | spin_lock_bh(&qdio->req_q_lock); |
1284 | if (zfcp_fsf_req_sbal_get(adapter)) | 1271 | if (zfcp_fsf_req_sbal_get(qdio)) |
1285 | goto out_unlock; | 1272 | goto out_unlock; |
1286 | 1273 | ||
1287 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_EXCHANGE_CONFIG_DATA, | 1274 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_CONFIG_DATA, NULL); |
1288 | 0, NULL); | 1275 | |
1289 | if (IS_ERR(req)) { | 1276 | if (IS_ERR(req)) { |
1290 | retval = PTR_ERR(req); | 1277 | retval = PTR_ERR(req); |
1291 | goto out_unlock; | 1278 | goto out_unlock; |
1292 | } | 1279 | } |
1293 | 1280 | ||
1294 | sbale = zfcp_qdio_sbale_req(req); | 1281 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); |
1295 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; | 1282 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; |
1296 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 1283 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
1297 | req->handler = zfcp_fsf_exchange_config_data_handler; | 1284 | req->handler = zfcp_fsf_exchange_config_data_handler; |
@@ -1307,16 +1294,15 @@ int zfcp_fsf_exchange_config_data_sync(struct zfcp_adapter *adapter, | |||
1307 | 1294 | ||
1308 | zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); | 1295 | zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); |
1309 | retval = zfcp_fsf_req_send(req); | 1296 | retval = zfcp_fsf_req_send(req); |
1310 | spin_unlock_bh(&adapter->req_q_lock); | 1297 | spin_unlock_bh(&qdio->req_q_lock); |
1311 | if (!retval) | 1298 | if (!retval) |
1312 | wait_event(req->completion_wq, | 1299 | wait_for_completion(&req->completion); |
1313 | req->status & ZFCP_STATUS_FSFREQ_COMPLETED); | ||
1314 | 1300 | ||
1315 | zfcp_fsf_req_free(req); | 1301 | zfcp_fsf_req_free(req); |
1316 | return retval; | 1302 | return retval; |
1317 | 1303 | ||
1318 | out_unlock: | 1304 | out_unlock: |
1319 | spin_unlock_bh(&adapter->req_q_lock); | 1305 | spin_unlock_bh(&qdio->req_q_lock); |
1320 | return retval; | 1306 | return retval; |
1321 | } | 1307 | } |
1322 | 1308 | ||
@@ -1327,26 +1313,28 @@ out_unlock: | |||
1327 | */ | 1313 | */ |
1328 | int zfcp_fsf_exchange_port_data(struct zfcp_erp_action *erp_action) | 1314 | int zfcp_fsf_exchange_port_data(struct zfcp_erp_action *erp_action) |
1329 | { | 1315 | { |
1316 | struct zfcp_qdio *qdio = erp_action->adapter->qdio; | ||
1330 | struct qdio_buffer_element *sbale; | 1317 | struct qdio_buffer_element *sbale; |
1331 | struct zfcp_fsf_req *req; | 1318 | struct zfcp_fsf_req *req; |
1332 | struct zfcp_adapter *adapter = erp_action->adapter; | ||
1333 | int retval = -EIO; | 1319 | int retval = -EIO; |
1334 | 1320 | ||
1335 | if (!(adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT)) | 1321 | if (!(qdio->adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT)) |
1336 | return -EOPNOTSUPP; | 1322 | return -EOPNOTSUPP; |
1337 | 1323 | ||
1338 | spin_lock_bh(&adapter->req_q_lock); | 1324 | spin_lock_bh(&qdio->req_q_lock); |
1339 | if (zfcp_fsf_req_sbal_get(adapter)) | 1325 | if (zfcp_fsf_req_sbal_get(qdio)) |
1340 | goto out; | 1326 | goto out; |
1341 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_EXCHANGE_PORT_DATA, | 1327 | |
1342 | ZFCP_REQ_AUTO_CLEANUP, | 1328 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_PORT_DATA, |
1343 | adapter->pool.fsf_req_erp); | 1329 | qdio->adapter->pool.erp_req); |
1330 | |||
1344 | if (IS_ERR(req)) { | 1331 | if (IS_ERR(req)) { |
1345 | retval = PTR_ERR(req); | 1332 | retval = PTR_ERR(req); |
1346 | goto out; | 1333 | goto out; |
1347 | } | 1334 | } |
1348 | 1335 | ||
1349 | sbale = zfcp_qdio_sbale_req(req); | 1336 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; |
1337 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); | ||
1350 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; | 1338 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; |
1351 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 1339 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
1352 | 1340 | ||
@@ -1361,32 +1349,32 @@ int zfcp_fsf_exchange_port_data(struct zfcp_erp_action *erp_action) | |||
1361 | erp_action->fsf_req = NULL; | 1349 | erp_action->fsf_req = NULL; |
1362 | } | 1350 | } |
1363 | out: | 1351 | out: |
1364 | spin_unlock_bh(&adapter->req_q_lock); | 1352 | spin_unlock_bh(&qdio->req_q_lock); |
1365 | return retval; | 1353 | return retval; |
1366 | } | 1354 | } |
1367 | 1355 | ||
1368 | /** | 1356 | /** |
1369 | * zfcp_fsf_exchange_port_data_sync - request information about local port | 1357 | * zfcp_fsf_exchange_port_data_sync - request information about local port |
1370 | * @adapter: pointer to struct zfcp_adapter | 1358 | * @qdio: pointer to struct zfcp_qdio |
1371 | * @data: pointer to struct fsf_qtcb_bottom_port | 1359 | * @data: pointer to struct fsf_qtcb_bottom_port |
1372 | * Returns: 0 on success, error otherwise | 1360 | * Returns: 0 on success, error otherwise |
1373 | */ | 1361 | */ |
1374 | int zfcp_fsf_exchange_port_data_sync(struct zfcp_adapter *adapter, | 1362 | int zfcp_fsf_exchange_port_data_sync(struct zfcp_qdio *qdio, |
1375 | struct fsf_qtcb_bottom_port *data) | 1363 | struct fsf_qtcb_bottom_port *data) |
1376 | { | 1364 | { |
1377 | struct qdio_buffer_element *sbale; | 1365 | struct qdio_buffer_element *sbale; |
1378 | struct zfcp_fsf_req *req = NULL; | 1366 | struct zfcp_fsf_req *req = NULL; |
1379 | int retval = -EIO; | 1367 | int retval = -EIO; |
1380 | 1368 | ||
1381 | if (!(adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT)) | 1369 | if (!(qdio->adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT)) |
1382 | return -EOPNOTSUPP; | 1370 | return -EOPNOTSUPP; |
1383 | 1371 | ||
1384 | spin_lock_bh(&adapter->req_q_lock); | 1372 | spin_lock_bh(&qdio->req_q_lock); |
1385 | if (zfcp_fsf_req_sbal_get(adapter)) | 1373 | if (zfcp_fsf_req_sbal_get(qdio)) |
1386 | goto out_unlock; | 1374 | goto out_unlock; |
1387 | 1375 | ||
1388 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_EXCHANGE_PORT_DATA, 0, | 1376 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_PORT_DATA, NULL); |
1389 | NULL); | 1377 | |
1390 | if (IS_ERR(req)) { | 1378 | if (IS_ERR(req)) { |
1391 | retval = PTR_ERR(req); | 1379 | retval = PTR_ERR(req); |
1392 | goto out_unlock; | 1380 | goto out_unlock; |
@@ -1395,24 +1383,24 @@ int zfcp_fsf_exchange_port_data_sync(struct zfcp_adapter *adapter, | |||
1395 | if (data) | 1383 | if (data) |
1396 | req->data = data; | 1384 | req->data = data; |
1397 | 1385 | ||
1398 | sbale = zfcp_qdio_sbale_req(req); | 1386 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); |
1399 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; | 1387 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; |
1400 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 1388 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
1401 | 1389 | ||
1402 | req->handler = zfcp_fsf_exchange_port_data_handler; | 1390 | req->handler = zfcp_fsf_exchange_port_data_handler; |
1403 | zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); | 1391 | zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); |
1404 | retval = zfcp_fsf_req_send(req); | 1392 | retval = zfcp_fsf_req_send(req); |
1405 | spin_unlock_bh(&adapter->req_q_lock); | 1393 | spin_unlock_bh(&qdio->req_q_lock); |
1406 | 1394 | ||
1407 | if (!retval) | 1395 | if (!retval) |
1408 | wait_event(req->completion_wq, | 1396 | wait_for_completion(&req->completion); |
1409 | req->status & ZFCP_STATUS_FSFREQ_COMPLETED); | 1397 | |
1410 | zfcp_fsf_req_free(req); | 1398 | zfcp_fsf_req_free(req); |
1411 | 1399 | ||
1412 | return retval; | 1400 | return retval; |
1413 | 1401 | ||
1414 | out_unlock: | 1402 | out_unlock: |
1415 | spin_unlock_bh(&adapter->req_q_lock); | 1403 | spin_unlock_bh(&qdio->req_q_lock); |
1416 | return retval; | 1404 | return retval; |
1417 | } | 1405 | } |
1418 | 1406 | ||
@@ -1498,25 +1486,25 @@ out: | |||
1498 | int zfcp_fsf_open_port(struct zfcp_erp_action *erp_action) | 1486 | int zfcp_fsf_open_port(struct zfcp_erp_action *erp_action) |
1499 | { | 1487 | { |
1500 | struct qdio_buffer_element *sbale; | 1488 | struct qdio_buffer_element *sbale; |
1501 | struct zfcp_adapter *adapter = erp_action->adapter; | 1489 | struct zfcp_qdio *qdio = erp_action->adapter->qdio; |
1502 | struct zfcp_fsf_req *req; | ||
1503 | struct zfcp_port *port = erp_action->port; | 1490 | struct zfcp_port *port = erp_action->port; |
1491 | struct zfcp_fsf_req *req; | ||
1504 | int retval = -EIO; | 1492 | int retval = -EIO; |
1505 | 1493 | ||
1506 | spin_lock_bh(&adapter->req_q_lock); | 1494 | spin_lock_bh(&qdio->req_q_lock); |
1507 | if (zfcp_fsf_req_sbal_get(adapter)) | 1495 | if (zfcp_fsf_req_sbal_get(qdio)) |
1508 | goto out; | 1496 | goto out; |
1509 | 1497 | ||
1510 | req = zfcp_fsf_req_create(adapter, | 1498 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_OPEN_PORT_WITH_DID, |
1511 | FSF_QTCB_OPEN_PORT_WITH_DID, | 1499 | qdio->adapter->pool.erp_req); |
1512 | ZFCP_REQ_AUTO_CLEANUP, | 1500 | |
1513 | adapter->pool.fsf_req_erp); | ||
1514 | if (IS_ERR(req)) { | 1501 | if (IS_ERR(req)) { |
1515 | retval = PTR_ERR(req); | 1502 | retval = PTR_ERR(req); |
1516 | goto out; | 1503 | goto out; |
1517 | } | 1504 | } |
1518 | 1505 | ||
1519 | sbale = zfcp_qdio_sbale_req(req); | 1506 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; |
1507 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); | ||
1520 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; | 1508 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; |
1521 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 1509 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
1522 | 1510 | ||
@@ -1535,7 +1523,7 @@ int zfcp_fsf_open_port(struct zfcp_erp_action *erp_action) | |||
1535 | zfcp_port_put(port); | 1523 | zfcp_port_put(port); |
1536 | } | 1524 | } |
1537 | out: | 1525 | out: |
1538 | spin_unlock_bh(&adapter->req_q_lock); | 1526 | spin_unlock_bh(&qdio->req_q_lock); |
1539 | return retval; | 1527 | return retval; |
1540 | } | 1528 | } |
1541 | 1529 | ||
@@ -1569,23 +1557,24 @@ static void zfcp_fsf_close_port_handler(struct zfcp_fsf_req *req) | |||
1569 | int zfcp_fsf_close_port(struct zfcp_erp_action *erp_action) | 1557 | int zfcp_fsf_close_port(struct zfcp_erp_action *erp_action) |
1570 | { | 1558 | { |
1571 | struct qdio_buffer_element *sbale; | 1559 | struct qdio_buffer_element *sbale; |
1572 | struct zfcp_adapter *adapter = erp_action->adapter; | 1560 | struct zfcp_qdio *qdio = erp_action->adapter->qdio; |
1573 | struct zfcp_fsf_req *req; | 1561 | struct zfcp_fsf_req *req; |
1574 | int retval = -EIO; | 1562 | int retval = -EIO; |
1575 | 1563 | ||
1576 | spin_lock_bh(&adapter->req_q_lock); | 1564 | spin_lock_bh(&qdio->req_q_lock); |
1577 | if (zfcp_fsf_req_sbal_get(adapter)) | 1565 | if (zfcp_fsf_req_sbal_get(qdio)) |
1578 | goto out; | 1566 | goto out; |
1579 | 1567 | ||
1580 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_CLOSE_PORT, | 1568 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_PORT, |
1581 | ZFCP_REQ_AUTO_CLEANUP, | 1569 | qdio->adapter->pool.erp_req); |
1582 | adapter->pool.fsf_req_erp); | 1570 | |
1583 | if (IS_ERR(req)) { | 1571 | if (IS_ERR(req)) { |
1584 | retval = PTR_ERR(req); | 1572 | retval = PTR_ERR(req); |
1585 | goto out; | 1573 | goto out; |
1586 | } | 1574 | } |
1587 | 1575 | ||
1588 | sbale = zfcp_qdio_sbale_req(req); | 1576 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; |
1577 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); | ||
1589 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; | 1578 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; |
1590 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 1579 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
1591 | 1580 | ||
@@ -1602,7 +1591,7 @@ int zfcp_fsf_close_port(struct zfcp_erp_action *erp_action) | |||
1602 | erp_action->fsf_req = NULL; | 1591 | erp_action->fsf_req = NULL; |
1603 | } | 1592 | } |
1604 | out: | 1593 | out: |
1605 | spin_unlock_bh(&adapter->req_q_lock); | 1594 | spin_unlock_bh(&qdio->req_q_lock); |
1606 | return retval; | 1595 | return retval; |
1607 | } | 1596 | } |
1608 | 1597 | ||
@@ -1645,24 +1634,24 @@ out: | |||
1645 | int zfcp_fsf_open_wka_port(struct zfcp_wka_port *wka_port) | 1634 | int zfcp_fsf_open_wka_port(struct zfcp_wka_port *wka_port) |
1646 | { | 1635 | { |
1647 | struct qdio_buffer_element *sbale; | 1636 | struct qdio_buffer_element *sbale; |
1648 | struct zfcp_adapter *adapter = wka_port->adapter; | 1637 | struct zfcp_qdio *qdio = wka_port->adapter->qdio; |
1649 | struct zfcp_fsf_req *req; | 1638 | struct zfcp_fsf_req *req; |
1650 | int retval = -EIO; | 1639 | int retval = -EIO; |
1651 | 1640 | ||
1652 | spin_lock_bh(&adapter->req_q_lock); | 1641 | spin_lock_bh(&qdio->req_q_lock); |
1653 | if (zfcp_fsf_req_sbal_get(adapter)) | 1642 | if (zfcp_fsf_req_sbal_get(qdio)) |
1654 | goto out; | 1643 | goto out; |
1655 | 1644 | ||
1656 | req = zfcp_fsf_req_create(adapter, | 1645 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_OPEN_PORT_WITH_DID, |
1657 | FSF_QTCB_OPEN_PORT_WITH_DID, | 1646 | qdio->adapter->pool.erp_req); |
1658 | ZFCP_REQ_AUTO_CLEANUP, | 1647 | |
1659 | adapter->pool.fsf_req_erp); | ||
1660 | if (unlikely(IS_ERR(req))) { | 1648 | if (unlikely(IS_ERR(req))) { |
1661 | retval = PTR_ERR(req); | 1649 | retval = PTR_ERR(req); |
1662 | goto out; | 1650 | goto out; |
1663 | } | 1651 | } |
1664 | 1652 | ||
1665 | sbale = zfcp_qdio_sbale_req(req); | 1653 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; |
1654 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); | ||
1666 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; | 1655 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; |
1667 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 1656 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
1668 | 1657 | ||
@@ -1675,7 +1664,7 @@ int zfcp_fsf_open_wka_port(struct zfcp_wka_port *wka_port) | |||
1675 | if (retval) | 1664 | if (retval) |
1676 | zfcp_fsf_req_free(req); | 1665 | zfcp_fsf_req_free(req); |
1677 | out: | 1666 | out: |
1678 | spin_unlock_bh(&adapter->req_q_lock); | 1667 | spin_unlock_bh(&qdio->req_q_lock); |
1679 | return retval; | 1668 | return retval; |
1680 | } | 1669 | } |
1681 | 1670 | ||
@@ -1700,23 +1689,24 @@ static void zfcp_fsf_close_wka_port_handler(struct zfcp_fsf_req *req) | |||
1700 | int zfcp_fsf_close_wka_port(struct zfcp_wka_port *wka_port) | 1689 | int zfcp_fsf_close_wka_port(struct zfcp_wka_port *wka_port) |
1701 | { | 1690 | { |
1702 | struct qdio_buffer_element *sbale; | 1691 | struct qdio_buffer_element *sbale; |
1703 | struct zfcp_adapter *adapter = wka_port->adapter; | 1692 | struct zfcp_qdio *qdio = wka_port->adapter->qdio; |
1704 | struct zfcp_fsf_req *req; | 1693 | struct zfcp_fsf_req *req; |
1705 | int retval = -EIO; | 1694 | int retval = -EIO; |
1706 | 1695 | ||
1707 | spin_lock_bh(&adapter->req_q_lock); | 1696 | spin_lock_bh(&qdio->req_q_lock); |
1708 | if (zfcp_fsf_req_sbal_get(adapter)) | 1697 | if (zfcp_fsf_req_sbal_get(qdio)) |
1709 | goto out; | 1698 | goto out; |
1710 | 1699 | ||
1711 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_CLOSE_PORT, | 1700 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_PORT, |
1712 | ZFCP_REQ_AUTO_CLEANUP, | 1701 | qdio->adapter->pool.erp_req); |
1713 | adapter->pool.fsf_req_erp); | 1702 | |
1714 | if (unlikely(IS_ERR(req))) { | 1703 | if (unlikely(IS_ERR(req))) { |
1715 | retval = PTR_ERR(req); | 1704 | retval = PTR_ERR(req); |
1716 | goto out; | 1705 | goto out; |
1717 | } | 1706 | } |
1718 | 1707 | ||
1719 | sbale = zfcp_qdio_sbale_req(req); | 1708 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; |
1709 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); | ||
1720 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; | 1710 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; |
1721 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 1711 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
1722 | 1712 | ||
@@ -1729,7 +1719,7 @@ int zfcp_fsf_close_wka_port(struct zfcp_wka_port *wka_port) | |||
1729 | if (retval) | 1719 | if (retval) |
1730 | zfcp_fsf_req_free(req); | 1720 | zfcp_fsf_req_free(req); |
1731 | out: | 1721 | out: |
1732 | spin_unlock_bh(&adapter->req_q_lock); | 1722 | spin_unlock_bh(&qdio->req_q_lock); |
1733 | return retval; | 1723 | return retval; |
1734 | } | 1724 | } |
1735 | 1725 | ||
@@ -1791,23 +1781,24 @@ static void zfcp_fsf_close_physical_port_handler(struct zfcp_fsf_req *req) | |||
1791 | int zfcp_fsf_close_physical_port(struct zfcp_erp_action *erp_action) | 1781 | int zfcp_fsf_close_physical_port(struct zfcp_erp_action *erp_action) |
1792 | { | 1782 | { |
1793 | struct qdio_buffer_element *sbale; | 1783 | struct qdio_buffer_element *sbale; |
1794 | struct zfcp_adapter *adapter = erp_action->adapter; | 1784 | struct zfcp_qdio *qdio = erp_action->adapter->qdio; |
1795 | struct zfcp_fsf_req *req; | 1785 | struct zfcp_fsf_req *req; |
1796 | int retval = -EIO; | 1786 | int retval = -EIO; |
1797 | 1787 | ||
1798 | spin_lock_bh(&adapter->req_q_lock); | 1788 | spin_lock_bh(&qdio->req_q_lock); |
1799 | if (zfcp_fsf_req_sbal_get(adapter)) | 1789 | if (zfcp_fsf_req_sbal_get(qdio)) |
1800 | goto out; | 1790 | goto out; |
1801 | 1791 | ||
1802 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_CLOSE_PHYSICAL_PORT, | 1792 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_PHYSICAL_PORT, |
1803 | ZFCP_REQ_AUTO_CLEANUP, | 1793 | qdio->adapter->pool.erp_req); |
1804 | adapter->pool.fsf_req_erp); | 1794 | |
1805 | if (IS_ERR(req)) { | 1795 | if (IS_ERR(req)) { |
1806 | retval = PTR_ERR(req); | 1796 | retval = PTR_ERR(req); |
1807 | goto out; | 1797 | goto out; |
1808 | } | 1798 | } |
1809 | 1799 | ||
1810 | sbale = zfcp_qdio_sbale_req(req); | 1800 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; |
1801 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); | ||
1811 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; | 1802 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; |
1812 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 1803 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
1813 | 1804 | ||
@@ -1824,7 +1815,7 @@ int zfcp_fsf_close_physical_port(struct zfcp_erp_action *erp_action) | |||
1824 | erp_action->fsf_req = NULL; | 1815 | erp_action->fsf_req = NULL; |
1825 | } | 1816 | } |
1826 | out: | 1817 | out: |
1827 | spin_unlock_bh(&adapter->req_q_lock); | 1818 | spin_unlock_bh(&qdio->req_q_lock); |
1828 | return retval; | 1819 | return retval; |
1829 | } | 1820 | } |
1830 | 1821 | ||
@@ -1895,7 +1886,7 @@ static void zfcp_fsf_open_unit_handler(struct zfcp_fsf_req *req) | |||
1895 | case FSF_ADAPTER_STATUS_AVAILABLE: | 1886 | case FSF_ADAPTER_STATUS_AVAILABLE: |
1896 | switch (header->fsf_status_qual.word[0]) { | 1887 | switch (header->fsf_status_qual.word[0]) { |
1897 | case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: | 1888 | case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: |
1898 | zfcp_test_link(unit->port); | 1889 | zfcp_fc_test_link(unit->port); |
1899 | /* fall through */ | 1890 | /* fall through */ |
1900 | case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: | 1891 | case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: |
1901 | req->status |= ZFCP_STATUS_FSFREQ_ERROR; | 1892 | req->status |= ZFCP_STATUS_FSFREQ_ERROR; |
@@ -1964,22 +1955,24 @@ int zfcp_fsf_open_unit(struct zfcp_erp_action *erp_action) | |||
1964 | { | 1955 | { |
1965 | struct qdio_buffer_element *sbale; | 1956 | struct qdio_buffer_element *sbale; |
1966 | struct zfcp_adapter *adapter = erp_action->adapter; | 1957 | struct zfcp_adapter *adapter = erp_action->adapter; |
1958 | struct zfcp_qdio *qdio = adapter->qdio; | ||
1967 | struct zfcp_fsf_req *req; | 1959 | struct zfcp_fsf_req *req; |
1968 | int retval = -EIO; | 1960 | int retval = -EIO; |
1969 | 1961 | ||
1970 | spin_lock_bh(&adapter->req_q_lock); | 1962 | spin_lock_bh(&qdio->req_q_lock); |
1971 | if (zfcp_fsf_req_sbal_get(adapter)) | 1963 | if (zfcp_fsf_req_sbal_get(qdio)) |
1972 | goto out; | 1964 | goto out; |
1973 | 1965 | ||
1974 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_OPEN_LUN, | 1966 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_OPEN_LUN, |
1975 | ZFCP_REQ_AUTO_CLEANUP, | 1967 | adapter->pool.erp_req); |
1976 | adapter->pool.fsf_req_erp); | 1968 | |
1977 | if (IS_ERR(req)) { | 1969 | if (IS_ERR(req)) { |
1978 | retval = PTR_ERR(req); | 1970 | retval = PTR_ERR(req); |
1979 | goto out; | 1971 | goto out; |
1980 | } | 1972 | } |
1981 | 1973 | ||
1982 | sbale = zfcp_qdio_sbale_req(req); | 1974 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; |
1975 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); | ||
1983 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; | 1976 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; |
1984 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 1977 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
1985 | 1978 | ||
@@ -2000,7 +1993,7 @@ int zfcp_fsf_open_unit(struct zfcp_erp_action *erp_action) | |||
2000 | erp_action->fsf_req = NULL; | 1993 | erp_action->fsf_req = NULL; |
2001 | } | 1994 | } |
2002 | out: | 1995 | out: |
2003 | spin_unlock_bh(&adapter->req_q_lock); | 1996 | spin_unlock_bh(&qdio->req_q_lock); |
2004 | return retval; | 1997 | return retval; |
2005 | } | 1998 | } |
2006 | 1999 | ||
@@ -2028,7 +2021,7 @@ static void zfcp_fsf_close_unit_handler(struct zfcp_fsf_req *req) | |||
2028 | case FSF_ADAPTER_STATUS_AVAILABLE: | 2021 | case FSF_ADAPTER_STATUS_AVAILABLE: |
2029 | switch (req->qtcb->header.fsf_status_qual.word[0]) { | 2022 | switch (req->qtcb->header.fsf_status_qual.word[0]) { |
2030 | case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: | 2023 | case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: |
2031 | zfcp_test_link(unit->port); | 2024 | zfcp_fc_test_link(unit->port); |
2032 | /* fall through */ | 2025 | /* fall through */ |
2033 | case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: | 2026 | case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: |
2034 | req->status |= ZFCP_STATUS_FSFREQ_ERROR; | 2027 | req->status |= ZFCP_STATUS_FSFREQ_ERROR; |
@@ -2049,22 +2042,24 @@ static void zfcp_fsf_close_unit_handler(struct zfcp_fsf_req *req) | |||
2049 | int zfcp_fsf_close_unit(struct zfcp_erp_action *erp_action) | 2042 | int zfcp_fsf_close_unit(struct zfcp_erp_action *erp_action) |
2050 | { | 2043 | { |
2051 | struct qdio_buffer_element *sbale; | 2044 | struct qdio_buffer_element *sbale; |
2052 | struct zfcp_adapter *adapter = erp_action->adapter; | 2045 | struct zfcp_qdio *qdio = erp_action->adapter->qdio; |
2053 | struct zfcp_fsf_req *req; | 2046 | struct zfcp_fsf_req *req; |
2054 | int retval = -EIO; | 2047 | int retval = -EIO; |
2055 | 2048 | ||
2056 | spin_lock_bh(&adapter->req_q_lock); | 2049 | spin_lock_bh(&qdio->req_q_lock); |
2057 | if (zfcp_fsf_req_sbal_get(adapter)) | 2050 | if (zfcp_fsf_req_sbal_get(qdio)) |
2058 | goto out; | 2051 | goto out; |
2059 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_CLOSE_LUN, | 2052 | |
2060 | ZFCP_REQ_AUTO_CLEANUP, | 2053 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_LUN, |
2061 | adapter->pool.fsf_req_erp); | 2054 | qdio->adapter->pool.erp_req); |
2055 | |||
2062 | if (IS_ERR(req)) { | 2056 | if (IS_ERR(req)) { |
2063 | retval = PTR_ERR(req); | 2057 | retval = PTR_ERR(req); |
2064 | goto out; | 2058 | goto out; |
2065 | } | 2059 | } |
2066 | 2060 | ||
2067 | sbale = zfcp_qdio_sbale_req(req); | 2061 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; |
2062 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); | ||
2068 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; | 2063 | sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; |
2069 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 2064 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
2070 | 2065 | ||
@@ -2082,7 +2077,7 @@ int zfcp_fsf_close_unit(struct zfcp_erp_action *erp_action) | |||
2082 | erp_action->fsf_req = NULL; | 2077 | erp_action->fsf_req = NULL; |
2083 | } | 2078 | } |
2084 | out: | 2079 | out: |
2085 | spin_unlock_bh(&adapter->req_q_lock); | 2080 | spin_unlock_bh(&qdio->req_q_lock); |
2086 | return retval; | 2081 | return retval; |
2087 | } | 2082 | } |
2088 | 2083 | ||
@@ -2141,8 +2136,8 @@ static void zfcp_fsf_trace_latency(struct zfcp_fsf_req *fsf_req) | |||
2141 | } | 2136 | } |
2142 | if (fsf_req->status & ZFCP_STATUS_FSFREQ_ERROR) | 2137 | if (fsf_req->status & ZFCP_STATUS_FSFREQ_ERROR) |
2143 | trace.flags |= ZFCP_BLK_REQ_ERROR; | 2138 | trace.flags |= ZFCP_BLK_REQ_ERROR; |
2144 | trace.inb_usage = fsf_req->qdio_inb_usage; | 2139 | trace.inb_usage = fsf_req->queue_req.qdio_inb_usage; |
2145 | trace.outb_usage = fsf_req->qdio_outb_usage; | 2140 | trace.outb_usage = fsf_req->queue_req.qdio_outb_usage; |
2146 | 2141 | ||
2147 | blk_add_driver_data(req->q, req, &trace, sizeof(trace)); | 2142 | blk_add_driver_data(req->q, req, &trace, sizeof(trace)); |
2148 | } | 2143 | } |
@@ -2215,11 +2210,11 @@ static void zfcp_fsf_send_fcp_command_task_handler(struct zfcp_fsf_req *req) | |||
2215 | } | 2210 | } |
2216 | skip_fsfstatus: | 2211 | skip_fsfstatus: |
2217 | if (scpnt->result != 0) | 2212 | if (scpnt->result != 0) |
2218 | zfcp_scsi_dbf_event_result("erro", 3, req->adapter, scpnt, req); | 2213 | zfcp_dbf_scsi_result("erro", 3, req->adapter->dbf, scpnt, req); |
2219 | else if (scpnt->retries > 0) | 2214 | else if (scpnt->retries > 0) |
2220 | zfcp_scsi_dbf_event_result("retr", 4, req->adapter, scpnt, req); | 2215 | zfcp_dbf_scsi_result("retr", 4, req->adapter->dbf, scpnt, req); |
2221 | else | 2216 | else |
2222 | zfcp_scsi_dbf_event_result("norm", 6, req->adapter, scpnt, req); | 2217 | zfcp_dbf_scsi_result("norm", 6, req->adapter->dbf, scpnt, req); |
2223 | 2218 | ||
2224 | scpnt->host_scribble = NULL; | 2219 | scpnt->host_scribble = NULL; |
2225 | (scpnt->scsi_done) (scpnt); | 2220 | (scpnt->scsi_done) (scpnt); |
@@ -2309,7 +2304,7 @@ static void zfcp_fsf_send_fcp_command_handler(struct zfcp_fsf_req *req) | |||
2309 | case FSF_ADAPTER_STATUS_AVAILABLE: | 2304 | case FSF_ADAPTER_STATUS_AVAILABLE: |
2310 | if (header->fsf_status_qual.word[0] == | 2305 | if (header->fsf_status_qual.word[0] == |
2311 | FSF_SQ_INVOKE_LINK_TEST_PROCEDURE) | 2306 | FSF_SQ_INVOKE_LINK_TEST_PROCEDURE) |
2312 | zfcp_test_link(unit->port); | 2307 | zfcp_fc_test_link(unit->port); |
2313 | req->status |= ZFCP_STATUS_FSFREQ_ERROR; | 2308 | req->status |= ZFCP_STATUS_FSFREQ_ERROR; |
2314 | break; | 2309 | break; |
2315 | } | 2310 | } |
@@ -2350,24 +2345,27 @@ int zfcp_fsf_send_fcp_command_task(struct zfcp_unit *unit, | |||
2350 | unsigned int sbtype = SBAL_FLAGS0_TYPE_READ; | 2345 | unsigned int sbtype = SBAL_FLAGS0_TYPE_READ; |
2351 | int real_bytes, retval = -EIO; | 2346 | int real_bytes, retval = -EIO; |
2352 | struct zfcp_adapter *adapter = unit->port->adapter; | 2347 | struct zfcp_adapter *adapter = unit->port->adapter; |
2348 | struct zfcp_qdio *qdio = adapter->qdio; | ||
2353 | 2349 | ||
2354 | if (unlikely(!(atomic_read(&unit->status) & | 2350 | if (unlikely(!(atomic_read(&unit->status) & |
2355 | ZFCP_STATUS_COMMON_UNBLOCKED))) | 2351 | ZFCP_STATUS_COMMON_UNBLOCKED))) |
2356 | return -EBUSY; | 2352 | return -EBUSY; |
2357 | 2353 | ||
2358 | spin_lock(&adapter->req_q_lock); | 2354 | spin_lock(&qdio->req_q_lock); |
2359 | if (atomic_read(&adapter->req_q.count) <= 0) { | 2355 | if (atomic_read(&qdio->req_q.count) <= 0) { |
2360 | atomic_inc(&adapter->qdio_outb_full); | 2356 | atomic_inc(&qdio->req_q_full); |
2361 | goto out; | 2357 | goto out; |
2362 | } | 2358 | } |
2363 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_FCP_CMND, | 2359 | |
2364 | ZFCP_REQ_AUTO_CLEANUP, | 2360 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_FCP_CMND, |
2365 | adapter->pool.fsf_req_scsi); | 2361 | adapter->pool.scsi_req); |
2362 | |||
2366 | if (IS_ERR(req)) { | 2363 | if (IS_ERR(req)) { |
2367 | retval = PTR_ERR(req); | 2364 | retval = PTR_ERR(req); |
2368 | goto out; | 2365 | goto out; |
2369 | } | 2366 | } |
2370 | 2367 | ||
2368 | req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; | ||
2371 | zfcp_unit_get(unit); | 2369 | zfcp_unit_get(unit); |
2372 | req->unit = unit; | 2370 | req->unit = unit; |
2373 | req->data = scsi_cmnd; | 2371 | req->data = scsi_cmnd; |
@@ -2419,11 +2417,11 @@ int zfcp_fsf_send_fcp_command_task(struct zfcp_unit *unit, | |||
2419 | req->qtcb->bottom.io.fcp_cmnd_length = sizeof(struct fcp_cmnd_iu) + | 2417 | req->qtcb->bottom.io.fcp_cmnd_length = sizeof(struct fcp_cmnd_iu) + |
2420 | fcp_cmnd_iu->add_fcp_cdb_length + sizeof(u32); | 2418 | fcp_cmnd_iu->add_fcp_cdb_length + sizeof(u32); |
2421 | 2419 | ||
2422 | real_bytes = zfcp_qdio_sbals_from_sg(req, sbtype, | 2420 | real_bytes = zfcp_qdio_sbals_from_sg(qdio, &req->queue_req, sbtype, |
2423 | scsi_sglist(scsi_cmnd), | 2421 | scsi_sglist(scsi_cmnd), |
2424 | FSF_MAX_SBALS_PER_REQ); | 2422 | FSF_MAX_SBALS_PER_REQ); |
2425 | if (unlikely(real_bytes < 0)) { | 2423 | if (unlikely(real_bytes < 0)) { |
2426 | if (req->sbal_number >= FSF_MAX_SBALS_PER_REQ) { | 2424 | if (req->queue_req.sbal_number >= FSF_MAX_SBALS_PER_REQ) { |
2427 | dev_err(&adapter->ccw_device->dev, | 2425 | dev_err(&adapter->ccw_device->dev, |
2428 | "Oversize data package, unit 0x%016Lx " | 2426 | "Oversize data package, unit 0x%016Lx " |
2429 | "on port 0x%016Lx closed\n", | 2427 | "on port 0x%016Lx closed\n", |
@@ -2448,7 +2446,7 @@ failed_scsi_cmnd: | |||
2448 | zfcp_fsf_req_free(req); | 2446 | zfcp_fsf_req_free(req); |
2449 | scsi_cmnd->host_scribble = NULL; | 2447 | scsi_cmnd->host_scribble = NULL; |
2450 | out: | 2448 | out: |
2451 | spin_unlock(&adapter->req_q_lock); | 2449 | spin_unlock(&qdio->req_q_lock); |
2452 | return retval; | 2450 | return retval; |
2453 | } | 2451 | } |
2454 | 2452 | ||
@@ -2463,17 +2461,19 @@ struct zfcp_fsf_req *zfcp_fsf_send_fcp_ctm(struct zfcp_unit *unit, u8 tm_flags) | |||
2463 | struct qdio_buffer_element *sbale; | 2461 | struct qdio_buffer_element *sbale; |
2464 | struct zfcp_fsf_req *req = NULL; | 2462 | struct zfcp_fsf_req *req = NULL; |
2465 | struct fcp_cmnd_iu *fcp_cmnd_iu; | 2463 | struct fcp_cmnd_iu *fcp_cmnd_iu; |
2466 | struct zfcp_adapter *adapter = unit->port->adapter; | 2464 | struct zfcp_qdio *qdio = unit->port->adapter->qdio; |
2467 | 2465 | ||
2468 | if (unlikely(!(atomic_read(&unit->status) & | 2466 | if (unlikely(!(atomic_read(&unit->status) & |
2469 | ZFCP_STATUS_COMMON_UNBLOCKED))) | 2467 | ZFCP_STATUS_COMMON_UNBLOCKED))) |
2470 | return NULL; | 2468 | return NULL; |
2471 | 2469 | ||
2472 | spin_lock_bh(&adapter->req_q_lock); | 2470 | spin_lock_bh(&qdio->req_q_lock); |
2473 | if (zfcp_fsf_req_sbal_get(adapter)) | 2471 | if (zfcp_fsf_req_sbal_get(qdio)) |
2474 | goto out; | 2472 | goto out; |
2475 | req = zfcp_fsf_req_create(adapter, FSF_QTCB_FCP_CMND, 0, | 2473 | |
2476 | adapter->pool.fsf_req_scsi); | 2474 | req = zfcp_fsf_req_create(qdio, FSF_QTCB_FCP_CMND, |
2475 | qdio->adapter->pool.scsi_req); | ||
2476 | |||
2477 | if (IS_ERR(req)) { | 2477 | if (IS_ERR(req)) { |
2478 | req = NULL; | 2478 | req = NULL; |
2479 | goto out; | 2479 | goto out; |
@@ -2489,7 +2489,7 @@ struct zfcp_fsf_req *zfcp_fsf_send_fcp_ctm(struct zfcp_unit *unit, u8 tm_flags) | |||
2489 | req->qtcb->bottom.io.fcp_cmnd_length = sizeof(struct fcp_cmnd_iu) + | 2489 | req->qtcb->bottom.io.fcp_cmnd_length = sizeof(struct fcp_cmnd_iu) + |
2490 | sizeof(u32); | 2490 | sizeof(u32); |
2491 | 2491 | ||
2492 | sbale = zfcp_qdio_sbale_req(req); | 2492 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); |
2493 | sbale[0].flags |= SBAL_FLAGS0_TYPE_WRITE; | 2493 | sbale[0].flags |= SBAL_FLAGS0_TYPE_WRITE; |
2494 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; | 2494 | sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; |
2495 | 2495 | ||
@@ -2504,7 +2504,7 @@ struct zfcp_fsf_req *zfcp_fsf_send_fcp_ctm(struct zfcp_unit *unit, u8 tm_flags) | |||
2504 | zfcp_fsf_req_free(req); | 2504 | zfcp_fsf_req_free(req); |
2505 | req = NULL; | 2505 | req = NULL; |
2506 | out: | 2506 | out: |
2507 | spin_unlock_bh(&adapter->req_q_lock); | 2507 | spin_unlock_bh(&qdio->req_q_lock); |
2508 | return req; | 2508 | return req; |
2509 | } | 2509 | } |
2510 | 2510 | ||
@@ -2522,6 +2522,7 @@ struct zfcp_fsf_req *zfcp_fsf_control_file(struct zfcp_adapter *adapter, | |||
2522 | struct zfcp_fsf_cfdc *fsf_cfdc) | 2522 | struct zfcp_fsf_cfdc *fsf_cfdc) |
2523 | { | 2523 | { |
2524 | struct qdio_buffer_element *sbale; | 2524 | struct qdio_buffer_element *sbale; |
2525 | struct zfcp_qdio *qdio = adapter->qdio; | ||
2525 | struct zfcp_fsf_req *req = NULL; | 2526 | struct zfcp_fsf_req *req = NULL; |
2526 | struct fsf_qtcb_bottom_support *bottom; | 2527 | struct fsf_qtcb_bottom_support *bottom; |
2527 | int direction, retval = -EIO, bytes; | 2528 | int direction, retval = -EIO, bytes; |
@@ -2540,11 +2541,11 @@ struct zfcp_fsf_req *zfcp_fsf_control_file(struct zfcp_adapter *adapter, | |||
2540 | return ERR_PTR(-EINVAL); | 2541 | return ERR_PTR(-EINVAL); |
2541 | } | 2542 | } |
2542 | 2543 | ||
2543 | spin_lock_bh(&adapter->req_q_lock); | 2544 | spin_lock_bh(&qdio->req_q_lock); |
2544 | if (zfcp_fsf_req_sbal_get(adapter)) | 2545 | if (zfcp_fsf_req_sbal_get(qdio)) |
2545 | goto out; | 2546 | goto out; |
2546 | 2547 | ||
2547 | req = zfcp_fsf_req_create(adapter, fsf_cfdc->command, 0, NULL); | 2548 | req = zfcp_fsf_req_create(qdio, fsf_cfdc->command, NULL); |
2548 | if (IS_ERR(req)) { | 2549 | if (IS_ERR(req)) { |
2549 | retval = -EPERM; | 2550 | retval = -EPERM; |
2550 | goto out; | 2551 | goto out; |
@@ -2552,14 +2553,15 @@ struct zfcp_fsf_req *zfcp_fsf_control_file(struct zfcp_adapter *adapter, | |||
2552 | 2553 | ||
2553 | req->handler = zfcp_fsf_control_file_handler; | 2554 | req->handler = zfcp_fsf_control_file_handler; |
2554 | 2555 | ||
2555 | sbale = zfcp_qdio_sbale_req(req); | 2556 | sbale = zfcp_qdio_sbale_req(qdio, &req->queue_req); |
2556 | sbale[0].flags |= direction; | 2557 | sbale[0].flags |= direction; |
2557 | 2558 | ||
2558 | bottom = &req->qtcb->bottom.support; | 2559 | bottom = &req->qtcb->bottom.support; |
2559 | bottom->operation_subtype = FSF_CFDC_OPERATION_SUBTYPE; | 2560 | bottom->operation_subtype = FSF_CFDC_OPERATION_SUBTYPE; |
2560 | bottom->option = fsf_cfdc->option; | 2561 | bottom->option = fsf_cfdc->option; |
2561 | 2562 | ||
2562 | bytes = zfcp_qdio_sbals_from_sg(req, direction, fsf_cfdc->sg, | 2563 | bytes = zfcp_qdio_sbals_from_sg(qdio, &req->queue_req, |
2564 | direction, fsf_cfdc->sg, | ||
2563 | FSF_MAX_SBALS_PER_REQ); | 2565 | FSF_MAX_SBALS_PER_REQ); |
2564 | if (bytes != ZFCP_CFDC_MAX_SIZE) { | 2566 | if (bytes != ZFCP_CFDC_MAX_SIZE) { |
2565 | zfcp_fsf_req_free(req); | 2567 | zfcp_fsf_req_free(req); |
@@ -2569,12 +2571,53 @@ struct zfcp_fsf_req *zfcp_fsf_control_file(struct zfcp_adapter *adapter, | |||
2569 | zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); | 2571 | zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); |
2570 | retval = zfcp_fsf_req_send(req); | 2572 | retval = zfcp_fsf_req_send(req); |
2571 | out: | 2573 | out: |
2572 | spin_unlock_bh(&adapter->req_q_lock); | 2574 | spin_unlock_bh(&qdio->req_q_lock); |
2573 | 2575 | ||
2574 | if (!retval) { | 2576 | if (!retval) { |
2575 | wait_event(req->completion_wq, | 2577 | wait_for_completion(&req->completion); |
2576 | req->status & ZFCP_STATUS_FSFREQ_COMPLETED); | ||
2577 | return req; | 2578 | return req; |
2578 | } | 2579 | } |
2579 | return ERR_PTR(retval); | 2580 | return ERR_PTR(retval); |
2580 | } | 2581 | } |
2582 | |||
2583 | /** | ||
2584 | * zfcp_fsf_reqid_check - validate req_id contained in SBAL returned by QDIO | ||
2585 | * @adapter: pointer to struct zfcp_adapter | ||
2586 | * @sbal_idx: response queue index of SBAL to be processed | ||
2587 | */ | ||
2588 | void zfcp_fsf_reqid_check(struct zfcp_qdio *qdio, int sbal_idx) | ||
2589 | { | ||
2590 | struct zfcp_adapter *adapter = qdio->adapter; | ||
2591 | struct qdio_buffer *sbal = qdio->resp_q.sbal[sbal_idx]; | ||
2592 | struct qdio_buffer_element *sbale; | ||
2593 | struct zfcp_fsf_req *fsf_req; | ||
2594 | unsigned long flags, req_id; | ||
2595 | int idx; | ||
2596 | |||
2597 | for (idx = 0; idx < QDIO_MAX_ELEMENTS_PER_BUFFER; idx++) { | ||
2598 | |||
2599 | sbale = &sbal->element[idx]; | ||
2600 | req_id = (unsigned long) sbale->addr; | ||
2601 | spin_lock_irqsave(&adapter->req_list_lock, flags); | ||
2602 | fsf_req = zfcp_reqlist_find(adapter, req_id); | ||
2603 | |||
2604 | if (!fsf_req) | ||
2605 | /* | ||
2606 | * Unknown request means that we have potentially memory | ||
2607 | * corruption and must stop the machine immediately. | ||
2608 | */ | ||
2609 | panic("error: unknown req_id (%lx) on adapter %s.\n", | ||
2610 | req_id, dev_name(&adapter->ccw_device->dev)); | ||
2611 | |||
2612 | list_del(&fsf_req->list); | ||
2613 | spin_unlock_irqrestore(&adapter->req_list_lock, flags); | ||
2614 | |||
2615 | fsf_req->queue_req.sbal_response = sbal_idx; | ||
2616 | fsf_req->queue_req.qdio_inb_usage = | ||
2617 | atomic_read(&qdio->resp_q.count); | ||
2618 | zfcp_fsf_req_complete(fsf_req); | ||
2619 | |||
2620 | if (likely(sbale->flags & SBAL_FLAGS_LAST_ENTRY)) | ||
2621 | break; | ||
2622 | } | ||
2623 | } | ||
diff --git a/drivers/s390/scsi/zfcp_fsf.h b/drivers/s390/scsi/zfcp_fsf.h index df7f232faba8..dcc7c1dbcf58 100644 --- a/drivers/s390/scsi/zfcp_fsf.h +++ b/drivers/s390/scsi/zfcp_fsf.h | |||
@@ -3,13 +3,14 @@ | |||
3 | * | 3 | * |
4 | * Interface to the FSF support functions. | 4 | * Interface to the FSF support functions. |
5 | * | 5 | * |
6 | * Copyright IBM Corporation 2002, 2008 | 6 | * Copyright IBM Corporation 2002, 2009 |
7 | */ | 7 | */ |
8 | 8 | ||
9 | #ifndef FSF_H | 9 | #ifndef FSF_H |
10 | #define FSF_H | 10 | #define FSF_H |
11 | 11 | ||
12 | #include <linux/pfn.h> | 12 | #include <linux/pfn.h> |
13 | #include <linux/scatterlist.h> | ||
13 | 14 | ||
14 | #define FSF_QTCB_CURRENT_VERSION 0x00000001 | 15 | #define FSF_QTCB_CURRENT_VERSION 0x00000001 |
15 | 16 | ||
diff --git a/drivers/s390/scsi/zfcp_qdio.c b/drivers/s390/scsi/zfcp_qdio.c index e0a215309df0..6c5228b627fc 100644 --- a/drivers/s390/scsi/zfcp_qdio.c +++ b/drivers/s390/scsi/zfcp_qdio.c | |||
@@ -3,7 +3,7 @@ | |||
3 | * | 3 | * |
4 | * Setup and helper functions to access QDIO. | 4 | * Setup and helper functions to access QDIO. |
5 | * | 5 | * |
6 | * Copyright IBM Corporation 2002, 2008 | 6 | * Copyright IBM Corporation 2002, 2009 |
7 | */ | 7 | */ |
8 | 8 | ||
9 | #define KMSG_COMPONENT "zfcp" | 9 | #define KMSG_COMPONENT "zfcp" |
@@ -34,29 +34,10 @@ zfcp_qdio_sbale(struct zfcp_qdio_queue *q, int sbal_idx, int sbale_idx) | |||
34 | return &q->sbal[sbal_idx]->element[sbale_idx]; | 34 | return &q->sbal[sbal_idx]->element[sbale_idx]; |
35 | } | 35 | } |
36 | 36 | ||
37 | /** | 37 | static void zfcp_qdio_handler_error(struct zfcp_qdio *qdio, char *id) |
38 | * zfcp_qdio_free - free memory used by request- and resposne queue | ||
39 | * @adapter: pointer to the zfcp_adapter structure | ||
40 | */ | ||
41 | void zfcp_qdio_free(struct zfcp_adapter *adapter) | ||
42 | { | 38 | { |
43 | struct qdio_buffer **sbal_req, **sbal_resp; | 39 | struct zfcp_adapter *adapter = qdio->adapter; |
44 | int p; | ||
45 | |||
46 | if (adapter->ccw_device) | ||
47 | qdio_free(adapter->ccw_device); | ||
48 | |||
49 | sbal_req = adapter->req_q.sbal; | ||
50 | sbal_resp = adapter->resp_q.sbal; | ||
51 | |||
52 | for (p = 0; p < QDIO_MAX_BUFFERS_PER_Q; p += QBUFF_PER_PAGE) { | ||
53 | free_page((unsigned long) sbal_req[p]); | ||
54 | free_page((unsigned long) sbal_resp[p]); | ||
55 | } | ||
56 | } | ||
57 | 40 | ||
58 | static void zfcp_qdio_handler_error(struct zfcp_adapter *adapter, char *id) | ||
59 | { | ||
60 | dev_warn(&adapter->ccw_device->dev, "A QDIO problem occurred\n"); | 41 | dev_warn(&adapter->ccw_device->dev, "A QDIO problem occurred\n"); |
61 | 42 | ||
62 | zfcp_erp_adapter_reopen(adapter, | 43 | zfcp_erp_adapter_reopen(adapter, |
@@ -75,72 +56,47 @@ static void zfcp_qdio_zero_sbals(struct qdio_buffer *sbal[], int first, int cnt) | |||
75 | } | 56 | } |
76 | 57 | ||
77 | /* this needs to be called prior to updating the queue fill level */ | 58 | /* this needs to be called prior to updating the queue fill level */ |
78 | static void zfcp_qdio_account(struct zfcp_adapter *adapter) | 59 | static inline void zfcp_qdio_account(struct zfcp_qdio *qdio) |
79 | { | 60 | { |
80 | ktime_t now; | 61 | unsigned long long now, span; |
81 | s64 span; | ||
82 | int free, used; | 62 | int free, used; |
83 | 63 | ||
84 | spin_lock(&adapter->qdio_stat_lock); | 64 | spin_lock(&qdio->stat_lock); |
85 | now = ktime_get(); | 65 | now = get_clock_monotonic(); |
86 | span = ktime_us_delta(now, adapter->req_q_time); | 66 | span = (now - qdio->req_q_time) >> 12; |
87 | free = max(0, atomic_read(&adapter->req_q.count)); | 67 | free = atomic_read(&qdio->req_q.count); |
88 | used = QDIO_MAX_BUFFERS_PER_Q - free; | 68 | used = QDIO_MAX_BUFFERS_PER_Q - free; |
89 | adapter->req_q_util += used * span; | 69 | qdio->req_q_util += used * span; |
90 | adapter->req_q_time = now; | 70 | qdio->req_q_time = now; |
91 | spin_unlock(&adapter->qdio_stat_lock); | 71 | spin_unlock(&qdio->stat_lock); |
92 | } | 72 | } |
93 | 73 | ||
94 | static void zfcp_qdio_int_req(struct ccw_device *cdev, unsigned int qdio_err, | 74 | static void zfcp_qdio_int_req(struct ccw_device *cdev, unsigned int qdio_err, |
95 | int queue_no, int first, int count, | 75 | int queue_no, int first, int count, |
96 | unsigned long parm) | 76 | unsigned long parm) |
97 | { | 77 | { |
98 | struct zfcp_adapter *adapter = (struct zfcp_adapter *) parm; | 78 | struct zfcp_qdio *qdio = (struct zfcp_qdio *) parm; |
99 | struct zfcp_qdio_queue *queue = &adapter->req_q; | 79 | struct zfcp_qdio_queue *queue = &qdio->req_q; |
100 | 80 | ||
101 | if (unlikely(qdio_err)) { | 81 | if (unlikely(qdio_err)) { |
102 | zfcp_hba_dbf_event_qdio(adapter, qdio_err, first, count); | 82 | zfcp_dbf_hba_qdio(qdio->adapter->dbf, qdio_err, first, |
103 | zfcp_qdio_handler_error(adapter, "qdireq1"); | 83 | count); |
84 | zfcp_qdio_handler_error(qdio, "qdireq1"); | ||
104 | return; | 85 | return; |
105 | } | 86 | } |
106 | 87 | ||
107 | /* cleanup all SBALs being program-owned now */ | 88 | /* cleanup all SBALs being program-owned now */ |
108 | zfcp_qdio_zero_sbals(queue->sbal, first, count); | 89 | zfcp_qdio_zero_sbals(queue->sbal, first, count); |
109 | 90 | ||
110 | zfcp_qdio_account(adapter); | 91 | zfcp_qdio_account(qdio); |
111 | atomic_add(count, &queue->count); | 92 | atomic_add(count, &queue->count); |
112 | wake_up(&adapter->request_wq); | 93 | wake_up(&qdio->req_q_wq); |
113 | } | ||
114 | |||
115 | static void zfcp_qdio_reqid_check(struct zfcp_adapter *adapter, | ||
116 | unsigned long req_id, int sbal_idx) | ||
117 | { | ||
118 | struct zfcp_fsf_req *fsf_req; | ||
119 | unsigned long flags; | ||
120 | |||
121 | spin_lock_irqsave(&adapter->req_list_lock, flags); | ||
122 | fsf_req = zfcp_reqlist_find(adapter, req_id); | ||
123 | |||
124 | if (!fsf_req) | ||
125 | /* | ||
126 | * Unknown request means that we have potentially memory | ||
127 | * corruption and must stop the machine immediatly. | ||
128 | */ | ||
129 | panic("error: unknown request id (%lx) on adapter %s.\n", | ||
130 | req_id, dev_name(&adapter->ccw_device->dev)); | ||
131 | |||
132 | zfcp_reqlist_remove(adapter, fsf_req); | ||
133 | spin_unlock_irqrestore(&adapter->req_list_lock, flags); | ||
134 | |||
135 | fsf_req->sbal_response = sbal_idx; | ||
136 | fsf_req->qdio_inb_usage = atomic_read(&adapter->resp_q.count); | ||
137 | zfcp_fsf_req_complete(fsf_req); | ||
138 | } | 94 | } |
139 | 95 | ||
140 | static void zfcp_qdio_resp_put_back(struct zfcp_adapter *adapter, int processed) | 96 | static void zfcp_qdio_resp_put_back(struct zfcp_qdio *qdio, int processed) |
141 | { | 97 | { |
142 | struct zfcp_qdio_queue *queue = &adapter->resp_q; | 98 | struct zfcp_qdio_queue *queue = &qdio->resp_q; |
143 | struct ccw_device *cdev = adapter->ccw_device; | 99 | struct ccw_device *cdev = qdio->adapter->ccw_device; |
144 | u8 count, start = queue->first; | 100 | u8 count, start = queue->first; |
145 | unsigned int retval; | 101 | unsigned int retval; |
146 | 102 | ||
@@ -162,14 +118,13 @@ static void zfcp_qdio_int_resp(struct ccw_device *cdev, unsigned int qdio_err, | |||
162 | int queue_no, int first, int count, | 118 | int queue_no, int first, int count, |
163 | unsigned long parm) | 119 | unsigned long parm) |
164 | { | 120 | { |
165 | struct zfcp_adapter *adapter = (struct zfcp_adapter *) parm; | 121 | struct zfcp_qdio *qdio = (struct zfcp_qdio *) parm; |
166 | struct zfcp_qdio_queue *queue = &adapter->resp_q; | 122 | int sbal_idx, sbal_no; |
167 | struct qdio_buffer_element *sbale; | ||
168 | int sbal_idx, sbale_idx, sbal_no; | ||
169 | 123 | ||
170 | if (unlikely(qdio_err)) { | 124 | if (unlikely(qdio_err)) { |
171 | zfcp_hba_dbf_event_qdio(adapter, qdio_err, first, count); | 125 | zfcp_dbf_hba_qdio(qdio->adapter->dbf, qdio_err, first, |
172 | zfcp_qdio_handler_error(adapter, "qdires1"); | 126 | count); |
127 | zfcp_qdio_handler_error(qdio, "qdires1"); | ||
173 | return; | 128 | return; |
174 | } | 129 | } |
175 | 130 | ||
@@ -179,39 +134,27 @@ static void zfcp_qdio_int_resp(struct ccw_device *cdev, unsigned int qdio_err, | |||
179 | */ | 134 | */ |
180 | for (sbal_no = 0; sbal_no < count; sbal_no++) { | 135 | for (sbal_no = 0; sbal_no < count; sbal_no++) { |
181 | sbal_idx = (first + sbal_no) % QDIO_MAX_BUFFERS_PER_Q; | 136 | sbal_idx = (first + sbal_no) % QDIO_MAX_BUFFERS_PER_Q; |
182 | |||
183 | /* go through all SBALEs of SBAL */ | 137 | /* go through all SBALEs of SBAL */ |
184 | for (sbale_idx = 0; sbale_idx < QDIO_MAX_ELEMENTS_PER_BUFFER; | 138 | zfcp_fsf_reqid_check(qdio, sbal_idx); |
185 | sbale_idx++) { | ||
186 | sbale = zfcp_qdio_sbale(queue, sbal_idx, sbale_idx); | ||
187 | zfcp_qdio_reqid_check(adapter, | ||
188 | (unsigned long) sbale->addr, | ||
189 | sbal_idx); | ||
190 | if (likely(sbale->flags & SBAL_FLAGS_LAST_ENTRY)) | ||
191 | break; | ||
192 | }; | ||
193 | |||
194 | if (unlikely(!(sbale->flags & SBAL_FLAGS_LAST_ENTRY))) | ||
195 | dev_warn(&adapter->ccw_device->dev, | ||
196 | "A QDIO protocol error occurred, " | ||
197 | "operations continue\n"); | ||
198 | } | 139 | } |
199 | 140 | ||
200 | /* | 141 | /* |
201 | * put range of SBALs back to response queue | 142 | * put range of SBALs back to response queue |
202 | * (including SBALs which have already been free before) | 143 | * (including SBALs which have already been free before) |
203 | */ | 144 | */ |
204 | zfcp_qdio_resp_put_back(adapter, count); | 145 | zfcp_qdio_resp_put_back(qdio, count); |
205 | } | 146 | } |
206 | 147 | ||
207 | /** | 148 | /** |
208 | * zfcp_qdio_sbale_req - return ptr to SBALE of req_q for a struct zfcp_fsf_req | 149 | * zfcp_qdio_sbale_req - return ptr to SBALE of req_q for a struct zfcp_fsf_req |
209 | * @fsf_req: pointer to struct fsf_req | 150 | * @qdio: pointer to struct zfcp_qdio |
151 | * @q_rec: pointer to struct zfcp_queue_rec | ||
210 | * Returns: pointer to qdio_buffer_element (SBALE) structure | 152 | * Returns: pointer to qdio_buffer_element (SBALE) structure |
211 | */ | 153 | */ |
212 | struct qdio_buffer_element *zfcp_qdio_sbale_req(struct zfcp_fsf_req *req) | 154 | struct qdio_buffer_element *zfcp_qdio_sbale_req(struct zfcp_qdio *qdio, |
155 | struct zfcp_queue_req *q_req) | ||
213 | { | 156 | { |
214 | return zfcp_qdio_sbale(&req->adapter->req_q, req->sbal_last, 0); | 157 | return zfcp_qdio_sbale(&qdio->req_q, q_req->sbal_last, 0); |
215 | } | 158 | } |
216 | 159 | ||
217 | /** | 160 | /** |
@@ -219,74 +162,80 @@ struct qdio_buffer_element *zfcp_qdio_sbale_req(struct zfcp_fsf_req *req) | |||
219 | * @fsf_req: pointer to struct fsf_req | 162 | * @fsf_req: pointer to struct fsf_req |
220 | * Returns: pointer to qdio_buffer_element (SBALE) structure | 163 | * Returns: pointer to qdio_buffer_element (SBALE) structure |
221 | */ | 164 | */ |
222 | struct qdio_buffer_element *zfcp_qdio_sbale_curr(struct zfcp_fsf_req *req) | 165 | struct qdio_buffer_element *zfcp_qdio_sbale_curr(struct zfcp_qdio *qdio, |
166 | struct zfcp_queue_req *q_req) | ||
223 | { | 167 | { |
224 | return zfcp_qdio_sbale(&req->adapter->req_q, req->sbal_last, | 168 | return zfcp_qdio_sbale(&qdio->req_q, q_req->sbal_last, |
225 | req->sbale_curr); | 169 | q_req->sbale_curr); |
226 | } | 170 | } |
227 | 171 | ||
228 | static void zfcp_qdio_sbal_limit(struct zfcp_fsf_req *fsf_req, int max_sbals) | 172 | static void zfcp_qdio_sbal_limit(struct zfcp_qdio *qdio, |
173 | struct zfcp_queue_req *q_req, int max_sbals) | ||
229 | { | 174 | { |
230 | int count = atomic_read(&fsf_req->adapter->req_q.count); | 175 | int count = atomic_read(&qdio->req_q.count); |
231 | count = min(count, max_sbals); | 176 | count = min(count, max_sbals); |
232 | fsf_req->sbal_limit = (fsf_req->sbal_first + count - 1) | 177 | q_req->sbal_limit = (q_req->sbal_first + count - 1) |
233 | % QDIO_MAX_BUFFERS_PER_Q; | 178 | % QDIO_MAX_BUFFERS_PER_Q; |
234 | } | 179 | } |
235 | 180 | ||
236 | static struct qdio_buffer_element * | 181 | static struct qdio_buffer_element * |
237 | zfcp_qdio_sbal_chain(struct zfcp_fsf_req *fsf_req, unsigned long sbtype) | 182 | zfcp_qdio_sbal_chain(struct zfcp_qdio *qdio, struct zfcp_queue_req *q_req, |
183 | unsigned long sbtype) | ||
238 | { | 184 | { |
239 | struct qdio_buffer_element *sbale; | 185 | struct qdio_buffer_element *sbale; |
240 | 186 | ||
241 | /* set last entry flag in current SBALE of current SBAL */ | 187 | /* set last entry flag in current SBALE of current SBAL */ |
242 | sbale = zfcp_qdio_sbale_curr(fsf_req); | 188 | sbale = zfcp_qdio_sbale_curr(qdio, q_req); |
243 | sbale->flags |= SBAL_FLAGS_LAST_ENTRY; | 189 | sbale->flags |= SBAL_FLAGS_LAST_ENTRY; |
244 | 190 | ||
245 | /* don't exceed last allowed SBAL */ | 191 | /* don't exceed last allowed SBAL */ |
246 | if (fsf_req->sbal_last == fsf_req->sbal_limit) | 192 | if (q_req->sbal_last == q_req->sbal_limit) |
247 | return NULL; | 193 | return NULL; |
248 | 194 | ||
249 | /* set chaining flag in first SBALE of current SBAL */ | 195 | /* set chaining flag in first SBALE of current SBAL */ |
250 | sbale = zfcp_qdio_sbale_req(fsf_req); | 196 | sbale = zfcp_qdio_sbale_req(qdio, q_req); |
251 | sbale->flags |= SBAL_FLAGS0_MORE_SBALS; | 197 | sbale->flags |= SBAL_FLAGS0_MORE_SBALS; |
252 | 198 | ||
253 | /* calculate index of next SBAL */ | 199 | /* calculate index of next SBAL */ |
254 | fsf_req->sbal_last++; | 200 | q_req->sbal_last++; |
255 | fsf_req->sbal_last %= QDIO_MAX_BUFFERS_PER_Q; | 201 | q_req->sbal_last %= QDIO_MAX_BUFFERS_PER_Q; |
256 | 202 | ||
257 | /* keep this requests number of SBALs up-to-date */ | 203 | /* keep this requests number of SBALs up-to-date */ |
258 | fsf_req->sbal_number++; | 204 | q_req->sbal_number++; |
259 | 205 | ||
260 | /* start at first SBALE of new SBAL */ | 206 | /* start at first SBALE of new SBAL */ |
261 | fsf_req->sbale_curr = 0; | 207 | q_req->sbale_curr = 0; |
262 | 208 | ||
263 | /* set storage-block type for new SBAL */ | 209 | /* set storage-block type for new SBAL */ |
264 | sbale = zfcp_qdio_sbale_curr(fsf_req); | 210 | sbale = zfcp_qdio_sbale_curr(qdio, q_req); |
265 | sbale->flags |= sbtype; | 211 | sbale->flags |= sbtype; |
266 | 212 | ||
267 | return sbale; | 213 | return sbale; |
268 | } | 214 | } |
269 | 215 | ||
270 | static struct qdio_buffer_element * | 216 | static struct qdio_buffer_element * |
271 | zfcp_qdio_sbale_next(struct zfcp_fsf_req *fsf_req, unsigned long sbtype) | 217 | zfcp_qdio_sbale_next(struct zfcp_qdio *qdio, struct zfcp_queue_req *q_req, |
218 | unsigned int sbtype) | ||
272 | { | 219 | { |
273 | if (fsf_req->sbale_curr == ZFCP_LAST_SBALE_PER_SBAL) | 220 | if (q_req->sbale_curr == ZFCP_LAST_SBALE_PER_SBAL) |
274 | return zfcp_qdio_sbal_chain(fsf_req, sbtype); | 221 | return zfcp_qdio_sbal_chain(qdio, q_req, sbtype); |
275 | fsf_req->sbale_curr++; | 222 | q_req->sbale_curr++; |
276 | return zfcp_qdio_sbale_curr(fsf_req); | 223 | return zfcp_qdio_sbale_curr(qdio, q_req); |
277 | } | 224 | } |
278 | 225 | ||
279 | static void zfcp_qdio_undo_sbals(struct zfcp_fsf_req *fsf_req) | 226 | static void zfcp_qdio_undo_sbals(struct zfcp_qdio *qdio, |
227 | struct zfcp_queue_req *q_req) | ||
280 | { | 228 | { |
281 | struct qdio_buffer **sbal = fsf_req->adapter->req_q.sbal; | 229 | struct qdio_buffer **sbal = qdio->req_q.sbal; |
282 | int first = fsf_req->sbal_first; | 230 | int first = q_req->sbal_first; |
283 | int last = fsf_req->sbal_last; | 231 | int last = q_req->sbal_last; |
284 | int count = (last - first + QDIO_MAX_BUFFERS_PER_Q) % | 232 | int count = (last - first + QDIO_MAX_BUFFERS_PER_Q) % |
285 | QDIO_MAX_BUFFERS_PER_Q + 1; | 233 | QDIO_MAX_BUFFERS_PER_Q + 1; |
286 | zfcp_qdio_zero_sbals(sbal, first, count); | 234 | zfcp_qdio_zero_sbals(sbal, first, count); |
287 | } | 235 | } |
288 | 236 | ||
289 | static int zfcp_qdio_fill_sbals(struct zfcp_fsf_req *fsf_req, | 237 | static int zfcp_qdio_fill_sbals(struct zfcp_qdio *qdio, |
238 | struct zfcp_queue_req *q_req, | ||
290 | unsigned int sbtype, void *start_addr, | 239 | unsigned int sbtype, void *start_addr, |
291 | unsigned int total_length) | 240 | unsigned int total_length) |
292 | { | 241 | { |
@@ -297,10 +246,10 @@ static int zfcp_qdio_fill_sbals(struct zfcp_fsf_req *fsf_req, | |||
297 | /* split segment up */ | 246 | /* split segment up */ |
298 | for (addr = start_addr, remaining = total_length; remaining > 0; | 247 | for (addr = start_addr, remaining = total_length; remaining > 0; |
299 | addr += length, remaining -= length) { | 248 | addr += length, remaining -= length) { |
300 | sbale = zfcp_qdio_sbale_next(fsf_req, sbtype); | 249 | sbale = zfcp_qdio_sbale_next(qdio, q_req, sbtype); |
301 | if (!sbale) { | 250 | if (!sbale) { |
302 | atomic_inc(&fsf_req->adapter->qdio_outb_full); | 251 | atomic_inc(&qdio->req_q_full); |
303 | zfcp_qdio_undo_sbals(fsf_req); | 252 | zfcp_qdio_undo_sbals(qdio, q_req); |
304 | return -EINVAL; | 253 | return -EINVAL; |
305 | } | 254 | } |
306 | 255 | ||
@@ -322,29 +271,31 @@ static int zfcp_qdio_fill_sbals(struct zfcp_fsf_req *fsf_req, | |||
322 | * @max_sbals: upper bound for number of SBALs to be used | 271 | * @max_sbals: upper bound for number of SBALs to be used |
323 | * Returns: number of bytes, or error (negativ) | 272 | * Returns: number of bytes, or error (negativ) |
324 | */ | 273 | */ |
325 | int zfcp_qdio_sbals_from_sg(struct zfcp_fsf_req *fsf_req, unsigned long sbtype, | 274 | int zfcp_qdio_sbals_from_sg(struct zfcp_qdio *qdio, |
326 | struct scatterlist *sg, int max_sbals) | 275 | struct zfcp_queue_req *q_req, |
276 | unsigned long sbtype, struct scatterlist *sg, | ||
277 | int max_sbals) | ||
327 | { | 278 | { |
328 | struct qdio_buffer_element *sbale; | 279 | struct qdio_buffer_element *sbale; |
329 | int retval, bytes = 0; | 280 | int retval, bytes = 0; |
330 | 281 | ||
331 | /* figure out last allowed SBAL */ | 282 | /* figure out last allowed SBAL */ |
332 | zfcp_qdio_sbal_limit(fsf_req, max_sbals); | 283 | zfcp_qdio_sbal_limit(qdio, q_req, max_sbals); |
333 | 284 | ||
334 | /* set storage-block type for this request */ | 285 | /* set storage-block type for this request */ |
335 | sbale = zfcp_qdio_sbale_req(fsf_req); | 286 | sbale = zfcp_qdio_sbale_req(qdio, q_req); |
336 | sbale->flags |= sbtype; | 287 | sbale->flags |= sbtype; |
337 | 288 | ||
338 | for (; sg; sg = sg_next(sg)) { | 289 | for (; sg; sg = sg_next(sg)) { |
339 | retval = zfcp_qdio_fill_sbals(fsf_req, sbtype, sg_virt(sg), | 290 | retval = zfcp_qdio_fill_sbals(qdio, q_req, sbtype, |
340 | sg->length); | 291 | sg_virt(sg), sg->length); |
341 | if (retval < 0) | 292 | if (retval < 0) |
342 | return retval; | 293 | return retval; |
343 | bytes += sg->length; | 294 | bytes += sg->length; |
344 | } | 295 | } |
345 | 296 | ||
346 | /* assume that no other SBALEs are to follow in the same SBAL */ | 297 | /* assume that no other SBALEs are to follow in the same SBAL */ |
347 | sbale = zfcp_qdio_sbale_curr(fsf_req); | 298 | sbale = zfcp_qdio_sbale_curr(qdio, q_req); |
348 | sbale->flags |= SBAL_FLAGS_LAST_ENTRY; | 299 | sbale->flags |= SBAL_FLAGS_LAST_ENTRY; |
349 | 300 | ||
350 | return bytes; | 301 | return bytes; |
@@ -352,21 +303,22 @@ int zfcp_qdio_sbals_from_sg(struct zfcp_fsf_req *fsf_req, unsigned long sbtype, | |||
352 | 303 | ||
353 | /** | 304 | /** |
354 | * zfcp_qdio_send - set PCI flag in first SBALE and send req to QDIO | 305 | * zfcp_qdio_send - set PCI flag in first SBALE and send req to QDIO |
355 | * @fsf_req: pointer to struct zfcp_fsf_req | 306 | * @qdio: pointer to struct zfcp_qdio |
307 | * @q_req: pointer to struct zfcp_queue_req | ||
356 | * Returns: 0 on success, error otherwise | 308 | * Returns: 0 on success, error otherwise |
357 | */ | 309 | */ |
358 | int zfcp_qdio_send(struct zfcp_fsf_req *fsf_req) | 310 | int zfcp_qdio_send(struct zfcp_qdio *qdio, struct zfcp_queue_req *q_req) |
359 | { | 311 | { |
360 | struct zfcp_adapter *adapter = fsf_req->adapter; | 312 | struct zfcp_qdio_queue *req_q = &qdio->req_q; |
361 | struct zfcp_qdio_queue *req_q = &adapter->req_q; | 313 | int first = q_req->sbal_first; |
362 | int first = fsf_req->sbal_first; | 314 | int count = q_req->sbal_number; |
363 | int count = fsf_req->sbal_number; | ||
364 | int retval; | 315 | int retval; |
365 | unsigned int qdio_flags = QDIO_FLAG_SYNC_OUTPUT; | 316 | unsigned int qdio_flags = QDIO_FLAG_SYNC_OUTPUT; |
366 | 317 | ||
367 | zfcp_qdio_account(adapter); | 318 | zfcp_qdio_account(qdio); |
368 | 319 | ||
369 | retval = do_QDIO(adapter->ccw_device, qdio_flags, 0, first, count); | 320 | retval = do_QDIO(qdio->adapter->ccw_device, qdio_flags, 0, first, |
321 | count); | ||
370 | if (unlikely(retval)) { | 322 | if (unlikely(retval)) { |
371 | zfcp_qdio_zero_sbals(req_q->sbal, first, count); | 323 | zfcp_qdio_zero_sbals(req_q->sbal, first, count); |
372 | return retval; | 324 | return retval; |
@@ -379,63 +331,69 @@ int zfcp_qdio_send(struct zfcp_fsf_req *fsf_req) | |||
379 | return 0; | 331 | return 0; |
380 | } | 332 | } |
381 | 333 | ||
334 | |||
335 | static void zfcp_qdio_setup_init_data(struct qdio_initialize *id, | ||
336 | struct zfcp_qdio *qdio) | ||
337 | { | ||
338 | |||
339 | id->cdev = qdio->adapter->ccw_device; | ||
340 | id->q_format = QDIO_ZFCP_QFMT; | ||
341 | memcpy(id->adapter_name, dev_name(&id->cdev->dev), 8); | ||
342 | ASCEBC(id->adapter_name, 8); | ||
343 | id->qib_param_field_format = 0; | ||
344 | id->qib_param_field = NULL; | ||
345 | id->input_slib_elements = NULL; | ||
346 | id->output_slib_elements = NULL; | ||
347 | id->no_input_qs = 1; | ||
348 | id->no_output_qs = 1; | ||
349 | id->input_handler = zfcp_qdio_int_resp; | ||
350 | id->output_handler = zfcp_qdio_int_req; | ||
351 | id->int_parm = (unsigned long) qdio; | ||
352 | id->flags = QDIO_INBOUND_0COPY_SBALS | | ||
353 | QDIO_OUTBOUND_0COPY_SBALS | QDIO_USE_OUTBOUND_PCIS; | ||
354 | id->input_sbal_addr_array = (void **) (qdio->resp_q.sbal); | ||
355 | id->output_sbal_addr_array = (void **) (qdio->req_q.sbal); | ||
356 | |||
357 | } | ||
382 | /** | 358 | /** |
383 | * zfcp_qdio_allocate - allocate queue memory and initialize QDIO data | 359 | * zfcp_qdio_allocate - allocate queue memory and initialize QDIO data |
384 | * @adapter: pointer to struct zfcp_adapter | 360 | * @adapter: pointer to struct zfcp_adapter |
385 | * Returns: -ENOMEM on memory allocation error or return value from | 361 | * Returns: -ENOMEM on memory allocation error or return value from |
386 | * qdio_allocate | 362 | * qdio_allocate |
387 | */ | 363 | */ |
388 | int zfcp_qdio_allocate(struct zfcp_adapter *adapter) | 364 | static int zfcp_qdio_allocate(struct zfcp_qdio *qdio) |
389 | { | 365 | { |
390 | struct qdio_initialize *init_data; | 366 | struct qdio_initialize init_data; |
391 | 367 | ||
392 | if (zfcp_qdio_buffers_enqueue(adapter->req_q.sbal) || | 368 | if (zfcp_qdio_buffers_enqueue(qdio->req_q.sbal) || |
393 | zfcp_qdio_buffers_enqueue(adapter->resp_q.sbal)) | 369 | zfcp_qdio_buffers_enqueue(qdio->resp_q.sbal)) |
394 | return -ENOMEM; | 370 | return -ENOMEM; |
395 | 371 | ||
396 | init_data = &adapter->qdio_init_data; | 372 | zfcp_qdio_setup_init_data(&init_data, qdio); |
397 | 373 | ||
398 | init_data->cdev = adapter->ccw_device; | 374 | return qdio_allocate(&init_data); |
399 | init_data->q_format = QDIO_ZFCP_QFMT; | ||
400 | memcpy(init_data->adapter_name, dev_name(&adapter->ccw_device->dev), 8); | ||
401 | ASCEBC(init_data->adapter_name, 8); | ||
402 | init_data->qib_param_field_format = 0; | ||
403 | init_data->qib_param_field = NULL; | ||
404 | init_data->input_slib_elements = NULL; | ||
405 | init_data->output_slib_elements = NULL; | ||
406 | init_data->no_input_qs = 1; | ||
407 | init_data->no_output_qs = 1; | ||
408 | init_data->input_handler = zfcp_qdio_int_resp; | ||
409 | init_data->output_handler = zfcp_qdio_int_req; | ||
410 | init_data->int_parm = (unsigned long) adapter; | ||
411 | init_data->flags = QDIO_INBOUND_0COPY_SBALS | | ||
412 | QDIO_OUTBOUND_0COPY_SBALS | QDIO_USE_OUTBOUND_PCIS; | ||
413 | init_data->input_sbal_addr_array = | ||
414 | (void **) (adapter->resp_q.sbal); | ||
415 | init_data->output_sbal_addr_array = | ||
416 | (void **) (adapter->req_q.sbal); | ||
417 | |||
418 | return qdio_allocate(init_data); | ||
419 | } | 375 | } |
420 | 376 | ||
421 | /** | 377 | /** |
422 | * zfcp_close_qdio - close qdio queues for an adapter | 378 | * zfcp_close_qdio - close qdio queues for an adapter |
379 | * @qdio: pointer to structure zfcp_qdio | ||
423 | */ | 380 | */ |
424 | void zfcp_qdio_close(struct zfcp_adapter *adapter) | 381 | void zfcp_qdio_close(struct zfcp_qdio *qdio) |
425 | { | 382 | { |
426 | struct zfcp_qdio_queue *req_q; | 383 | struct zfcp_qdio_queue *req_q; |
427 | int first, count; | 384 | int first, count; |
428 | 385 | ||
429 | if (!(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP)) | 386 | if (!(atomic_read(&qdio->adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP)) |
430 | return; | 387 | return; |
431 | 388 | ||
432 | /* clear QDIOUP flag, thus do_QDIO is not called during qdio_shutdown */ | 389 | /* clear QDIOUP flag, thus do_QDIO is not called during qdio_shutdown */ |
433 | req_q = &adapter->req_q; | 390 | req_q = &qdio->req_q; |
434 | spin_lock_bh(&adapter->req_q_lock); | 391 | spin_lock_bh(&qdio->req_q_lock); |
435 | atomic_clear_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status); | 392 | atomic_clear_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &qdio->adapter->status); |
436 | spin_unlock_bh(&adapter->req_q_lock); | 393 | spin_unlock_bh(&qdio->req_q_lock); |
437 | 394 | ||
438 | qdio_shutdown(adapter->ccw_device, QDIO_FLAG_CLEANUP_USING_CLEAR); | 395 | qdio_shutdown(qdio->adapter->ccw_device, |
396 | QDIO_FLAG_CLEANUP_USING_CLEAR); | ||
439 | 397 | ||
440 | /* cleanup used outbound sbals */ | 398 | /* cleanup used outbound sbals */ |
441 | count = atomic_read(&req_q->count); | 399 | count = atomic_read(&req_q->count); |
@@ -446,50 +404,99 @@ void zfcp_qdio_close(struct zfcp_adapter *adapter) | |||
446 | } | 404 | } |
447 | req_q->first = 0; | 405 | req_q->first = 0; |
448 | atomic_set(&req_q->count, 0); | 406 | atomic_set(&req_q->count, 0); |
449 | adapter->resp_q.first = 0; | 407 | qdio->resp_q.first = 0; |
450 | atomic_set(&adapter->resp_q.count, 0); | 408 | atomic_set(&qdio->resp_q.count, 0); |
451 | } | 409 | } |
452 | 410 | ||
453 | /** | 411 | /** |
454 | * zfcp_qdio_open - prepare and initialize response queue | 412 | * zfcp_qdio_open - prepare and initialize response queue |
455 | * @adapter: pointer to struct zfcp_adapter | 413 | * @qdio: pointer to struct zfcp_qdio |
456 | * Returns: 0 on success, otherwise -EIO | 414 | * Returns: 0 on success, otherwise -EIO |
457 | */ | 415 | */ |
458 | int zfcp_qdio_open(struct zfcp_adapter *adapter) | 416 | int zfcp_qdio_open(struct zfcp_qdio *qdio) |
459 | { | 417 | { |
460 | struct qdio_buffer_element *sbale; | 418 | struct qdio_buffer_element *sbale; |
419 | struct qdio_initialize init_data; | ||
420 | struct ccw_device *cdev = qdio->adapter->ccw_device; | ||
461 | int cc; | 421 | int cc; |
462 | 422 | ||
463 | if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP) | 423 | if (atomic_read(&qdio->adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP) |
464 | return -EIO; | 424 | return -EIO; |
465 | 425 | ||
466 | if (qdio_establish(&adapter->qdio_init_data)) | 426 | zfcp_qdio_setup_init_data(&init_data, qdio); |
427 | |||
428 | if (qdio_establish(&init_data)) | ||
467 | goto failed_establish; | 429 | goto failed_establish; |
468 | 430 | ||
469 | if (qdio_activate(adapter->ccw_device)) | 431 | if (qdio_activate(cdev)) |
470 | goto failed_qdio; | 432 | goto failed_qdio; |
471 | 433 | ||
472 | for (cc = 0; cc < QDIO_MAX_BUFFERS_PER_Q; cc++) { | 434 | for (cc = 0; cc < QDIO_MAX_BUFFERS_PER_Q; cc++) { |
473 | sbale = &(adapter->resp_q.sbal[cc]->element[0]); | 435 | sbale = &(qdio->resp_q.sbal[cc]->element[0]); |
474 | sbale->length = 0; | 436 | sbale->length = 0; |
475 | sbale->flags = SBAL_FLAGS_LAST_ENTRY; | 437 | sbale->flags = SBAL_FLAGS_LAST_ENTRY; |
476 | sbale->addr = NULL; | 438 | sbale->addr = NULL; |
477 | } | 439 | } |
478 | 440 | ||
479 | if (do_QDIO(adapter->ccw_device, QDIO_FLAG_SYNC_INPUT, 0, 0, | 441 | if (do_QDIO(cdev, QDIO_FLAG_SYNC_INPUT, 0, 0, |
480 | QDIO_MAX_BUFFERS_PER_Q)) | 442 | QDIO_MAX_BUFFERS_PER_Q)) |
481 | goto failed_qdio; | 443 | goto failed_qdio; |
482 | 444 | ||
483 | /* set index of first avalable SBALS / number of available SBALS */ | 445 | /* set index of first avalable SBALS / number of available SBALS */ |
484 | adapter->req_q.first = 0; | 446 | qdio->req_q.first = 0; |
485 | atomic_set(&adapter->req_q.count, QDIO_MAX_BUFFERS_PER_Q); | 447 | atomic_set(&qdio->req_q.count, QDIO_MAX_BUFFERS_PER_Q); |
486 | 448 | ||
487 | return 0; | 449 | return 0; |
488 | 450 | ||
489 | failed_qdio: | 451 | failed_qdio: |
490 | qdio_shutdown(adapter->ccw_device, QDIO_FLAG_CLEANUP_USING_CLEAR); | 452 | qdio_shutdown(cdev, QDIO_FLAG_CLEANUP_USING_CLEAR); |
491 | failed_establish: | 453 | failed_establish: |
492 | dev_err(&adapter->ccw_device->dev, | 454 | dev_err(&cdev->dev, |
493 | "Setting up the QDIO connection to the FCP adapter failed\n"); | 455 | "Setting up the QDIO connection to the FCP adapter failed\n"); |
494 | return -EIO; | 456 | return -EIO; |
495 | } | 457 | } |
458 | |||
459 | void zfcp_qdio_destroy(struct zfcp_qdio *qdio) | ||
460 | { | ||
461 | struct qdio_buffer **sbal_req, **sbal_resp; | ||
462 | int p; | ||
463 | |||
464 | if (!qdio) | ||
465 | return; | ||
466 | |||
467 | if (qdio->adapter->ccw_device) | ||
468 | qdio_free(qdio->adapter->ccw_device); | ||
469 | |||
470 | sbal_req = qdio->req_q.sbal; | ||
471 | sbal_resp = qdio->resp_q.sbal; | ||
472 | |||
473 | for (p = 0; p < QDIO_MAX_BUFFERS_PER_Q; p += QBUFF_PER_PAGE) { | ||
474 | free_page((unsigned long) sbal_req[p]); | ||
475 | free_page((unsigned long) sbal_resp[p]); | ||
476 | } | ||
477 | |||
478 | kfree(qdio); | ||
479 | } | ||
480 | |||
481 | int zfcp_qdio_setup(struct zfcp_adapter *adapter) | ||
482 | { | ||
483 | struct zfcp_qdio *qdio; | ||
484 | |||
485 | qdio = kzalloc(sizeof(struct zfcp_qdio), GFP_KERNEL); | ||
486 | if (!qdio) | ||
487 | return -ENOMEM; | ||
488 | |||
489 | qdio->adapter = adapter; | ||
490 | |||
491 | if (zfcp_qdio_allocate(qdio)) { | ||
492 | zfcp_qdio_destroy(qdio); | ||
493 | return -ENOMEM; | ||
494 | } | ||
495 | |||
496 | spin_lock_init(&qdio->req_q_lock); | ||
497 | spin_lock_init(&qdio->stat_lock); | ||
498 | |||
499 | adapter->qdio = qdio; | ||
500 | return 0; | ||
501 | } | ||
502 | |||
diff --git a/drivers/s390/scsi/zfcp_scsi.c b/drivers/s390/scsi/zfcp_scsi.c index 6925a1784682..3ff726afafc6 100644 --- a/drivers/s390/scsi/zfcp_scsi.c +++ b/drivers/s390/scsi/zfcp_scsi.c | |||
@@ -9,8 +9,9 @@ | |||
9 | #define KMSG_COMPONENT "zfcp" | 9 | #define KMSG_COMPONENT "zfcp" |
10 | #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt | 10 | #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt |
11 | 11 | ||
12 | #include "zfcp_ext.h" | ||
13 | #include <asm/atomic.h> | 12 | #include <asm/atomic.h> |
13 | #include "zfcp_ext.h" | ||
14 | #include "zfcp_dbf.h" | ||
14 | 15 | ||
15 | static unsigned int default_depth = 32; | 16 | static unsigned int default_depth = 32; |
16 | module_param_named(queue_depth, default_depth, uint, 0600); | 17 | module_param_named(queue_depth, default_depth, uint, 0600); |
@@ -52,11 +53,11 @@ static int zfcp_scsi_slave_configure(struct scsi_device *sdp) | |||
52 | 53 | ||
53 | static void zfcp_scsi_command_fail(struct scsi_cmnd *scpnt, int result) | 54 | static void zfcp_scsi_command_fail(struct scsi_cmnd *scpnt, int result) |
54 | { | 55 | { |
56 | struct zfcp_adapter *adapter = | ||
57 | (struct zfcp_adapter *) scpnt->device->host->hostdata[0]; | ||
55 | set_host_byte(scpnt, result); | 58 | set_host_byte(scpnt, result); |
56 | if ((scpnt->device != NULL) && (scpnt->device->host != NULL)) | 59 | if ((scpnt->device != NULL) && (scpnt->device->host != NULL)) |
57 | zfcp_scsi_dbf_event_result("fail", 4, | 60 | zfcp_dbf_scsi_result("fail", 4, adapter->dbf, scpnt, NULL); |
58 | (struct zfcp_adapter*) scpnt->device->host->hostdata[0], | ||
59 | scpnt, NULL); | ||
60 | /* return directly */ | 61 | /* return directly */ |
61 | scpnt->scsi_done(scpnt); | 62 | scpnt->scsi_done(scpnt); |
62 | } | 63 | } |
@@ -92,7 +93,7 @@ static int zfcp_scsi_queuecommand(struct scsi_cmnd *scpnt, | |||
92 | scsi_result = fc_remote_port_chkready(rport); | 93 | scsi_result = fc_remote_port_chkready(rport); |
93 | if (unlikely(scsi_result)) { | 94 | if (unlikely(scsi_result)) { |
94 | scpnt->result = scsi_result; | 95 | scpnt->result = scsi_result; |
95 | zfcp_scsi_dbf_event_result("fail", 4, adapter, scpnt, NULL); | 96 | zfcp_dbf_scsi_result("fail", 4, adapter->dbf, scpnt, NULL); |
96 | scpnt->scsi_done(scpnt); | 97 | scpnt->scsi_done(scpnt); |
97 | return 0; | 98 | return 0; |
98 | } | 99 | } |
@@ -180,8 +181,8 @@ static int zfcp_scsi_eh_abort_handler(struct scsi_cmnd *scpnt) | |||
180 | spin_unlock(&adapter->req_list_lock); | 181 | spin_unlock(&adapter->req_list_lock); |
181 | if (!old_req) { | 182 | if (!old_req) { |
182 | write_unlock_irqrestore(&adapter->abort_lock, flags); | 183 | write_unlock_irqrestore(&adapter->abort_lock, flags); |
183 | zfcp_scsi_dbf_event_abort("lte1", adapter, scpnt, NULL, | 184 | zfcp_dbf_scsi_abort("lte1", adapter->dbf, scpnt, NULL, |
184 | old_reqid); | 185 | old_reqid); |
185 | return FAILED; /* completion could be in progress */ | 186 | return FAILED; /* completion could be in progress */ |
186 | } | 187 | } |
187 | old_req->data = NULL; | 188 | old_req->data = NULL; |
@@ -197,16 +198,15 @@ static int zfcp_scsi_eh_abort_handler(struct scsi_cmnd *scpnt) | |||
197 | zfcp_erp_wait(adapter); | 198 | zfcp_erp_wait(adapter); |
198 | if (!(atomic_read(&adapter->status) & | 199 | if (!(atomic_read(&adapter->status) & |
199 | ZFCP_STATUS_COMMON_RUNNING)) { | 200 | ZFCP_STATUS_COMMON_RUNNING)) { |
200 | zfcp_scsi_dbf_event_abort("nres", adapter, scpnt, NULL, | 201 | zfcp_dbf_scsi_abort("nres", adapter->dbf, scpnt, NULL, |
201 | old_reqid); | 202 | old_reqid); |
202 | return SUCCESS; | 203 | return SUCCESS; |
203 | } | 204 | } |
204 | } | 205 | } |
205 | if (!abrt_req) | 206 | if (!abrt_req) |
206 | return FAILED; | 207 | return FAILED; |
207 | 208 | ||
208 | wait_event(abrt_req->completion_wq, | 209 | wait_for_completion(&abrt_req->completion); |
209 | abrt_req->status & ZFCP_STATUS_FSFREQ_COMPLETED); | ||
210 | 210 | ||
211 | if (abrt_req->status & ZFCP_STATUS_FSFREQ_ABORTSUCCEEDED) | 211 | if (abrt_req->status & ZFCP_STATUS_FSFREQ_ABORTSUCCEEDED) |
212 | dbf_tag = "okay"; | 212 | dbf_tag = "okay"; |
@@ -216,7 +216,7 @@ static int zfcp_scsi_eh_abort_handler(struct scsi_cmnd *scpnt) | |||
216 | dbf_tag = "fail"; | 216 | dbf_tag = "fail"; |
217 | retval = FAILED; | 217 | retval = FAILED; |
218 | } | 218 | } |
219 | zfcp_scsi_dbf_event_abort(dbf_tag, adapter, scpnt, abrt_req, old_reqid); | 219 | zfcp_dbf_scsi_abort(dbf_tag, adapter->dbf, scpnt, abrt_req, old_reqid); |
220 | zfcp_fsf_req_free(abrt_req); | 220 | zfcp_fsf_req_free(abrt_req); |
221 | return retval; | 221 | return retval; |
222 | } | 222 | } |
@@ -225,7 +225,7 @@ static int zfcp_task_mgmt_function(struct scsi_cmnd *scpnt, u8 tm_flags) | |||
225 | { | 225 | { |
226 | struct zfcp_unit *unit = scpnt->device->hostdata; | 226 | struct zfcp_unit *unit = scpnt->device->hostdata; |
227 | struct zfcp_adapter *adapter = unit->port->adapter; | 227 | struct zfcp_adapter *adapter = unit->port->adapter; |
228 | struct zfcp_fsf_req *fsf_req; | 228 | struct zfcp_fsf_req *fsf_req = NULL; |
229 | int retval = SUCCESS; | 229 | int retval = SUCCESS; |
230 | int retry = 3; | 230 | int retry = 3; |
231 | 231 | ||
@@ -237,25 +237,23 @@ static int zfcp_task_mgmt_function(struct scsi_cmnd *scpnt, u8 tm_flags) | |||
237 | zfcp_erp_wait(adapter); | 237 | zfcp_erp_wait(adapter); |
238 | if (!(atomic_read(&adapter->status) & | 238 | if (!(atomic_read(&adapter->status) & |
239 | ZFCP_STATUS_COMMON_RUNNING)) { | 239 | ZFCP_STATUS_COMMON_RUNNING)) { |
240 | zfcp_scsi_dbf_event_devreset("nres", tm_flags, unit, | 240 | zfcp_dbf_scsi_devreset("nres", tm_flags, unit, scpnt); |
241 | scpnt); | ||
242 | return SUCCESS; | 241 | return SUCCESS; |
243 | } | 242 | } |
244 | } | 243 | } |
245 | if (!fsf_req) | 244 | if (!fsf_req) |
246 | return FAILED; | 245 | return FAILED; |
247 | 246 | ||
248 | wait_event(fsf_req->completion_wq, | 247 | wait_for_completion(&fsf_req->completion); |
249 | fsf_req->status & ZFCP_STATUS_FSFREQ_COMPLETED); | ||
250 | 248 | ||
251 | if (fsf_req->status & ZFCP_STATUS_FSFREQ_TMFUNCFAILED) { | 249 | if (fsf_req->status & ZFCP_STATUS_FSFREQ_TMFUNCFAILED) { |
252 | zfcp_scsi_dbf_event_devreset("fail", tm_flags, unit, scpnt); | 250 | zfcp_dbf_scsi_devreset("fail", tm_flags, unit, scpnt); |
253 | retval = FAILED; | 251 | retval = FAILED; |
254 | } else if (fsf_req->status & ZFCP_STATUS_FSFREQ_TMFUNCNOTSUPP) { | 252 | } else if (fsf_req->status & ZFCP_STATUS_FSFREQ_TMFUNCNOTSUPP) { |
255 | zfcp_scsi_dbf_event_devreset("nsup", tm_flags, unit, scpnt); | 253 | zfcp_dbf_scsi_devreset("nsup", tm_flags, unit, scpnt); |
256 | retval = FAILED; | 254 | retval = FAILED; |
257 | } else | 255 | } else |
258 | zfcp_scsi_dbf_event_devreset("okay", tm_flags, unit, scpnt); | 256 | zfcp_dbf_scsi_devreset("okay", tm_flags, unit, scpnt); |
259 | 257 | ||
260 | zfcp_fsf_req_free(fsf_req); | 258 | zfcp_fsf_req_free(fsf_req); |
261 | return retval; | 259 | return retval; |
@@ -430,7 +428,7 @@ static struct fc_host_statistics *zfcp_get_fc_host_stats(struct Scsi_Host *host) | |||
430 | if (!data) | 428 | if (!data) |
431 | return NULL; | 429 | return NULL; |
432 | 430 | ||
433 | ret = zfcp_fsf_exchange_port_data_sync(adapter, data); | 431 | ret = zfcp_fsf_exchange_port_data_sync(adapter->qdio, data); |
434 | if (ret) { | 432 | if (ret) { |
435 | kfree(data); | 433 | kfree(data); |
436 | return NULL; | 434 | return NULL; |
@@ -459,7 +457,7 @@ static void zfcp_reset_fc_host_stats(struct Scsi_Host *shost) | |||
459 | if (!data) | 457 | if (!data) |
460 | return; | 458 | return; |
461 | 459 | ||
462 | ret = zfcp_fsf_exchange_port_data_sync(adapter, data); | 460 | ret = zfcp_fsf_exchange_port_data_sync(adapter->qdio, data); |
463 | if (ret) | 461 | if (ret) |
464 | kfree(data); | 462 | kfree(data); |
465 | else { | 463 | else { |
@@ -493,21 +491,6 @@ static void zfcp_set_rport_dev_loss_tmo(struct fc_rport *rport, u32 timeout) | |||
493 | } | 491 | } |
494 | 492 | ||
495 | /** | 493 | /** |
496 | * zfcp_scsi_dev_loss_tmo_callbk - Free any reference to rport | ||
497 | * @rport: The rport that is about to be deleted. | ||
498 | */ | ||
499 | static void zfcp_scsi_dev_loss_tmo_callbk(struct fc_rport *rport) | ||
500 | { | ||
501 | struct zfcp_port *port; | ||
502 | |||
503 | write_lock_irq(&zfcp_data.config_lock); | ||
504 | port = rport->dd_data; | ||
505 | if (port) | ||
506 | port->rport = NULL; | ||
507 | write_unlock_irq(&zfcp_data.config_lock); | ||
508 | } | ||
509 | |||
510 | /** | ||
511 | * zfcp_scsi_terminate_rport_io - Terminate all I/O on a rport | 494 | * zfcp_scsi_terminate_rport_io - Terminate all I/O on a rport |
512 | * @rport: The FC rport where to teminate I/O | 495 | * @rport: The FC rport where to teminate I/O |
513 | * | 496 | * |
@@ -518,9 +501,12 @@ static void zfcp_scsi_dev_loss_tmo_callbk(struct fc_rport *rport) | |||
518 | static void zfcp_scsi_terminate_rport_io(struct fc_rport *rport) | 501 | static void zfcp_scsi_terminate_rport_io(struct fc_rport *rport) |
519 | { | 502 | { |
520 | struct zfcp_port *port; | 503 | struct zfcp_port *port; |
504 | struct Scsi_Host *shost = rport_to_shost(rport); | ||
505 | struct zfcp_adapter *adapter = | ||
506 | (struct zfcp_adapter *)shost->hostdata[0]; | ||
521 | 507 | ||
522 | write_lock_irq(&zfcp_data.config_lock); | 508 | write_lock_irq(&zfcp_data.config_lock); |
523 | port = rport->dd_data; | 509 | port = zfcp_get_port_by_wwpn(adapter, rport->port_name); |
524 | if (port) | 510 | if (port) |
525 | zfcp_port_get(port); | 511 | zfcp_port_get(port); |
526 | write_unlock_irq(&zfcp_data.config_lock); | 512 | write_unlock_irq(&zfcp_data.config_lock); |
@@ -552,7 +538,6 @@ static void zfcp_scsi_rport_register(struct zfcp_port *port) | |||
552 | return; | 538 | return; |
553 | } | 539 | } |
554 | 540 | ||
555 | rport->dd_data = port; | ||
556 | rport->maxframe_size = port->maxframe_size; | 541 | rport->maxframe_size = port->maxframe_size; |
557 | rport->supported_classes = port->supported_classes; | 542 | rport->supported_classes = port->supported_classes; |
558 | port->rport = rport; | 543 | port->rport = rport; |
@@ -573,7 +558,7 @@ void zfcp_scsi_schedule_rport_register(struct zfcp_port *port) | |||
573 | zfcp_port_get(port); | 558 | zfcp_port_get(port); |
574 | port->rport_task = RPORT_ADD; | 559 | port->rport_task = RPORT_ADD; |
575 | 560 | ||
576 | if (!queue_work(zfcp_data.work_queue, &port->rport_work)) | 561 | if (!queue_work(port->adapter->work_queue, &port->rport_work)) |
577 | zfcp_port_put(port); | 562 | zfcp_port_put(port); |
578 | } | 563 | } |
579 | 564 | ||
@@ -582,8 +567,11 @@ void zfcp_scsi_schedule_rport_block(struct zfcp_port *port) | |||
582 | zfcp_port_get(port); | 567 | zfcp_port_get(port); |
583 | port->rport_task = RPORT_DEL; | 568 | port->rport_task = RPORT_DEL; |
584 | 569 | ||
585 | if (!queue_work(zfcp_data.work_queue, &port->rport_work)) | 570 | if (port->rport && queue_work(port->adapter->work_queue, |
586 | zfcp_port_put(port); | 571 | &port->rport_work)) |
572 | return; | ||
573 | |||
574 | zfcp_port_put(port); | ||
587 | } | 575 | } |
588 | 576 | ||
589 | void zfcp_scsi_schedule_rports_block(struct zfcp_adapter *adapter) | 577 | void zfcp_scsi_schedule_rports_block(struct zfcp_adapter *adapter) |
@@ -662,7 +650,6 @@ struct fc_function_template zfcp_transport_functions = { | |||
662 | .reset_fc_host_stats = zfcp_reset_fc_host_stats, | 650 | .reset_fc_host_stats = zfcp_reset_fc_host_stats, |
663 | .set_rport_dev_loss_tmo = zfcp_set_rport_dev_loss_tmo, | 651 | .set_rport_dev_loss_tmo = zfcp_set_rport_dev_loss_tmo, |
664 | .get_host_port_state = zfcp_get_host_port_state, | 652 | .get_host_port_state = zfcp_get_host_port_state, |
665 | .dev_loss_tmo_callbk = zfcp_scsi_dev_loss_tmo_callbk, | ||
666 | .terminate_rport_io = zfcp_scsi_terminate_rport_io, | 653 | .terminate_rport_io = zfcp_scsi_terminate_rport_io, |
667 | .show_host_port_state = 1, | 654 | .show_host_port_state = 1, |
668 | .bsg_request = zfcp_execute_fc_job, | 655 | .bsg_request = zfcp_execute_fc_job, |
diff --git a/drivers/s390/scsi/zfcp_sysfs.c b/drivers/s390/scsi/zfcp_sysfs.c index 0fe5cce818cb..079a8cf518a3 100644 --- a/drivers/s390/scsi/zfcp_sysfs.c +++ b/drivers/s390/scsi/zfcp_sysfs.c | |||
@@ -88,7 +88,7 @@ static ssize_t zfcp_sysfs_##_feat##_failed_store(struct device *dev, \ | |||
88 | unsigned long val; \ | 88 | unsigned long val; \ |
89 | int retval = 0; \ | 89 | int retval = 0; \ |
90 | \ | 90 | \ |
91 | down(&zfcp_data.config_sema); \ | 91 | mutex_lock(&zfcp_data.config_mutex); \ |
92 | if (atomic_read(&_feat->status) & ZFCP_STATUS_COMMON_REMOVE) { \ | 92 | if (atomic_read(&_feat->status) & ZFCP_STATUS_COMMON_REMOVE) { \ |
93 | retval = -EBUSY; \ | 93 | retval = -EBUSY; \ |
94 | goto out; \ | 94 | goto out; \ |
@@ -105,7 +105,7 @@ static ssize_t zfcp_sysfs_##_feat##_failed_store(struct device *dev, \ | |||
105 | _reopen_id, NULL); \ | 105 | _reopen_id, NULL); \ |
106 | zfcp_erp_wait(_adapter); \ | 106 | zfcp_erp_wait(_adapter); \ |
107 | out: \ | 107 | out: \ |
108 | up(&zfcp_data.config_sema); \ | 108 | mutex_unlock(&zfcp_data.config_mutex); \ |
109 | return retval ? retval : (ssize_t) count; \ | 109 | return retval ? retval : (ssize_t) count; \ |
110 | } \ | 110 | } \ |
111 | static ZFCP_DEV_ATTR(_feat, failed, S_IWUSR | S_IRUGO, \ | 111 | static ZFCP_DEV_ATTR(_feat, failed, S_IWUSR | S_IRUGO, \ |
@@ -126,7 +126,7 @@ static ssize_t zfcp_sysfs_port_rescan_store(struct device *dev, | |||
126 | if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_REMOVE) | 126 | if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_REMOVE) |
127 | return -EBUSY; | 127 | return -EBUSY; |
128 | 128 | ||
129 | ret = zfcp_scan_ports(adapter); | 129 | ret = zfcp_fc_scan_ports(adapter); |
130 | return ret ? ret : (ssize_t) count; | 130 | return ret ? ret : (ssize_t) count; |
131 | } | 131 | } |
132 | static ZFCP_DEV_ATTR(adapter, port_rescan, S_IWUSR, NULL, | 132 | static ZFCP_DEV_ATTR(adapter, port_rescan, S_IWUSR, NULL, |
@@ -142,7 +142,7 @@ static ssize_t zfcp_sysfs_port_remove_store(struct device *dev, | |||
142 | int retval = 0; | 142 | int retval = 0; |
143 | LIST_HEAD(port_remove_lh); | 143 | LIST_HEAD(port_remove_lh); |
144 | 144 | ||
145 | down(&zfcp_data.config_sema); | 145 | mutex_lock(&zfcp_data.config_mutex); |
146 | if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_REMOVE) { | 146 | if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_REMOVE) { |
147 | retval = -EBUSY; | 147 | retval = -EBUSY; |
148 | goto out; | 148 | goto out; |
@@ -173,7 +173,7 @@ static ssize_t zfcp_sysfs_port_remove_store(struct device *dev, | |||
173 | zfcp_port_put(port); | 173 | zfcp_port_put(port); |
174 | zfcp_port_dequeue(port); | 174 | zfcp_port_dequeue(port); |
175 | out: | 175 | out: |
176 | up(&zfcp_data.config_sema); | 176 | mutex_unlock(&zfcp_data.config_mutex); |
177 | return retval ? retval : (ssize_t) count; | 177 | return retval ? retval : (ssize_t) count; |
178 | } | 178 | } |
179 | static ZFCP_DEV_ATTR(adapter, port_remove, S_IWUSR, NULL, | 179 | static ZFCP_DEV_ATTR(adapter, port_remove, S_IWUSR, NULL, |
@@ -207,7 +207,7 @@ static ssize_t zfcp_sysfs_unit_add_store(struct device *dev, | |||
207 | u64 fcp_lun; | 207 | u64 fcp_lun; |
208 | int retval = -EINVAL; | 208 | int retval = -EINVAL; |
209 | 209 | ||
210 | down(&zfcp_data.config_sema); | 210 | mutex_lock(&zfcp_data.config_mutex); |
211 | if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_REMOVE) { | 211 | if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_REMOVE) { |
212 | retval = -EBUSY; | 212 | retval = -EBUSY; |
213 | goto out; | 213 | goto out; |
@@ -226,7 +226,7 @@ static ssize_t zfcp_sysfs_unit_add_store(struct device *dev, | |||
226 | zfcp_erp_wait(unit->port->adapter); | 226 | zfcp_erp_wait(unit->port->adapter); |
227 | zfcp_unit_put(unit); | 227 | zfcp_unit_put(unit); |
228 | out: | 228 | out: |
229 | up(&zfcp_data.config_sema); | 229 | mutex_unlock(&zfcp_data.config_mutex); |
230 | return retval ? retval : (ssize_t) count; | 230 | return retval ? retval : (ssize_t) count; |
231 | } | 231 | } |
232 | static DEVICE_ATTR(unit_add, S_IWUSR, NULL, zfcp_sysfs_unit_add_store); | 232 | static DEVICE_ATTR(unit_add, S_IWUSR, NULL, zfcp_sysfs_unit_add_store); |
@@ -241,7 +241,7 @@ static ssize_t zfcp_sysfs_unit_remove_store(struct device *dev, | |||
241 | int retval = 0; | 241 | int retval = 0; |
242 | LIST_HEAD(unit_remove_lh); | 242 | LIST_HEAD(unit_remove_lh); |
243 | 243 | ||
244 | down(&zfcp_data.config_sema); | 244 | mutex_lock(&zfcp_data.config_mutex); |
245 | if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_REMOVE) { | 245 | if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_REMOVE) { |
246 | retval = -EBUSY; | 246 | retval = -EBUSY; |
247 | goto out; | 247 | goto out; |
@@ -282,7 +282,7 @@ static ssize_t zfcp_sysfs_unit_remove_store(struct device *dev, | |||
282 | zfcp_unit_put(unit); | 282 | zfcp_unit_put(unit); |
283 | zfcp_unit_dequeue(unit); | 283 | zfcp_unit_dequeue(unit); |
284 | out: | 284 | out: |
285 | up(&zfcp_data.config_sema); | 285 | mutex_unlock(&zfcp_data.config_mutex); |
286 | return retval ? retval : (ssize_t) count; | 286 | return retval ? retval : (ssize_t) count; |
287 | } | 287 | } |
288 | static DEVICE_ATTR(unit_remove, S_IWUSR, NULL, zfcp_sysfs_unit_remove_store); | 288 | static DEVICE_ATTR(unit_remove, S_IWUSR, NULL, zfcp_sysfs_unit_remove_store); |
@@ -425,7 +425,7 @@ static ssize_t zfcp_sysfs_adapter_util_show(struct device *dev, | |||
425 | if (!qtcb_port) | 425 | if (!qtcb_port) |
426 | return -ENOMEM; | 426 | return -ENOMEM; |
427 | 427 | ||
428 | retval = zfcp_fsf_exchange_port_data_sync(adapter, qtcb_port); | 428 | retval = zfcp_fsf_exchange_port_data_sync(adapter->qdio, qtcb_port); |
429 | if (!retval) | 429 | if (!retval) |
430 | retval = sprintf(buf, "%u %u %u\n", qtcb_port->cp_util, | 430 | retval = sprintf(buf, "%u %u %u\n", qtcb_port->cp_util, |
431 | qtcb_port->cb_util, qtcb_port->a_util); | 431 | qtcb_port->cb_util, qtcb_port->a_util); |
@@ -451,7 +451,7 @@ static int zfcp_sysfs_adapter_ex_config(struct device *dev, | |||
451 | if (!qtcb_config) | 451 | if (!qtcb_config) |
452 | return -ENOMEM; | 452 | return -ENOMEM; |
453 | 453 | ||
454 | retval = zfcp_fsf_exchange_config_data_sync(adapter, qtcb_config); | 454 | retval = zfcp_fsf_exchange_config_data_sync(adapter->qdio, qtcb_config); |
455 | if (!retval) | 455 | if (!retval) |
456 | *stat_inf = qtcb_config->stat_info; | 456 | *stat_inf = qtcb_config->stat_info; |
457 | 457 | ||
@@ -492,15 +492,15 @@ static ssize_t zfcp_sysfs_adapter_q_full_show(struct device *dev, | |||
492 | char *buf) | 492 | char *buf) |
493 | { | 493 | { |
494 | struct Scsi_Host *scsi_host = class_to_shost(dev); | 494 | struct Scsi_Host *scsi_host = class_to_shost(dev); |
495 | struct zfcp_adapter *adapter = | 495 | struct zfcp_qdio *qdio = |
496 | (struct zfcp_adapter *) scsi_host->hostdata[0]; | 496 | ((struct zfcp_adapter *) scsi_host->hostdata[0])->qdio; |
497 | u64 util; | 497 | u64 util; |
498 | 498 | ||
499 | spin_lock_bh(&adapter->qdio_stat_lock); | 499 | spin_lock_bh(&qdio->stat_lock); |
500 | util = adapter->req_q_util; | 500 | util = qdio->req_q_util; |
501 | spin_unlock_bh(&adapter->qdio_stat_lock); | 501 | spin_unlock_bh(&qdio->stat_lock); |
502 | 502 | ||
503 | return sprintf(buf, "%d %llu\n", atomic_read(&adapter->qdio_outb_full), | 503 | return sprintf(buf, "%d %llu\n", atomic_read(&qdio->req_q_full), |
504 | (unsigned long long)util); | 504 | (unsigned long long)util); |
505 | } | 505 | } |
506 | static DEVICE_ATTR(queue_full, S_IRUGO, zfcp_sysfs_adapter_q_full_show, NULL); | 506 | static DEVICE_ATTR(queue_full, S_IRUGO, zfcp_sysfs_adapter_q_full_show, NULL); |
diff --git a/drivers/scsi/Kconfig b/drivers/scsi/Kconfig index 9c23122f755f..82bb3b2d207a 100644 --- a/drivers/scsi/Kconfig +++ b/drivers/scsi/Kconfig | |||
@@ -1811,6 +1811,12 @@ config ZFCP | |||
1811 | called zfcp. If you want to compile it as a module, say M here | 1811 | called zfcp. If you want to compile it as a module, say M here |
1812 | and read <file:Documentation/kbuild/modules.txt>. | 1812 | and read <file:Documentation/kbuild/modules.txt>. |
1813 | 1813 | ||
1814 | config SCSI_PMCRAID | ||
1815 | tristate "PMC SIERRA Linux MaxRAID adapter support" | ||
1816 | depends on PCI && SCSI | ||
1817 | ---help--- | ||
1818 | This driver supports the PMC SIERRA MaxRAID adapters. | ||
1819 | |||
1814 | config SCSI_SRP | 1820 | config SCSI_SRP |
1815 | tristate "SCSI RDMA Protocol helper library" | 1821 | tristate "SCSI RDMA Protocol helper library" |
1816 | depends on SCSI && PCI | 1822 | depends on SCSI && PCI |
diff --git a/drivers/scsi/Makefile b/drivers/scsi/Makefile index 25429ea63d0a..61a94af3cee7 100644 --- a/drivers/scsi/Makefile +++ b/drivers/scsi/Makefile | |||
@@ -130,6 +130,7 @@ obj-$(CONFIG_SCSI_MVSAS) += mvsas/ | |||
130 | obj-$(CONFIG_PS3_ROM) += ps3rom.o | 130 | obj-$(CONFIG_PS3_ROM) += ps3rom.o |
131 | obj-$(CONFIG_SCSI_CXGB3_ISCSI) += libiscsi.o libiscsi_tcp.o cxgb3i/ | 131 | obj-$(CONFIG_SCSI_CXGB3_ISCSI) += libiscsi.o libiscsi_tcp.o cxgb3i/ |
132 | obj-$(CONFIG_SCSI_BNX2_ISCSI) += libiscsi.o bnx2i/ | 132 | obj-$(CONFIG_SCSI_BNX2_ISCSI) += libiscsi.o bnx2i/ |
133 | obj-$(CONFIG_SCSI_PMCRAID) += pmcraid.o | ||
133 | 134 | ||
134 | obj-$(CONFIG_ARM) += arm/ | 135 | obj-$(CONFIG_ARM) += arm/ |
135 | 136 | ||
diff --git a/drivers/scsi/bnx2i/bnx2i_init.c b/drivers/scsi/bnx2i/bnx2i_init.c index ae4b2d588fd3..0c4210d48ee8 100644 --- a/drivers/scsi/bnx2i/bnx2i_init.c +++ b/drivers/scsi/bnx2i/bnx2i_init.c | |||
@@ -15,11 +15,10 @@ | |||
15 | 15 | ||
16 | static struct list_head adapter_list = LIST_HEAD_INIT(adapter_list); | 16 | static struct list_head adapter_list = LIST_HEAD_INIT(adapter_list); |
17 | static u32 adapter_count; | 17 | static u32 adapter_count; |
18 | static int bnx2i_reg_device; | ||
19 | 18 | ||
20 | #define DRV_MODULE_NAME "bnx2i" | 19 | #define DRV_MODULE_NAME "bnx2i" |
21 | #define DRV_MODULE_VERSION "2.0.1d" | 20 | #define DRV_MODULE_VERSION "2.0.1e" |
22 | #define DRV_MODULE_RELDATE "Mar 25, 2009" | 21 | #define DRV_MODULE_RELDATE "June 22, 2009" |
23 | 22 | ||
24 | static char version[] __devinitdata = | 23 | static char version[] __devinitdata = |
25 | "Broadcom NetXtreme II iSCSI Driver " DRV_MODULE_NAME \ | 24 | "Broadcom NetXtreme II iSCSI Driver " DRV_MODULE_NAME \ |
@@ -31,7 +30,7 @@ MODULE_DESCRIPTION("Broadcom NetXtreme II BCM5706/5708/5709 iSCSI Driver"); | |||
31 | MODULE_LICENSE("GPL"); | 30 | MODULE_LICENSE("GPL"); |
32 | MODULE_VERSION(DRV_MODULE_VERSION); | 31 | MODULE_VERSION(DRV_MODULE_VERSION); |
33 | 32 | ||
34 | static DEFINE_RWLOCK(bnx2i_dev_lock); | 33 | static DEFINE_MUTEX(bnx2i_dev_lock); |
35 | 34 | ||
36 | unsigned int event_coal_div = 1; | 35 | unsigned int event_coal_div = 1; |
37 | module_param(event_coal_div, int, 0664); | 36 | module_param(event_coal_div, int, 0664); |
@@ -100,14 +99,14 @@ struct bnx2i_hba *get_adapter_list_head(void) | |||
100 | if (!adapter_count) | 99 | if (!adapter_count) |
101 | goto hba_not_found; | 100 | goto hba_not_found; |
102 | 101 | ||
103 | read_lock(&bnx2i_dev_lock); | 102 | mutex_lock(&bnx2i_dev_lock); |
104 | list_for_each_entry(tmp_hba, &adapter_list, link) { | 103 | list_for_each_entry(tmp_hba, &adapter_list, link) { |
105 | if (tmp_hba->cnic && tmp_hba->cnic->cm_select_dev) { | 104 | if (tmp_hba->cnic && tmp_hba->cnic->cm_select_dev) { |
106 | hba = tmp_hba; | 105 | hba = tmp_hba; |
107 | break; | 106 | break; |
108 | } | 107 | } |
109 | } | 108 | } |
110 | read_unlock(&bnx2i_dev_lock); | 109 | mutex_unlock(&bnx2i_dev_lock); |
111 | hba_not_found: | 110 | hba_not_found: |
112 | return hba; | 111 | return hba; |
113 | } | 112 | } |
@@ -122,14 +121,14 @@ struct bnx2i_hba *bnx2i_find_hba_for_cnic(struct cnic_dev *cnic) | |||
122 | { | 121 | { |
123 | struct bnx2i_hba *hba, *temp; | 122 | struct bnx2i_hba *hba, *temp; |
124 | 123 | ||
125 | read_lock(&bnx2i_dev_lock); | 124 | mutex_lock(&bnx2i_dev_lock); |
126 | list_for_each_entry_safe(hba, temp, &adapter_list, link) { | 125 | list_for_each_entry_safe(hba, temp, &adapter_list, link) { |
127 | if (hba->cnic == cnic) { | 126 | if (hba->cnic == cnic) { |
128 | read_unlock(&bnx2i_dev_lock); | 127 | mutex_unlock(&bnx2i_dev_lock); |
129 | return hba; | 128 | return hba; |
130 | } | 129 | } |
131 | } | 130 | } |
132 | read_unlock(&bnx2i_dev_lock); | 131 | mutex_unlock(&bnx2i_dev_lock); |
133 | return NULL; | 132 | return NULL; |
134 | } | 133 | } |
135 | 134 | ||
@@ -186,18 +185,17 @@ void bnx2i_stop(void *handle) | |||
186 | */ | 185 | */ |
187 | void bnx2i_register_device(struct bnx2i_hba *hba) | 186 | void bnx2i_register_device(struct bnx2i_hba *hba) |
188 | { | 187 | { |
188 | int rc; | ||
189 | |||
189 | if (test_bit(ADAPTER_STATE_GOING_DOWN, &hba->adapter_state) || | 190 | if (test_bit(ADAPTER_STATE_GOING_DOWN, &hba->adapter_state) || |
190 | test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic)) { | 191 | test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic)) { |
191 | return; | 192 | return; |
192 | } | 193 | } |
193 | 194 | ||
194 | hba->cnic->register_device(hba->cnic, CNIC_ULP_ISCSI, hba); | 195 | rc = hba->cnic->register_device(hba->cnic, CNIC_ULP_ISCSI, hba); |
195 | 196 | ||
196 | spin_lock(&hba->lock); | 197 | if (!rc) |
197 | bnx2i_reg_device++; | 198 | set_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic); |
198 | spin_unlock(&hba->lock); | ||
199 | |||
200 | set_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic); | ||
201 | } | 199 | } |
202 | 200 | ||
203 | 201 | ||
@@ -211,10 +209,10 @@ void bnx2i_reg_dev_all(void) | |||
211 | { | 209 | { |
212 | struct bnx2i_hba *hba, *temp; | 210 | struct bnx2i_hba *hba, *temp; |
213 | 211 | ||
214 | read_lock(&bnx2i_dev_lock); | 212 | mutex_lock(&bnx2i_dev_lock); |
215 | list_for_each_entry_safe(hba, temp, &adapter_list, link) | 213 | list_for_each_entry_safe(hba, temp, &adapter_list, link) |
216 | bnx2i_register_device(hba); | 214 | bnx2i_register_device(hba); |
217 | read_unlock(&bnx2i_dev_lock); | 215 | mutex_unlock(&bnx2i_dev_lock); |
218 | } | 216 | } |
219 | 217 | ||
220 | 218 | ||
@@ -234,10 +232,6 @@ static void bnx2i_unreg_one_device(struct bnx2i_hba *hba) | |||
234 | 232 | ||
235 | hba->cnic->unregister_device(hba->cnic, CNIC_ULP_ISCSI); | 233 | hba->cnic->unregister_device(hba->cnic, CNIC_ULP_ISCSI); |
236 | 234 | ||
237 | spin_lock(&hba->lock); | ||
238 | bnx2i_reg_device--; | ||
239 | spin_unlock(&hba->lock); | ||
240 | |||
241 | /* ep_disconnect could come before NETDEV_DOWN, driver won't | 235 | /* ep_disconnect could come before NETDEV_DOWN, driver won't |
242 | * see NETDEV_DOWN as it already unregistered itself. | 236 | * see NETDEV_DOWN as it already unregistered itself. |
243 | */ | 237 | */ |
@@ -255,10 +249,10 @@ void bnx2i_unreg_dev_all(void) | |||
255 | { | 249 | { |
256 | struct bnx2i_hba *hba, *temp; | 250 | struct bnx2i_hba *hba, *temp; |
257 | 251 | ||
258 | read_lock(&bnx2i_dev_lock); | 252 | mutex_lock(&bnx2i_dev_lock); |
259 | list_for_each_entry_safe(hba, temp, &adapter_list, link) | 253 | list_for_each_entry_safe(hba, temp, &adapter_list, link) |
260 | bnx2i_unreg_one_device(hba); | 254 | bnx2i_unreg_one_device(hba); |
261 | read_unlock(&bnx2i_dev_lock); | 255 | mutex_unlock(&bnx2i_dev_lock); |
262 | } | 256 | } |
263 | 257 | ||
264 | 258 | ||
@@ -267,35 +261,34 @@ void bnx2i_unreg_dev_all(void) | |||
267 | * @hba: bnx2i adapter instance | 261 | * @hba: bnx2i adapter instance |
268 | * @cnic: cnic device handle | 262 | * @cnic: cnic device handle |
269 | * | 263 | * |
270 | * Global resource lock and host adapter lock is held during critical sections | 264 | * Global resource lock is held during critical sections below. This routine is |
271 | * below. This routine is called from cnic_register_driver() context and | 265 | * called from either cnic_register_driver() or device hot plug context and |
272 | * work horse thread which does majority of device specific initialization | 266 | * and does majority of device specific initialization |
273 | */ | 267 | */ |
274 | static int bnx2i_init_one(struct bnx2i_hba *hba, struct cnic_dev *cnic) | 268 | static int bnx2i_init_one(struct bnx2i_hba *hba, struct cnic_dev *cnic) |
275 | { | 269 | { |
276 | int rc; | 270 | int rc; |
277 | 271 | ||
278 | read_lock(&bnx2i_dev_lock); | 272 | mutex_lock(&bnx2i_dev_lock); |
279 | if (bnx2i_reg_device && | 273 | rc = cnic->register_device(cnic, CNIC_ULP_ISCSI, hba); |
280 | !test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic)) { | 274 | if (!rc) { |
281 | rc = cnic->register_device(cnic, CNIC_ULP_ISCSI, hba); | ||
282 | if (rc) /* duplicate registration */ | ||
283 | printk(KERN_ERR "bnx2i- dev reg failed\n"); | ||
284 | |||
285 | spin_lock(&hba->lock); | ||
286 | bnx2i_reg_device++; | ||
287 | hba->age++; | 275 | hba->age++; |
288 | spin_unlock(&hba->lock); | ||
289 | |||
290 | set_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic); | 276 | set_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic); |
291 | } | 277 | list_add_tail(&hba->link, &adapter_list); |
292 | read_unlock(&bnx2i_dev_lock); | 278 | adapter_count++; |
293 | 279 | } else if (rc == -EBUSY) /* duplicate registration */ | |
294 | write_lock(&bnx2i_dev_lock); | 280 | printk(KERN_ALERT "bnx2i, duplicate registration" |
295 | list_add_tail(&hba->link, &adapter_list); | 281 | "hba=%p, cnic=%p\n", hba, cnic); |
296 | adapter_count++; | 282 | else if (rc == -EAGAIN) |
297 | write_unlock(&bnx2i_dev_lock); | 283 | printk(KERN_ERR "bnx2i, driver not registered\n"); |
298 | return 0; | 284 | else if (rc == -EINVAL) |
285 | printk(KERN_ERR "bnx2i, invalid type %d\n", CNIC_ULP_ISCSI); | ||
286 | else | ||
287 | printk(KERN_ERR "bnx2i dev reg, unknown error, %d\n", rc); | ||
288 | |||
289 | mutex_unlock(&bnx2i_dev_lock); | ||
290 | |||
291 | return rc; | ||
299 | } | 292 | } |
300 | 293 | ||
301 | 294 | ||
@@ -343,19 +336,15 @@ void bnx2i_ulp_exit(struct cnic_dev *dev) | |||
343 | "found, dev 0x%p\n", dev); | 336 | "found, dev 0x%p\n", dev); |
344 | return; | 337 | return; |
345 | } | 338 | } |
346 | write_lock(&bnx2i_dev_lock); | 339 | mutex_lock(&bnx2i_dev_lock); |
347 | list_del_init(&hba->link); | 340 | list_del_init(&hba->link); |
348 | adapter_count--; | 341 | adapter_count--; |
349 | 342 | ||
350 | if (test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic)) { | 343 | if (test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic)) { |
351 | hba->cnic->unregister_device(hba->cnic, CNIC_ULP_ISCSI); | 344 | hba->cnic->unregister_device(hba->cnic, CNIC_ULP_ISCSI); |
352 | clear_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic); | 345 | clear_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic); |
353 | |||
354 | spin_lock(&hba->lock); | ||
355 | bnx2i_reg_device--; | ||
356 | spin_unlock(&hba->lock); | ||
357 | } | 346 | } |
358 | write_unlock(&bnx2i_dev_lock); | 347 | mutex_unlock(&bnx2i_dev_lock); |
359 | 348 | ||
360 | bnx2i_free_hba(hba); | 349 | bnx2i_free_hba(hba); |
361 | } | 350 | } |
@@ -377,6 +366,8 @@ static int __init bnx2i_mod_init(void) | |||
377 | if (!is_power_of_2(sq_size)) | 366 | if (!is_power_of_2(sq_size)) |
378 | sq_size = roundup_pow_of_two(sq_size); | 367 | sq_size = roundup_pow_of_two(sq_size); |
379 | 368 | ||
369 | mutex_init(&bnx2i_dev_lock); | ||
370 | |||
380 | bnx2i_scsi_xport_template = | 371 | bnx2i_scsi_xport_template = |
381 | iscsi_register_transport(&bnx2i_iscsi_transport); | 372 | iscsi_register_transport(&bnx2i_iscsi_transport); |
382 | if (!bnx2i_scsi_xport_template) { | 373 | if (!bnx2i_scsi_xport_template) { |
@@ -412,7 +403,7 @@ static void __exit bnx2i_mod_exit(void) | |||
412 | { | 403 | { |
413 | struct bnx2i_hba *hba; | 404 | struct bnx2i_hba *hba; |
414 | 405 | ||
415 | write_lock(&bnx2i_dev_lock); | 406 | mutex_lock(&bnx2i_dev_lock); |
416 | while (!list_empty(&adapter_list)) { | 407 | while (!list_empty(&adapter_list)) { |
417 | hba = list_entry(adapter_list.next, struct bnx2i_hba, link); | 408 | hba = list_entry(adapter_list.next, struct bnx2i_hba, link); |
418 | list_del(&hba->link); | 409 | list_del(&hba->link); |
@@ -421,14 +412,11 @@ static void __exit bnx2i_mod_exit(void) | |||
421 | if (test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic)) { | 412 | if (test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic)) { |
422 | hba->cnic->unregister_device(hba->cnic, CNIC_ULP_ISCSI); | 413 | hba->cnic->unregister_device(hba->cnic, CNIC_ULP_ISCSI); |
423 | clear_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic); | 414 | clear_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic); |
424 | bnx2i_reg_device--; | ||
425 | } | 415 | } |
426 | 416 | ||
427 | write_unlock(&bnx2i_dev_lock); | ||
428 | bnx2i_free_hba(hba); | 417 | bnx2i_free_hba(hba); |
429 | write_lock(&bnx2i_dev_lock); | ||
430 | } | 418 | } |
431 | write_unlock(&bnx2i_dev_lock); | 419 | mutex_unlock(&bnx2i_dev_lock); |
432 | 420 | ||
433 | iscsi_unregister_transport(&bnx2i_iscsi_transport); | 421 | iscsi_unregister_transport(&bnx2i_iscsi_transport); |
434 | cnic_unregister_driver(CNIC_ULP_ISCSI); | 422 | cnic_unregister_driver(CNIC_ULP_ISCSI); |
diff --git a/drivers/scsi/bnx2i/bnx2i_iscsi.c b/drivers/scsi/bnx2i/bnx2i_iscsi.c index f7412196f2f8..9a7ba71f1af4 100644 --- a/drivers/scsi/bnx2i/bnx2i_iscsi.c +++ b/drivers/scsi/bnx2i/bnx2i_iscsi.c | |||
@@ -387,6 +387,7 @@ static struct iscsi_endpoint *bnx2i_alloc_ep(struct bnx2i_hba *hba) | |||
387 | bnx2i_ep = ep->dd_data; | 387 | bnx2i_ep = ep->dd_data; |
388 | INIT_LIST_HEAD(&bnx2i_ep->link); | 388 | INIT_LIST_HEAD(&bnx2i_ep->link); |
389 | bnx2i_ep->state = EP_STATE_IDLE; | 389 | bnx2i_ep->state = EP_STATE_IDLE; |
390 | bnx2i_ep->ep_iscsi_cid = (u16) -1; | ||
390 | bnx2i_ep->hba = hba; | 391 | bnx2i_ep->hba = hba; |
391 | bnx2i_ep->hba_age = hba->age; | 392 | bnx2i_ep->hba_age = hba->age; |
392 | hba->ofld_conns_active++; | 393 | hba->ofld_conns_active++; |
@@ -1160,9 +1161,6 @@ static int bnx2i_task_xmit(struct iscsi_task *task) | |||
1160 | struct bnx2i_cmd *cmd = task->dd_data; | 1161 | struct bnx2i_cmd *cmd = task->dd_data; |
1161 | struct iscsi_cmd *hdr = (struct iscsi_cmd *) task->hdr; | 1162 | struct iscsi_cmd *hdr = (struct iscsi_cmd *) task->hdr; |
1162 | 1163 | ||
1163 | if (test_bit(ADAPTER_STATE_LINK_DOWN, &hba->adapter_state)) | ||
1164 | return -ENOTCONN; | ||
1165 | |||
1166 | if (!bnx2i_conn->is_bound) | 1164 | if (!bnx2i_conn->is_bound) |
1167 | return -ENOTCONN; | 1165 | return -ENOTCONN; |
1168 | 1166 | ||
@@ -1653,15 +1651,18 @@ static struct iscsi_endpoint *bnx2i_ep_connect(struct Scsi_Host *shost, | |||
1653 | struct iscsi_endpoint *ep; | 1651 | struct iscsi_endpoint *ep; |
1654 | int rc = 0; | 1652 | int rc = 0; |
1655 | 1653 | ||
1656 | if (shost) | 1654 | if (shost) { |
1657 | /* driver is given scsi host to work with */ | 1655 | /* driver is given scsi host to work with */ |
1658 | hba = iscsi_host_priv(shost); | 1656 | hba = iscsi_host_priv(shost); |
1659 | else | 1657 | /* Register the device with cnic if not already done so */ |
1658 | bnx2i_register_device(hba); | ||
1659 | } else | ||
1660 | /* | 1660 | /* |
1661 | * check if the given destination can be reached through | 1661 | * check if the given destination can be reached through |
1662 | * a iscsi capable NetXtreme2 device | 1662 | * a iscsi capable NetXtreme2 device |
1663 | */ | 1663 | */ |
1664 | hba = bnx2i_check_route(dst_addr); | 1664 | hba = bnx2i_check_route(dst_addr); |
1665 | |||
1665 | if (!hba) { | 1666 | if (!hba) { |
1666 | rc = -ENOMEM; | 1667 | rc = -ENOMEM; |
1667 | goto check_busy; | 1668 | goto check_busy; |
@@ -1681,8 +1682,6 @@ static struct iscsi_endpoint *bnx2i_ep_connect(struct Scsi_Host *shost, | |||
1681 | goto net_if_down; | 1682 | goto net_if_down; |
1682 | } | 1683 | } |
1683 | 1684 | ||
1684 | bnx2i_ep->state = EP_STATE_IDLE; | ||
1685 | bnx2i_ep->ep_iscsi_cid = (u16) -1; | ||
1686 | bnx2i_ep->num_active_cmds = 0; | 1685 | bnx2i_ep->num_active_cmds = 0; |
1687 | iscsi_cid = bnx2i_alloc_iscsi_cid(hba); | 1686 | iscsi_cid = bnx2i_alloc_iscsi_cid(hba); |
1688 | if (iscsi_cid == -1) { | 1687 | if (iscsi_cid == -1) { |
diff --git a/drivers/scsi/ch.c b/drivers/scsi/ch.c index 7b1633a8c15a..fe11c1d4b31d 100644 --- a/drivers/scsi/ch.c +++ b/drivers/scsi/ch.c | |||
@@ -353,6 +353,12 @@ ch_readconfig(scsi_changer *ch) | |||
353 | /* look up the devices of the data transfer elements */ | 353 | /* look up the devices of the data transfer elements */ |
354 | ch->dt = kmalloc(ch->counts[CHET_DT]*sizeof(struct scsi_device), | 354 | ch->dt = kmalloc(ch->counts[CHET_DT]*sizeof(struct scsi_device), |
355 | GFP_KERNEL); | 355 | GFP_KERNEL); |
356 | |||
357 | if (!ch->dt) { | ||
358 | kfree(buffer); | ||
359 | return -ENOMEM; | ||
360 | } | ||
361 | |||
356 | for (elem = 0; elem < ch->counts[CHET_DT]; elem++) { | 362 | for (elem = 0; elem < ch->counts[CHET_DT]; elem++) { |
357 | id = -1; | 363 | id = -1; |
358 | lun = 0; | 364 | lun = 0; |
diff --git a/drivers/scsi/constants.c b/drivers/scsi/constants.c index e79e18101f87..63abb06c4edb 100644 --- a/drivers/scsi/constants.c +++ b/drivers/scsi/constants.c | |||
@@ -4,8 +4,7 @@ | |||
4 | * Additions for SCSI 2 and Linux 2.2.x by D. Gilbert (990422) | 4 | * Additions for SCSI 2 and Linux 2.2.x by D. Gilbert (990422) |
5 | * Additions for SCSI 3+ (SPC-3 T10/1416-D Rev 07 3 May 2002) | 5 | * Additions for SCSI 3+ (SPC-3 T10/1416-D Rev 07 3 May 2002) |
6 | * by D. Gilbert and aeb (20020609) | 6 | * by D. Gilbert and aeb (20020609) |
7 | * Additions for SPC-3 T10/1416-D Rev 21 22 Sept 2004, D. Gilbert 20041025 | 7 | * Update to SPC-4 T10/1713-D Rev 20, 22 May 2009, D. Gilbert 20090624 |
8 | * Update to SPC-4 T10/1713-D Rev 5a, 14 June 2006, D. Gilbert 20060702 | ||
9 | */ | 8 | */ |
10 | 9 | ||
11 | #include <linux/blkdev.h> | 10 | #include <linux/blkdev.h> |
@@ -56,9 +55,9 @@ static const char * cdb_byte0_names[] = { | |||
56 | "Read Buffer", | 55 | "Read Buffer", |
57 | /* 3d-3f */ "Update Block", "Read Long(10)", "Write Long(10)", | 56 | /* 3d-3f */ "Update Block", "Read Long(10)", "Write Long(10)", |
58 | /* 40-41 */ "Change Definition", "Write Same(10)", | 57 | /* 40-41 */ "Change Definition", "Write Same(10)", |
59 | /* 42-48 */ "Read sub-channel", "Read TOC/PMA/ATIP", "Read density support", | 58 | /* 42-48 */ "Unmap/Read sub-channel", "Read TOC/PMA/ATIP", |
60 | "Play audio(10)", "Get configuration", "Play audio msf", | 59 | "Read density support", "Play audio(10)", "Get configuration", |
61 | "Play audio track/index", | 60 | "Play audio msf", "Play audio track/index", |
62 | /* 49-4f */ "Play track relative(10)", "Get event status notification", | 61 | /* 49-4f */ "Play track relative(10)", "Get event status notification", |
63 | "Pause/resume", "Log Select", "Log Sense", "Stop play/scan", | 62 | "Pause/resume", "Log Select", "Log Sense", "Stop play/scan", |
64 | NULL, | 63 | NULL, |
@@ -71,12 +70,13 @@ static const char * cdb_byte0_names[] = { | |||
71 | /* 60-67 */ NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, | 70 | /* 60-67 */ NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, |
72 | /* 68-6f */ NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, | 71 | /* 68-6f */ NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, |
73 | /* 70-77 */ NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, | 72 | /* 70-77 */ NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, |
74 | /* 78-7f */ NULL, NULL, NULL, NULL, NULL, NULL, NULL, "Variable length", | 73 | /* 78-7f */ NULL, NULL, NULL, NULL, NULL, NULL, "Extended CDB", |
74 | "Variable length", | ||
75 | /* 80-84 */ "Xdwrite(16)", "Rebuild(16)", "Regenerate(16)", "Extended copy", | 75 | /* 80-84 */ "Xdwrite(16)", "Rebuild(16)", "Regenerate(16)", "Extended copy", |
76 | "Receive copy results", | 76 | "Receive copy results", |
77 | /* 85-89 */ "ATA command pass through(16)", "Access control in", | 77 | /* 85-89 */ "ATA command pass through(16)", "Access control in", |
78 | "Access control out", "Read(16)", "Memory Export Out(16)", | 78 | "Access control out", "Read(16)", "Memory Export Out(16)", |
79 | /* 8a-8f */ "Write(16)", NULL, "Read attributes", "Write attributes", | 79 | /* 8a-8f */ "Write(16)", "ORWrite", "Read attributes", "Write attributes", |
80 | "Write and verify(16)", "Verify(16)", | 80 | "Write and verify(16)", "Verify(16)", |
81 | /* 90-94 */ "Pre-fetch(16)", "Synchronize cache(16)", | 81 | /* 90-94 */ "Pre-fetch(16)", "Synchronize cache(16)", |
82 | "Lock/unlock cache(16)", "Write same(16)", NULL, | 82 | "Lock/unlock cache(16)", "Write same(16)", NULL, |
@@ -107,22 +107,24 @@ struct value_name_pair { | |||
107 | }; | 107 | }; |
108 | 108 | ||
109 | static const struct value_name_pair maint_in_arr[] = { | 109 | static const struct value_name_pair maint_in_arr[] = { |
110 | {0x5, "Report device identifier"}, | 110 | {0x5, "Report identifying information"}, |
111 | {0xa, "Report target port groups"}, | 111 | {0xa, "Report target port groups"}, |
112 | {0xb, "Report aliases"}, | 112 | {0xb, "Report aliases"}, |
113 | {0xc, "Report supported operation codes"}, | 113 | {0xc, "Report supported operation codes"}, |
114 | {0xd, "Report supported task management functions"}, | 114 | {0xd, "Report supported task management functions"}, |
115 | {0xe, "Report priority"}, | 115 | {0xe, "Report priority"}, |
116 | {0xf, "Report timestamp"}, | 116 | {0xf, "Report timestamp"}, |
117 | {0x10, "Management protocol in"}, | ||
117 | }; | 118 | }; |
118 | #define MAINT_IN_SZ ARRAY_SIZE(maint_in_arr) | 119 | #define MAINT_IN_SZ ARRAY_SIZE(maint_in_arr) |
119 | 120 | ||
120 | static const struct value_name_pair maint_out_arr[] = { | 121 | static const struct value_name_pair maint_out_arr[] = { |
121 | {0x6, "Set device identifier"}, | 122 | {0x6, "Set identifying information"}, |
122 | {0xa, "Set target port groups"}, | 123 | {0xa, "Set target port groups"}, |
123 | {0xb, "Change aliases"}, | 124 | {0xb, "Change aliases"}, |
124 | {0xe, "Set priority"}, | 125 | {0xe, "Set priority"}, |
125 | {0xe, "Set timestamp"}, | 126 | {0xf, "Set timestamp"}, |
127 | {0x10, "Management protocol out"}, | ||
126 | }; | 128 | }; |
127 | #define MAINT_OUT_SZ ARRAY_SIZE(maint_out_arr) | 129 | #define MAINT_OUT_SZ ARRAY_SIZE(maint_out_arr) |
128 | 130 | ||
@@ -412,6 +414,7 @@ static const struct error_info additional[] = | |||
412 | {0x0004, "Beginning-of-partition/medium detected"}, | 414 | {0x0004, "Beginning-of-partition/medium detected"}, |
413 | {0x0005, "End-of-data detected"}, | 415 | {0x0005, "End-of-data detected"}, |
414 | {0x0006, "I/O process terminated"}, | 416 | {0x0006, "I/O process terminated"}, |
417 | {0x0007, "Programmable early warning detected"}, | ||
415 | {0x0011, "Audio play operation in progress"}, | 418 | {0x0011, "Audio play operation in progress"}, |
416 | {0x0012, "Audio play operation paused"}, | 419 | {0x0012, "Audio play operation paused"}, |
417 | {0x0013, "Audio play operation successfully completed"}, | 420 | {0x0013, "Audio play operation successfully completed"}, |
@@ -425,6 +428,7 @@ static const struct error_info additional[] = | |||
425 | {0x001B, "Set capacity operation in progress"}, | 428 | {0x001B, "Set capacity operation in progress"}, |
426 | {0x001C, "Verify operation in progress"}, | 429 | {0x001C, "Verify operation in progress"}, |
427 | {0x001D, "ATA pass through information available"}, | 430 | {0x001D, "ATA pass through information available"}, |
431 | {0x001E, "Conflicting SA creation request"}, | ||
428 | 432 | ||
429 | {0x0100, "No index/sector signal"}, | 433 | {0x0100, "No index/sector signal"}, |
430 | 434 | ||
@@ -449,9 +453,12 @@ static const struct error_info additional[] = | |||
449 | {0x040B, "Logical unit not accessible, target port in standby state"}, | 453 | {0x040B, "Logical unit not accessible, target port in standby state"}, |
450 | {0x040C, "Logical unit not accessible, target port in unavailable " | 454 | {0x040C, "Logical unit not accessible, target port in unavailable " |
451 | "state"}, | 455 | "state"}, |
456 | {0x040D, "Logical unit not ready, structure check required"}, | ||
452 | {0x0410, "Logical unit not ready, auxiliary memory not accessible"}, | 457 | {0x0410, "Logical unit not ready, auxiliary memory not accessible"}, |
453 | {0x0411, "Logical unit not ready, notify (enable spinup) required"}, | 458 | {0x0411, "Logical unit not ready, notify (enable spinup) required"}, |
454 | {0x0412, "Logical unit not ready, offline"}, | 459 | {0x0412, "Logical unit not ready, offline"}, |
460 | {0x0413, "Logical unit not ready, SA creation in progress"}, | ||
461 | {0x0414, "Logical unit not ready, space allocation in progress"}, | ||
455 | 462 | ||
456 | {0x0500, "Logical unit does not respond to selection"}, | 463 | {0x0500, "Logical unit does not respond to selection"}, |
457 | 464 | ||
@@ -479,6 +486,9 @@ static const struct error_info additional[] = | |||
479 | {0x0B03, "Warning - background self-test failed"}, | 486 | {0x0B03, "Warning - background self-test failed"}, |
480 | {0x0B04, "Warning - background pre-scan detected medium error"}, | 487 | {0x0B04, "Warning - background pre-scan detected medium error"}, |
481 | {0x0B05, "Warning - background medium scan detected medium error"}, | 488 | {0x0B05, "Warning - background medium scan detected medium error"}, |
489 | {0x0B06, "Warning - non-volatile cache now volatile"}, | ||
490 | {0x0B07, "Warning - degraded power to non-volatile cache"}, | ||
491 | {0x0B08, "Warning - power loss expected"}, | ||
482 | 492 | ||
483 | {0x0C00, "Write error"}, | 493 | {0x0C00, "Write error"}, |
484 | {0x0C01, "Write error - recovered with auto reallocation"}, | 494 | {0x0C01, "Write error - recovered with auto reallocation"}, |
@@ -593,6 +603,7 @@ static const struct error_info additional[] = | |||
593 | {0x1C02, "Grown defect list not found"}, | 603 | {0x1C02, "Grown defect list not found"}, |
594 | 604 | ||
595 | {0x1D00, "Miscompare during verify operation"}, | 605 | {0x1D00, "Miscompare during verify operation"}, |
606 | {0x1D01, "Miscompare verify of unmapped LBA"}, | ||
596 | 607 | ||
597 | {0x1E00, "Recovered id with ECC correction"}, | 608 | {0x1E00, "Recovered id with ECC correction"}, |
598 | 609 | ||
@@ -626,6 +637,7 @@ static const struct error_info additional[] = | |||
626 | {0x2405, "Security working key frozen"}, | 637 | {0x2405, "Security working key frozen"}, |
627 | {0x2406, "Nonce not unique"}, | 638 | {0x2406, "Nonce not unique"}, |
628 | {0x2407, "Nonce timestamp out of range"}, | 639 | {0x2407, "Nonce timestamp out of range"}, |
640 | {0x2408, "Invalid XCDB"}, | ||
629 | 641 | ||
630 | {0x2500, "Logical unit not supported"}, | 642 | {0x2500, "Logical unit not supported"}, |
631 | 643 | ||
@@ -656,10 +668,12 @@ static const struct error_info additional[] = | |||
656 | {0x2704, "Persistent write protect"}, | 668 | {0x2704, "Persistent write protect"}, |
657 | {0x2705, "Permanent write protect"}, | 669 | {0x2705, "Permanent write protect"}, |
658 | {0x2706, "Conditional write protect"}, | 670 | {0x2706, "Conditional write protect"}, |
671 | {0x2707, "Space allocation failed write protect"}, | ||
659 | 672 | ||
660 | {0x2800, "Not ready to ready change, medium may have changed"}, | 673 | {0x2800, "Not ready to ready change, medium may have changed"}, |
661 | {0x2801, "Import or export element accessed"}, | 674 | {0x2801, "Import or export element accessed"}, |
662 | {0x2802, "Format-layer may have changed"}, | 675 | {0x2802, "Format-layer may have changed"}, |
676 | {0x2803, "Import/export element accessed, medium changed"}, | ||
663 | 677 | ||
664 | {0x2900, "Power on, reset, or bus device reset occurred"}, | 678 | {0x2900, "Power on, reset, or bus device reset occurred"}, |
665 | {0x2901, "Power on occurred"}, | 679 | {0x2901, "Power on occurred"}, |
@@ -680,11 +694,16 @@ static const struct error_info additional[] = | |||
680 | {0x2A07, "Implicit asymmetric access state transition failed"}, | 694 | {0x2A07, "Implicit asymmetric access state transition failed"}, |
681 | {0x2A08, "Priority changed"}, | 695 | {0x2A08, "Priority changed"}, |
682 | {0x2A09, "Capacity data has changed"}, | 696 | {0x2A09, "Capacity data has changed"}, |
697 | {0x2A0A, "Error history I_T nexus cleared"}, | ||
698 | {0x2A0B, "Error history snapshot released"}, | ||
699 | {0x2A0C, "Error recovery attributes have changed"}, | ||
700 | {0x2A0D, "Data encryption capabilities changed"}, | ||
683 | {0x2A10, "Timestamp changed"}, | 701 | {0x2A10, "Timestamp changed"}, |
684 | {0x2A11, "Data encryption parameters changed by another i_t nexus"}, | 702 | {0x2A11, "Data encryption parameters changed by another i_t nexus"}, |
685 | {0x2A12, "Data encryption parameters changed by vendor specific " | 703 | {0x2A12, "Data encryption parameters changed by vendor specific " |
686 | "event"}, | 704 | "event"}, |
687 | {0x2A13, "Data encryption key instance counter has changed"}, | 705 | {0x2A13, "Data encryption key instance counter has changed"}, |
706 | {0x2A14, "SA creation capabilities data has changed"}, | ||
688 | 707 | ||
689 | {0x2B00, "Copy cannot execute since host cannot disconnect"}, | 708 | {0x2B00, "Copy cannot execute since host cannot disconnect"}, |
690 | 709 | ||
@@ -723,6 +742,8 @@ static const struct error_info additional[] = | |||
723 | {0x300C, "WORM medium - overwrite attempted"}, | 742 | {0x300C, "WORM medium - overwrite attempted"}, |
724 | {0x300D, "WORM medium - integrity check"}, | 743 | {0x300D, "WORM medium - integrity check"}, |
725 | {0x3010, "Medium not formatted"}, | 744 | {0x3010, "Medium not formatted"}, |
745 | {0x3011, "Incompatible volume type"}, | ||
746 | {0x3012, "Incompatible volume qualifier"}, | ||
726 | 747 | ||
727 | {0x3100, "Medium format corrupted"}, | 748 | {0x3100, "Medium format corrupted"}, |
728 | {0x3101, "Format command failed"}, | 749 | {0x3101, "Format command failed"}, |
@@ -782,6 +803,10 @@ static const struct error_info additional[] = | |||
782 | {0x3B15, "Medium magazine unlocked"}, | 803 | {0x3B15, "Medium magazine unlocked"}, |
783 | {0x3B16, "Mechanical positioning or changer error"}, | 804 | {0x3B16, "Mechanical positioning or changer error"}, |
784 | {0x3B17, "Read past end of user object"}, | 805 | {0x3B17, "Read past end of user object"}, |
806 | {0x3B18, "Element disabled"}, | ||
807 | {0x3B19, "Element enabled"}, | ||
808 | {0x3B1A, "Data transfer device removed"}, | ||
809 | {0x3B1B, "Data transfer device inserted"}, | ||
785 | 810 | ||
786 | {0x3D00, "Invalid bits in identify message"}, | 811 | {0x3D00, "Invalid bits in identify message"}, |
787 | 812 | ||
@@ -882,6 +907,8 @@ static const struct error_info additional[] = | |||
882 | {0x5506, "Auxiliary memory out of space"}, | 907 | {0x5506, "Auxiliary memory out of space"}, |
883 | {0x5507, "Quota error"}, | 908 | {0x5507, "Quota error"}, |
884 | {0x5508, "Maximum number of supplemental decryption keys exceeded"}, | 909 | {0x5508, "Maximum number of supplemental decryption keys exceeded"}, |
910 | {0x5509, "Medium auxiliary memory not accessible"}, | ||
911 | {0x550A, "Data currently unavailable"}, | ||
885 | 912 | ||
886 | {0x5700, "Unable to recover table-of-contents"}, | 913 | {0x5700, "Unable to recover table-of-contents"}, |
887 | 914 | ||
@@ -993,6 +1020,12 @@ static const struct error_info additional[] = | |||
993 | {0x5E02, "Standby condition activated by timer"}, | 1020 | {0x5E02, "Standby condition activated by timer"}, |
994 | {0x5E03, "Idle condition activated by command"}, | 1021 | {0x5E03, "Idle condition activated by command"}, |
995 | {0x5E04, "Standby condition activated by command"}, | 1022 | {0x5E04, "Standby condition activated by command"}, |
1023 | {0x5E05, "Idle_b condition activated by timer"}, | ||
1024 | {0x5E06, "Idle_b condition activated by command"}, | ||
1025 | {0x5E07, "Idle_c condition activated by timer"}, | ||
1026 | {0x5E08, "Idle_c condition activated by command"}, | ||
1027 | {0x5E09, "Standby_y condition activated by timer"}, | ||
1028 | {0x5E0A, "Standby_y condition activated by command"}, | ||
996 | {0x5E41, "Power state change to active"}, | 1029 | {0x5E41, "Power state change to active"}, |
997 | {0x5E42, "Power state change to idle"}, | 1030 | {0x5E42, "Power state change to idle"}, |
998 | {0x5E43, "Power state change to standby"}, | 1031 | {0x5E43, "Power state change to standby"}, |
@@ -1091,7 +1124,28 @@ static const struct error_info additional[] = | |||
1091 | {0x7403, "Incorrect data encryption key"}, | 1124 | {0x7403, "Incorrect data encryption key"}, |
1092 | {0x7404, "Cryptographic integrity validation failed"}, | 1125 | {0x7404, "Cryptographic integrity validation failed"}, |
1093 | {0x7405, "Error decrypting data"}, | 1126 | {0x7405, "Error decrypting data"}, |
1127 | {0x7406, "Unknown signature verification key"}, | ||
1128 | {0x7407, "Encryption parameters not useable"}, | ||
1129 | {0x7408, "Digital signature validation failure"}, | ||
1130 | {0x7409, "Encryption mode mismatch on read"}, | ||
1131 | {0x740A, "Encrypted block not raw read enabled"}, | ||
1132 | {0x740B, "Incorrect Encryption parameters"}, | ||
1133 | {0x740C, "Unable to decrypt parameter list"}, | ||
1134 | {0x740D, "Encryption algorithm disabled"}, | ||
1135 | {0x7410, "SA creation parameter value invalid"}, | ||
1136 | {0x7411, "SA creation parameter value rejected"}, | ||
1137 | {0x7412, "Invalid SA usage"}, | ||
1138 | {0x7421, "Data Encryption configuration prevented"}, | ||
1139 | {0x7430, "SA creation parameter not supported"}, | ||
1140 | {0x7440, "Authentication failed"}, | ||
1141 | {0x7461, "External data encryption key manager access error"}, | ||
1142 | {0x7462, "External data encryption key manager error"}, | ||
1143 | {0x7463, "External data encryption key not found"}, | ||
1144 | {0x7464, "External data encryption request not authorized"}, | ||
1145 | {0x746E, "External data encryption control timeout"}, | ||
1146 | {0x746F, "External data encryption control error"}, | ||
1094 | {0x7471, "Logical unit access not authorized"}, | 1147 | {0x7471, "Logical unit access not authorized"}, |
1148 | {0x7479, "Security conflict in translated device"}, | ||
1095 | 1149 | ||
1096 | {0, NULL} | 1150 | {0, NULL} |
1097 | }; | 1151 | }; |
@@ -1103,12 +1157,12 @@ struct error_info2 { | |||
1103 | 1157 | ||
1104 | static const struct error_info2 additional2[] = | 1158 | static const struct error_info2 additional2[] = |
1105 | { | 1159 | { |
1106 | {0x40,0x00,0x7f,"Ram failure (%x)"}, | 1160 | {0x40, 0x00, 0x7f, "Ram failure (%x)"}, |
1107 | {0x40,0x80,0xff,"Diagnostic failure on component (%x)"}, | 1161 | {0x40, 0x80, 0xff, "Diagnostic failure on component (%x)"}, |
1108 | {0x41,0x00,0xff,"Data path failure (%x)"}, | 1162 | {0x41, 0x00, 0xff, "Data path failure (%x)"}, |
1109 | {0x42,0x00,0xff,"Power-on or self-test failure (%x)"}, | 1163 | {0x42, 0x00, 0xff, "Power-on or self-test failure (%x)"}, |
1110 | {0x4D,0x00,0xff,"Tagged overlapped commands (queue tag %x)"}, | 1164 | {0x4D, 0x00, 0xff, "Tagged overlapped commands (task tag %x)"}, |
1111 | {0x70,0x00,0xff,"Decompression exception short algorithm id of %x"}, | 1165 | {0x70, 0x00, 0xff, "Decompression exception short algorithm id of %x"}, |
1112 | {0, 0, 0, NULL} | 1166 | {0, 0, 0, NULL} |
1113 | }; | 1167 | }; |
1114 | 1168 | ||
@@ -1157,14 +1211,15 @@ scsi_extd_sense_format(unsigned char asc, unsigned char ascq) { | |||
1157 | int i; | 1211 | int i; |
1158 | unsigned short code = ((asc << 8) | ascq); | 1212 | unsigned short code = ((asc << 8) | ascq); |
1159 | 1213 | ||
1160 | for (i=0; additional[i].text; i++) | 1214 | for (i = 0; additional[i].text; i++) |
1161 | if (additional[i].code12 == code) | 1215 | if (additional[i].code12 == code) |
1162 | return additional[i].text; | 1216 | return additional[i].text; |
1163 | for (i=0; additional2[i].fmt; i++) | 1217 | for (i = 0; additional2[i].fmt; i++) { |
1164 | if (additional2[i].code1 == asc && | 1218 | if (additional2[i].code1 == asc && |
1165 | additional2[i].code2_min >= ascq && | 1219 | ascq >= additional2[i].code2_min && |
1166 | additional2[i].code2_max <= ascq) | 1220 | ascq <= additional2[i].code2_max) |
1167 | return additional2[i].fmt; | 1221 | return additional2[i].fmt; |
1222 | } | ||
1168 | #endif | 1223 | #endif |
1169 | return NULL; | 1224 | return NULL; |
1170 | } | 1225 | } |
diff --git a/drivers/scsi/device_handler/scsi_dh.c b/drivers/scsi/device_handler/scsi_dh.c index a518f2eff19a..3ee1cbc89479 100644 --- a/drivers/scsi/device_handler/scsi_dh.c +++ b/drivers/scsi/device_handler/scsi_dh.c | |||
@@ -153,12 +153,24 @@ static int scsi_dh_handler_attach(struct scsi_device *sdev, | |||
153 | if (sdev->scsi_dh_data) { | 153 | if (sdev->scsi_dh_data) { |
154 | if (sdev->scsi_dh_data->scsi_dh != scsi_dh) | 154 | if (sdev->scsi_dh_data->scsi_dh != scsi_dh) |
155 | err = -EBUSY; | 155 | err = -EBUSY; |
156 | } else if (scsi_dh->attach) | 156 | else |
157 | kref_get(&sdev->scsi_dh_data->kref); | ||
158 | } else if (scsi_dh->attach) { | ||
157 | err = scsi_dh->attach(sdev); | 159 | err = scsi_dh->attach(sdev); |
158 | 160 | if (!err) { | |
161 | kref_init(&sdev->scsi_dh_data->kref); | ||
162 | sdev->scsi_dh_data->sdev = sdev; | ||
163 | } | ||
164 | } | ||
159 | return err; | 165 | return err; |
160 | } | 166 | } |
161 | 167 | ||
168 | static void __detach_handler (struct kref *kref) | ||
169 | { | ||
170 | struct scsi_dh_data *scsi_dh_data = container_of(kref, struct scsi_dh_data, kref); | ||
171 | scsi_dh_data->scsi_dh->detach(scsi_dh_data->sdev); | ||
172 | } | ||
173 | |||
162 | /* | 174 | /* |
163 | * scsi_dh_handler_detach - Detach a device handler from a device | 175 | * scsi_dh_handler_detach - Detach a device handler from a device |
164 | * @sdev - SCSI device the device handler should be detached from | 176 | * @sdev - SCSI device the device handler should be detached from |
@@ -180,7 +192,7 @@ static void scsi_dh_handler_detach(struct scsi_device *sdev, | |||
180 | scsi_dh = sdev->scsi_dh_data->scsi_dh; | 192 | scsi_dh = sdev->scsi_dh_data->scsi_dh; |
181 | 193 | ||
182 | if (scsi_dh && scsi_dh->detach) | 194 | if (scsi_dh && scsi_dh->detach) |
183 | scsi_dh->detach(sdev); | 195 | kref_put(&sdev->scsi_dh_data->kref, __detach_handler); |
184 | } | 196 | } |
185 | 197 | ||
186 | /* | 198 | /* |
@@ -440,6 +452,39 @@ int scsi_dh_activate(struct request_queue *q) | |||
440 | EXPORT_SYMBOL_GPL(scsi_dh_activate); | 452 | EXPORT_SYMBOL_GPL(scsi_dh_activate); |
441 | 453 | ||
442 | /* | 454 | /* |
455 | * scsi_dh_set_params - set the parameters for the device as per the | ||
456 | * string specified in params. | ||
457 | * @q - Request queue that is associated with the scsi_device for | ||
458 | * which the parameters to be set. | ||
459 | * @params - parameters in the following format | ||
460 | * "no_of_params\0param1\0param2\0param3\0...\0" | ||
461 | * for example, string for 2 parameters with value 10 and 21 | ||
462 | * is specified as "2\010\021\0". | ||
463 | */ | ||
464 | int scsi_dh_set_params(struct request_queue *q, const char *params) | ||
465 | { | ||
466 | int err = -SCSI_DH_NOSYS; | ||
467 | unsigned long flags; | ||
468 | struct scsi_device *sdev; | ||
469 | struct scsi_device_handler *scsi_dh = NULL; | ||
470 | |||
471 | spin_lock_irqsave(q->queue_lock, flags); | ||
472 | sdev = q->queuedata; | ||
473 | if (sdev && sdev->scsi_dh_data) | ||
474 | scsi_dh = sdev->scsi_dh_data->scsi_dh; | ||
475 | if (scsi_dh && scsi_dh->set_params && get_device(&sdev->sdev_gendev)) | ||
476 | err = 0; | ||
477 | spin_unlock_irqrestore(q->queue_lock, flags); | ||
478 | |||
479 | if (err) | ||
480 | return err; | ||
481 | err = scsi_dh->set_params(sdev, params); | ||
482 | put_device(&sdev->sdev_gendev); | ||
483 | return err; | ||
484 | } | ||
485 | EXPORT_SYMBOL_GPL(scsi_dh_set_params); | ||
486 | |||
487 | /* | ||
443 | * scsi_dh_handler_exist - Return TRUE(1) if a device handler exists for | 488 | * scsi_dh_handler_exist - Return TRUE(1) if a device handler exists for |
444 | * the given name. FALSE(0) otherwise. | 489 | * the given name. FALSE(0) otherwise. |
445 | * @name - name of the device handler. | 490 | * @name - name of the device handler. |
@@ -474,7 +519,6 @@ int scsi_dh_attach(struct request_queue *q, const char *name) | |||
474 | 519 | ||
475 | if (!err) { | 520 | if (!err) { |
476 | err = scsi_dh_handler_attach(sdev, scsi_dh); | 521 | err = scsi_dh_handler_attach(sdev, scsi_dh); |
477 | |||
478 | put_device(&sdev->sdev_gendev); | 522 | put_device(&sdev->sdev_gendev); |
479 | } | 523 | } |
480 | return err; | 524 | return err; |
@@ -505,10 +549,8 @@ void scsi_dh_detach(struct request_queue *q) | |||
505 | return; | 549 | return; |
506 | 550 | ||
507 | if (sdev->scsi_dh_data) { | 551 | if (sdev->scsi_dh_data) { |
508 | /* if sdev is not on internal list, detach */ | ||
509 | scsi_dh = sdev->scsi_dh_data->scsi_dh; | 552 | scsi_dh = sdev->scsi_dh_data->scsi_dh; |
510 | if (!device_handler_match(scsi_dh, sdev)) | 553 | scsi_dh_handler_detach(sdev, scsi_dh); |
511 | scsi_dh_handler_detach(sdev, scsi_dh); | ||
512 | } | 554 | } |
513 | put_device(&sdev->sdev_gendev); | 555 | put_device(&sdev->sdev_gendev); |
514 | } | 556 | } |
diff --git a/drivers/scsi/device_handler/scsi_dh_alua.c b/drivers/scsi/device_handler/scsi_dh_alua.c index dba154c8ff64..b5cdefaf2608 100644 --- a/drivers/scsi/device_handler/scsi_dh_alua.c +++ b/drivers/scsi/device_handler/scsi_dh_alua.c | |||
@@ -663,7 +663,7 @@ static int alua_activate(struct scsi_device *sdev) | |||
663 | goto out; | 663 | goto out; |
664 | } | 664 | } |
665 | 665 | ||
666 | if (h->tpgs == TPGS_MODE_EXPLICIT && h->state != TPGS_STATE_OPTIMIZED) | 666 | if (h->tpgs & TPGS_MODE_EXPLICIT && h->state != TPGS_STATE_OPTIMIZED) |
667 | err = alua_stpg(sdev, TPGS_STATE_OPTIMIZED, h); | 667 | err = alua_stpg(sdev, TPGS_STATE_OPTIMIZED, h); |
668 | 668 | ||
669 | out: | 669 | out: |
diff --git a/drivers/scsi/device_handler/scsi_dh_emc.c b/drivers/scsi/device_handler/scsi_dh_emc.c index 0e572d2c5b0a..0cffe84976fe 100644 --- a/drivers/scsi/device_handler/scsi_dh_emc.c +++ b/drivers/scsi/device_handler/scsi_dh_emc.c | |||
@@ -561,6 +561,61 @@ done: | |||
561 | 561 | ||
562 | return result; | 562 | return result; |
563 | } | 563 | } |
564 | /* | ||
565 | * params - parameters in the following format | ||
566 | * "no_of_params\0param1\0param2\0param3\0...\0" | ||
567 | * for example, string for 2 parameters with value 10 and 21 | ||
568 | * is specified as "2\010\021\0". | ||
569 | */ | ||
570 | static int clariion_set_params(struct scsi_device *sdev, const char *params) | ||
571 | { | ||
572 | struct clariion_dh_data *csdev = get_clariion_data(sdev); | ||
573 | unsigned int hr = 0, st = 0, argc; | ||
574 | const char *p = params; | ||
575 | int result = SCSI_DH_OK; | ||
576 | |||
577 | if ((sscanf(params, "%u", &argc) != 1) || (argc != 2)) | ||
578 | return -EINVAL; | ||
579 | |||
580 | while (*p++) | ||
581 | ; | ||
582 | if ((sscanf(p, "%u", &st) != 1) || (st > 1)) | ||
583 | return -EINVAL; | ||
584 | |||
585 | while (*p++) | ||
586 | ; | ||
587 | if ((sscanf(p, "%u", &hr) != 1) || (hr > 1)) | ||
588 | return -EINVAL; | ||
589 | |||
590 | if (st) | ||
591 | csdev->flags |= CLARIION_SHORT_TRESPASS; | ||
592 | else | ||
593 | csdev->flags &= ~CLARIION_SHORT_TRESPASS; | ||
594 | |||
595 | if (hr) | ||
596 | csdev->flags |= CLARIION_HONOR_RESERVATIONS; | ||
597 | else | ||
598 | csdev->flags &= ~CLARIION_HONOR_RESERVATIONS; | ||
599 | |||
600 | /* | ||
601 | * If this path is owned, we have to send a trespass command | ||
602 | * with the new parameters. If not, simply return. Next trespass | ||
603 | * command would use the parameters. | ||
604 | */ | ||
605 | if (csdev->lun_state != CLARIION_LUN_OWNED) | ||
606 | goto done; | ||
607 | |||
608 | csdev->lun_state = CLARIION_LUN_UNINITIALIZED; | ||
609 | result = send_trespass_cmd(sdev, csdev); | ||
610 | if (result != SCSI_DH_OK) | ||
611 | goto done; | ||
612 | |||
613 | /* Update status */ | ||
614 | result = clariion_send_inquiry(sdev, csdev); | ||
615 | |||
616 | done: | ||
617 | return result; | ||
618 | } | ||
564 | 619 | ||
565 | static const struct scsi_dh_devlist clariion_dev_list[] = { | 620 | static const struct scsi_dh_devlist clariion_dev_list[] = { |
566 | {"DGC", "RAID"}, | 621 | {"DGC", "RAID"}, |
@@ -581,11 +636,9 @@ static struct scsi_device_handler clariion_dh = { | |||
581 | .check_sense = clariion_check_sense, | 636 | .check_sense = clariion_check_sense, |
582 | .activate = clariion_activate, | 637 | .activate = clariion_activate, |
583 | .prep_fn = clariion_prep_fn, | 638 | .prep_fn = clariion_prep_fn, |
639 | .set_params = clariion_set_params, | ||
584 | }; | 640 | }; |
585 | 641 | ||
586 | /* | ||
587 | * TODO: need some interface so we can set trespass values | ||
588 | */ | ||
589 | static int clariion_bus_attach(struct scsi_device *sdev) | 642 | static int clariion_bus_attach(struct scsi_device *sdev) |
590 | { | 643 | { |
591 | struct scsi_dh_data *scsi_dh_data; | 644 | struct scsi_dh_data *scsi_dh_data; |
diff --git a/drivers/scsi/device_handler/scsi_dh_rdac.c b/drivers/scsi/device_handler/scsi_dh_rdac.c index fd0544f7da81..11c89311427e 100644 --- a/drivers/scsi/device_handler/scsi_dh_rdac.c +++ b/drivers/scsi/device_handler/scsi_dh_rdac.c | |||
@@ -112,6 +112,7 @@ struct c9_inquiry { | |||
112 | 112 | ||
113 | #define SUBSYS_ID_LEN 16 | 113 | #define SUBSYS_ID_LEN 16 |
114 | #define SLOT_ID_LEN 2 | 114 | #define SLOT_ID_LEN 2 |
115 | #define ARRAY_LABEL_LEN 31 | ||
115 | 116 | ||
116 | struct c4_inquiry { | 117 | struct c4_inquiry { |
117 | u8 peripheral_info; | 118 | u8 peripheral_info; |
@@ -135,6 +136,8 @@ struct rdac_controller { | |||
135 | struct rdac_pg_legacy legacy; | 136 | struct rdac_pg_legacy legacy; |
136 | struct rdac_pg_expanded expanded; | 137 | struct rdac_pg_expanded expanded; |
137 | } mode_select; | 138 | } mode_select; |
139 | u8 index; | ||
140 | u8 array_name[ARRAY_LABEL_LEN]; | ||
138 | }; | 141 | }; |
139 | struct c8_inquiry { | 142 | struct c8_inquiry { |
140 | u8 peripheral_info; | 143 | u8 peripheral_info; |
@@ -198,6 +201,31 @@ static const char *lun_state[] = | |||
198 | static LIST_HEAD(ctlr_list); | 201 | static LIST_HEAD(ctlr_list); |
199 | static DEFINE_SPINLOCK(list_lock); | 202 | static DEFINE_SPINLOCK(list_lock); |
200 | 203 | ||
204 | /* | ||
205 | * module parameter to enable rdac debug logging. | ||
206 | * 2 bits for each type of logging, only two types defined for now | ||
207 | * Can be enhanced if required at later point | ||
208 | */ | ||
209 | static int rdac_logging = 1; | ||
210 | module_param(rdac_logging, int, S_IRUGO|S_IWUSR); | ||
211 | MODULE_PARM_DESC(rdac_logging, "A bit mask of rdac logging levels, " | ||
212 | "Default is 1 - failover logging enabled, " | ||
213 | "set it to 0xF to enable all the logs"); | ||
214 | |||
215 | #define RDAC_LOG_FAILOVER 0 | ||
216 | #define RDAC_LOG_SENSE 2 | ||
217 | |||
218 | #define RDAC_LOG_BITS 2 | ||
219 | |||
220 | #define RDAC_LOG_LEVEL(SHIFT) \ | ||
221 | ((rdac_logging >> (SHIFT)) & ((1 << (RDAC_LOG_BITS)) - 1)) | ||
222 | |||
223 | #define RDAC_LOG(SHIFT, sdev, f, arg...) \ | ||
224 | do { \ | ||
225 | if (unlikely(RDAC_LOG_LEVEL(SHIFT))) \ | ||
226 | sdev_printk(KERN_INFO, sdev, RDAC_NAME ": " f "\n", ## arg); \ | ||
227 | } while (0); | ||
228 | |||
201 | static inline struct rdac_dh_data *get_rdac_data(struct scsi_device *sdev) | 229 | static inline struct rdac_dh_data *get_rdac_data(struct scsi_device *sdev) |
202 | { | 230 | { |
203 | struct scsi_dh_data *scsi_dh_data = sdev->scsi_dh_data; | 231 | struct scsi_dh_data *scsi_dh_data = sdev->scsi_dh_data; |
@@ -303,7 +331,8 @@ static void release_controller(struct kref *kref) | |||
303 | kfree(ctlr); | 331 | kfree(ctlr); |
304 | } | 332 | } |
305 | 333 | ||
306 | static struct rdac_controller *get_controller(u8 *subsys_id, u8 *slot_id) | 334 | static struct rdac_controller *get_controller(u8 *subsys_id, u8 *slot_id, |
335 | char *array_name) | ||
307 | { | 336 | { |
308 | struct rdac_controller *ctlr, *tmp; | 337 | struct rdac_controller *ctlr, *tmp; |
309 | 338 | ||
@@ -324,6 +353,14 @@ static struct rdac_controller *get_controller(u8 *subsys_id, u8 *slot_id) | |||
324 | /* initialize fields of controller */ | 353 | /* initialize fields of controller */ |
325 | memcpy(ctlr->subsys_id, subsys_id, SUBSYS_ID_LEN); | 354 | memcpy(ctlr->subsys_id, subsys_id, SUBSYS_ID_LEN); |
326 | memcpy(ctlr->slot_id, slot_id, SLOT_ID_LEN); | 355 | memcpy(ctlr->slot_id, slot_id, SLOT_ID_LEN); |
356 | memcpy(ctlr->array_name, array_name, ARRAY_LABEL_LEN); | ||
357 | |||
358 | /* update the controller index */ | ||
359 | if (slot_id[1] == 0x31) | ||
360 | ctlr->index = 0; | ||
361 | else | ||
362 | ctlr->index = 1; | ||
363 | |||
327 | kref_init(&ctlr->kref); | 364 | kref_init(&ctlr->kref); |
328 | ctlr->use_ms10 = -1; | 365 | ctlr->use_ms10 = -1; |
329 | list_add(&ctlr->node, &ctlr_list); | 366 | list_add(&ctlr->node, &ctlr_list); |
@@ -363,9 +400,10 @@ done: | |||
363 | return err; | 400 | return err; |
364 | } | 401 | } |
365 | 402 | ||
366 | static int get_lun(struct scsi_device *sdev, struct rdac_dh_data *h) | 403 | static int get_lun_info(struct scsi_device *sdev, struct rdac_dh_data *h, |
404 | char *array_name) | ||
367 | { | 405 | { |
368 | int err; | 406 | int err, i; |
369 | struct c8_inquiry *inqp; | 407 | struct c8_inquiry *inqp; |
370 | 408 | ||
371 | err = submit_inquiry(sdev, 0xC8, sizeof(struct c8_inquiry), h); | 409 | err = submit_inquiry(sdev, 0xC8, sizeof(struct c8_inquiry), h); |
@@ -377,6 +415,11 @@ static int get_lun(struct scsi_device *sdev, struct rdac_dh_data *h) | |||
377 | inqp->page_id[2] != 'i' || inqp->page_id[3] != 'd') | 415 | inqp->page_id[2] != 'i' || inqp->page_id[3] != 'd') |
378 | return SCSI_DH_NOSYS; | 416 | return SCSI_DH_NOSYS; |
379 | h->lun = inqp->lun[7]; /* Uses only the last byte */ | 417 | h->lun = inqp->lun[7]; /* Uses only the last byte */ |
418 | |||
419 | for(i=0; i<ARRAY_LABEL_LEN-1; ++i) | ||
420 | *(array_name+i) = inqp->array_user_label[(2*i)+1]; | ||
421 | |||
422 | *(array_name+ARRAY_LABEL_LEN-1) = '\0'; | ||
380 | } | 423 | } |
381 | return err; | 424 | return err; |
382 | } | 425 | } |
@@ -410,7 +453,7 @@ static int check_ownership(struct scsi_device *sdev, struct rdac_dh_data *h) | |||
410 | } | 453 | } |
411 | 454 | ||
412 | static int initialize_controller(struct scsi_device *sdev, | 455 | static int initialize_controller(struct scsi_device *sdev, |
413 | struct rdac_dh_data *h) | 456 | struct rdac_dh_data *h, char *array_name) |
414 | { | 457 | { |
415 | int err; | 458 | int err; |
416 | struct c4_inquiry *inqp; | 459 | struct c4_inquiry *inqp; |
@@ -418,7 +461,8 @@ static int initialize_controller(struct scsi_device *sdev, | |||
418 | err = submit_inquiry(sdev, 0xC4, sizeof(struct c4_inquiry), h); | 461 | err = submit_inquiry(sdev, 0xC4, sizeof(struct c4_inquiry), h); |
419 | if (err == SCSI_DH_OK) { | 462 | if (err == SCSI_DH_OK) { |
420 | inqp = &h->inq.c4; | 463 | inqp = &h->inq.c4; |
421 | h->ctlr = get_controller(inqp->subsys_id, inqp->slot_id); | 464 | h->ctlr = get_controller(inqp->subsys_id, inqp->slot_id, |
465 | array_name); | ||
422 | if (!h->ctlr) | 466 | if (!h->ctlr) |
423 | err = SCSI_DH_RES_TEMP_UNAVAIL; | 467 | err = SCSI_DH_RES_TEMP_UNAVAIL; |
424 | } | 468 | } |
@@ -450,6 +494,7 @@ static int mode_select_handle_sense(struct scsi_device *sdev, | |||
450 | { | 494 | { |
451 | struct scsi_sense_hdr sense_hdr; | 495 | struct scsi_sense_hdr sense_hdr; |
452 | int err = SCSI_DH_IO, ret; | 496 | int err = SCSI_DH_IO, ret; |
497 | struct rdac_dh_data *h = get_rdac_data(sdev); | ||
453 | 498 | ||
454 | ret = scsi_normalize_sense(sensebuf, SCSI_SENSE_BUFFERSIZE, &sense_hdr); | 499 | ret = scsi_normalize_sense(sensebuf, SCSI_SENSE_BUFFERSIZE, &sense_hdr); |
455 | if (!ret) | 500 | if (!ret) |
@@ -478,11 +523,14 @@ static int mode_select_handle_sense(struct scsi_device *sdev, | |||
478 | err = SCSI_DH_RETRY; | 523 | err = SCSI_DH_RETRY; |
479 | break; | 524 | break; |
480 | default: | 525 | default: |
481 | sdev_printk(KERN_INFO, sdev, | 526 | break; |
482 | "MODE_SELECT failed with sense %02x/%02x/%02x.\n", | ||
483 | sense_hdr.sense_key, sense_hdr.asc, sense_hdr.ascq); | ||
484 | } | 527 | } |
485 | 528 | ||
529 | RDAC_LOG(RDAC_LOG_FAILOVER, sdev, "array %s, ctlr %d, " | ||
530 | "MODE_SELECT returned with sense %02x/%02x/%02x", | ||
531 | (char *) h->ctlr->array_name, h->ctlr->index, | ||
532 | sense_hdr.sense_key, sense_hdr.asc, sense_hdr.ascq); | ||
533 | |||
486 | done: | 534 | done: |
487 | return err; | 535 | return err; |
488 | } | 536 | } |
@@ -499,7 +547,9 @@ retry: | |||
499 | if (!rq) | 547 | if (!rq) |
500 | goto done; | 548 | goto done; |
501 | 549 | ||
502 | sdev_printk(KERN_INFO, sdev, "%s MODE_SELECT command.\n", | 550 | RDAC_LOG(RDAC_LOG_FAILOVER, sdev, "array %s, ctlr %d, " |
551 | "%s MODE_SELECT command", | ||
552 | (char *) h->ctlr->array_name, h->ctlr->index, | ||
503 | (retry_cnt == RDAC_RETRY_COUNT) ? "queueing" : "retrying"); | 553 | (retry_cnt == RDAC_RETRY_COUNT) ? "queueing" : "retrying"); |
504 | 554 | ||
505 | err = blk_execute_rq(q, NULL, rq, 1); | 555 | err = blk_execute_rq(q, NULL, rq, 1); |
@@ -509,8 +559,12 @@ retry: | |||
509 | if (err == SCSI_DH_RETRY && retry_cnt--) | 559 | if (err == SCSI_DH_RETRY && retry_cnt--) |
510 | goto retry; | 560 | goto retry; |
511 | } | 561 | } |
512 | if (err == SCSI_DH_OK) | 562 | if (err == SCSI_DH_OK) { |
513 | h->state = RDAC_STATE_ACTIVE; | 563 | h->state = RDAC_STATE_ACTIVE; |
564 | RDAC_LOG(RDAC_LOG_FAILOVER, sdev, "array %s, ctlr %d, " | ||
565 | "MODE_SELECT completed", | ||
566 | (char *) h->ctlr->array_name, h->ctlr->index); | ||
567 | } | ||
514 | 568 | ||
515 | done: | 569 | done: |
516 | return err; | 570 | return err; |
@@ -525,17 +579,6 @@ static int rdac_activate(struct scsi_device *sdev) | |||
525 | if (err != SCSI_DH_OK) | 579 | if (err != SCSI_DH_OK) |
526 | goto done; | 580 | goto done; |
527 | 581 | ||
528 | if (!h->ctlr) { | ||
529 | err = initialize_controller(sdev, h); | ||
530 | if (err != SCSI_DH_OK) | ||
531 | goto done; | ||
532 | } | ||
533 | |||
534 | if (h->ctlr->use_ms10 == -1) { | ||
535 | err = set_mode_select(sdev, h); | ||
536 | if (err != SCSI_DH_OK) | ||
537 | goto done; | ||
538 | } | ||
539 | if (h->lun_state == RDAC_LUN_UNOWNED) | 582 | if (h->lun_state == RDAC_LUN_UNOWNED) |
540 | err = send_mode_select(sdev, h); | 583 | err = send_mode_select(sdev, h); |
541 | done: | 584 | done: |
@@ -559,6 +602,12 @@ static int rdac_check_sense(struct scsi_device *sdev, | |||
559 | struct scsi_sense_hdr *sense_hdr) | 602 | struct scsi_sense_hdr *sense_hdr) |
560 | { | 603 | { |
561 | struct rdac_dh_data *h = get_rdac_data(sdev); | 604 | struct rdac_dh_data *h = get_rdac_data(sdev); |
605 | |||
606 | RDAC_LOG(RDAC_LOG_SENSE, sdev, "array %s, ctlr %d, " | ||
607 | "I/O returned with sense %02x/%02x/%02x", | ||
608 | (char *) h->ctlr->array_name, h->ctlr->index, | ||
609 | sense_hdr->sense_key, sense_hdr->asc, sense_hdr->ascq); | ||
610 | |||
562 | switch (sense_hdr->sense_key) { | 611 | switch (sense_hdr->sense_key) { |
563 | case NOT_READY: | 612 | case NOT_READY: |
564 | if (sense_hdr->asc == 0x04 && sense_hdr->ascq == 0x01) | 613 | if (sense_hdr->asc == 0x04 && sense_hdr->ascq == 0x01) |
@@ -628,11 +677,18 @@ static const struct scsi_dh_devlist rdac_dev_list[] = { | |||
628 | {"SGI", "IS"}, | 677 | {"SGI", "IS"}, |
629 | {"STK", "OPENstorage D280"}, | 678 | {"STK", "OPENstorage D280"}, |
630 | {"SUN", "CSM200_R"}, | 679 | {"SUN", "CSM200_R"}, |
680 | {"SUN", "LCSM100_I"}, | ||
681 | {"SUN", "LCSM100_S"}, | ||
682 | {"SUN", "LCSM100_E"}, | ||
631 | {"SUN", "LCSM100_F"}, | 683 | {"SUN", "LCSM100_F"}, |
632 | {"DELL", "MD3000"}, | 684 | {"DELL", "MD3000"}, |
633 | {"DELL", "MD3000i"}, | 685 | {"DELL", "MD3000i"}, |
686 | {"DELL", "MD32xx"}, | ||
687 | {"DELL", "MD32xxi"}, | ||
634 | {"LSI", "INF-01-00"}, | 688 | {"LSI", "INF-01-00"}, |
635 | {"ENGENIO", "INF-01-00"}, | 689 | {"ENGENIO", "INF-01-00"}, |
690 | {"STK", "FLEXLINE 380"}, | ||
691 | {"SUN", "CSM100_R_FC"}, | ||
636 | {NULL, NULL}, | 692 | {NULL, NULL}, |
637 | }; | 693 | }; |
638 | 694 | ||
@@ -656,6 +712,7 @@ static int rdac_bus_attach(struct scsi_device *sdev) | |||
656 | struct rdac_dh_data *h; | 712 | struct rdac_dh_data *h; |
657 | unsigned long flags; | 713 | unsigned long flags; |
658 | int err; | 714 | int err; |
715 | char array_name[ARRAY_LABEL_LEN]; | ||
659 | 716 | ||
660 | scsi_dh_data = kzalloc(sizeof(struct scsi_device_handler *) | 717 | scsi_dh_data = kzalloc(sizeof(struct scsi_device_handler *) |
661 | + sizeof(*h) , GFP_KERNEL); | 718 | + sizeof(*h) , GFP_KERNEL); |
@@ -670,16 +727,24 @@ static int rdac_bus_attach(struct scsi_device *sdev) | |||
670 | h->lun = UNINITIALIZED_LUN; | 727 | h->lun = UNINITIALIZED_LUN; |
671 | h->state = RDAC_STATE_ACTIVE; | 728 | h->state = RDAC_STATE_ACTIVE; |
672 | 729 | ||
673 | err = get_lun(sdev, h); | 730 | err = get_lun_info(sdev, h, array_name); |
674 | if (err != SCSI_DH_OK) | 731 | if (err != SCSI_DH_OK) |
675 | goto failed; | 732 | goto failed; |
676 | 733 | ||
677 | err = check_ownership(sdev, h); | 734 | err = initialize_controller(sdev, h, array_name); |
678 | if (err != SCSI_DH_OK) | 735 | if (err != SCSI_DH_OK) |
679 | goto failed; | 736 | goto failed; |
680 | 737 | ||
738 | err = check_ownership(sdev, h); | ||
739 | if (err != SCSI_DH_OK) | ||
740 | goto clean_ctlr; | ||
741 | |||
742 | err = set_mode_select(sdev, h); | ||
743 | if (err != SCSI_DH_OK) | ||
744 | goto clean_ctlr; | ||
745 | |||
681 | if (!try_module_get(THIS_MODULE)) | 746 | if (!try_module_get(THIS_MODULE)) |
682 | goto failed; | 747 | goto clean_ctlr; |
683 | 748 | ||
684 | spin_lock_irqsave(sdev->request_queue->queue_lock, flags); | 749 | spin_lock_irqsave(sdev->request_queue->queue_lock, flags); |
685 | sdev->scsi_dh_data = scsi_dh_data; | 750 | sdev->scsi_dh_data = scsi_dh_data; |
@@ -691,6 +756,9 @@ static int rdac_bus_attach(struct scsi_device *sdev) | |||
691 | 756 | ||
692 | return 0; | 757 | return 0; |
693 | 758 | ||
759 | clean_ctlr: | ||
760 | kref_put(&h->ctlr->kref, release_controller); | ||
761 | |||
694 | failed: | 762 | failed: |
695 | kfree(scsi_dh_data); | 763 | kfree(scsi_dh_data); |
696 | sdev_printk(KERN_ERR, sdev, "%s: not attached\n", | 764 | sdev_printk(KERN_ERR, sdev, "%s: not attached\n", |
diff --git a/drivers/scsi/fcoe/fcoe.c b/drivers/scsi/fcoe/fcoe.c index 0a5609bb5817..704b8e034946 100644 --- a/drivers/scsi/fcoe/fcoe.c +++ b/drivers/scsi/fcoe/fcoe.c | |||
@@ -1,5 +1,5 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright(c) 2007 - 2008 Intel Corporation. All rights reserved. | 2 | * Copyright(c) 2007 - 2009 Intel Corporation. All rights reserved. |
3 | * | 3 | * |
4 | * This program is free software; you can redistribute it and/or modify it | 4 | * This program is free software; you can redistribute it and/or modify it |
5 | * under the terms and conditions of the GNU General Public License, | 5 | * under the terms and conditions of the GNU General Public License, |
@@ -49,9 +49,20 @@ MODULE_AUTHOR("Open-FCoE.org"); | |||
49 | MODULE_DESCRIPTION("FCoE"); | 49 | MODULE_DESCRIPTION("FCoE"); |
50 | MODULE_LICENSE("GPL v2"); | 50 | MODULE_LICENSE("GPL v2"); |
51 | 51 | ||
52 | /* Performance tuning parameters for fcoe */ | ||
53 | static unsigned int fcoe_ddp_min; | ||
54 | module_param_named(ddp_min, fcoe_ddp_min, uint, S_IRUGO | S_IWUSR); | ||
55 | MODULE_PARM_DESC(ddp_min, "Minimum I/O size in bytes for " \ | ||
56 | "Direct Data Placement (DDP)."); | ||
57 | |||
58 | DEFINE_MUTEX(fcoe_config_mutex); | ||
59 | |||
60 | /* fcoe_percpu_clean completion. Waiter protected by fcoe_create_mutex */ | ||
61 | static DECLARE_COMPLETION(fcoe_flush_completion); | ||
62 | |||
52 | /* fcoe host list */ | 63 | /* fcoe host list */ |
64 | /* must only by accessed under the RTNL mutex */ | ||
53 | LIST_HEAD(fcoe_hostlist); | 65 | LIST_HEAD(fcoe_hostlist); |
54 | DEFINE_RWLOCK(fcoe_hostlist_lock); | ||
55 | DEFINE_PER_CPU(struct fcoe_percpu_s, fcoe_percpu); | 66 | DEFINE_PER_CPU(struct fcoe_percpu_s, fcoe_percpu); |
56 | 67 | ||
57 | /* Function Prototypes */ | 68 | /* Function Prototypes */ |
@@ -66,12 +77,13 @@ static int fcoe_link_ok(struct fc_lport *lp); | |||
66 | 77 | ||
67 | static struct fc_lport *fcoe_hostlist_lookup(const struct net_device *); | 78 | static struct fc_lport *fcoe_hostlist_lookup(const struct net_device *); |
68 | static int fcoe_hostlist_add(const struct fc_lport *); | 79 | static int fcoe_hostlist_add(const struct fc_lport *); |
69 | static int fcoe_hostlist_remove(const struct fc_lport *); | ||
70 | 80 | ||
71 | static void fcoe_check_wait_queue(struct fc_lport *, struct sk_buff *); | 81 | static void fcoe_check_wait_queue(struct fc_lport *, struct sk_buff *); |
72 | static int fcoe_device_notification(struct notifier_block *, ulong, void *); | 82 | static int fcoe_device_notification(struct notifier_block *, ulong, void *); |
73 | static void fcoe_dev_setup(void); | 83 | static void fcoe_dev_setup(void); |
74 | static void fcoe_dev_cleanup(void); | 84 | static void fcoe_dev_cleanup(void); |
85 | static struct fcoe_interface * | ||
86 | fcoe_hostlist_lookup_port(const struct net_device *dev); | ||
75 | 87 | ||
76 | /* notification function from net device */ | 88 | /* notification function from net device */ |
77 | static struct notifier_block fcoe_notifier = { | 89 | static struct notifier_block fcoe_notifier = { |
@@ -132,6 +144,180 @@ static struct scsi_host_template fcoe_shost_template = { | |||
132 | .max_sectors = 0xffff, | 144 | .max_sectors = 0xffff, |
133 | }; | 145 | }; |
134 | 146 | ||
147 | static int fcoe_fip_recv(struct sk_buff *skb, struct net_device *dev, | ||
148 | struct packet_type *ptype, | ||
149 | struct net_device *orig_dev); | ||
150 | /** | ||
151 | * fcoe_interface_setup() | ||
152 | * @fcoe: new fcoe_interface | ||
153 | * @netdev : ptr to the associated netdevice struct | ||
154 | * | ||
155 | * Returns : 0 for success | ||
156 | * Locking: must be called with the RTNL mutex held | ||
157 | */ | ||
158 | static int fcoe_interface_setup(struct fcoe_interface *fcoe, | ||
159 | struct net_device *netdev) | ||
160 | { | ||
161 | struct fcoe_ctlr *fip = &fcoe->ctlr; | ||
162 | struct netdev_hw_addr *ha; | ||
163 | u8 flogi_maddr[ETH_ALEN]; | ||
164 | |||
165 | fcoe->netdev = netdev; | ||
166 | |||
167 | /* Do not support for bonding device */ | ||
168 | if ((netdev->priv_flags & IFF_MASTER_ALB) || | ||
169 | (netdev->priv_flags & IFF_SLAVE_INACTIVE) || | ||
170 | (netdev->priv_flags & IFF_MASTER_8023AD)) { | ||
171 | return -EOPNOTSUPP; | ||
172 | } | ||
173 | |||
174 | /* look for SAN MAC address, if multiple SAN MACs exist, only | ||
175 | * use the first one for SPMA */ | ||
176 | rcu_read_lock(); | ||
177 | for_each_dev_addr(netdev, ha) { | ||
178 | if ((ha->type == NETDEV_HW_ADDR_T_SAN) && | ||
179 | (is_valid_ether_addr(fip->ctl_src_addr))) { | ||
180 | memcpy(fip->ctl_src_addr, ha->addr, ETH_ALEN); | ||
181 | fip->spma = 1; | ||
182 | break; | ||
183 | } | ||
184 | } | ||
185 | rcu_read_unlock(); | ||
186 | |||
187 | /* setup Source Mac Address */ | ||
188 | if (!fip->spma) | ||
189 | memcpy(fip->ctl_src_addr, netdev->dev_addr, netdev->addr_len); | ||
190 | |||
191 | /* | ||
192 | * Add FCoE MAC address as second unicast MAC address | ||
193 | * or enter promiscuous mode if not capable of listening | ||
194 | * for multiple unicast MACs. | ||
195 | */ | ||
196 | memcpy(flogi_maddr, (u8[6]) FC_FCOE_FLOGI_MAC, ETH_ALEN); | ||
197 | dev_unicast_add(netdev, flogi_maddr); | ||
198 | if (fip->spma) | ||
199 | dev_unicast_add(netdev, fip->ctl_src_addr); | ||
200 | dev_mc_add(netdev, FIP_ALL_ENODE_MACS, ETH_ALEN, 0); | ||
201 | |||
202 | /* | ||
203 | * setup the receive function from ethernet driver | ||
204 | * on the ethertype for the given device | ||
205 | */ | ||
206 | fcoe->fcoe_packet_type.func = fcoe_rcv; | ||
207 | fcoe->fcoe_packet_type.type = __constant_htons(ETH_P_FCOE); | ||
208 | fcoe->fcoe_packet_type.dev = netdev; | ||
209 | dev_add_pack(&fcoe->fcoe_packet_type); | ||
210 | |||
211 | fcoe->fip_packet_type.func = fcoe_fip_recv; | ||
212 | fcoe->fip_packet_type.type = htons(ETH_P_FIP); | ||
213 | fcoe->fip_packet_type.dev = netdev; | ||
214 | dev_add_pack(&fcoe->fip_packet_type); | ||
215 | |||
216 | return 0; | ||
217 | } | ||
218 | |||
219 | static void fcoe_fip_send(struct fcoe_ctlr *fip, struct sk_buff *skb); | ||
220 | static void fcoe_update_src_mac(struct fcoe_ctlr *fip, u8 *old, u8 *new); | ||
221 | static void fcoe_destroy_work(struct work_struct *work); | ||
222 | |||
223 | /** | ||
224 | * fcoe_interface_create() | ||
225 | * @netdev: network interface | ||
226 | * | ||
227 | * Returns: pointer to a struct fcoe_interface or NULL on error | ||
228 | */ | ||
229 | static struct fcoe_interface *fcoe_interface_create(struct net_device *netdev) | ||
230 | { | ||
231 | struct fcoe_interface *fcoe; | ||
232 | |||
233 | fcoe = kzalloc(sizeof(*fcoe), GFP_KERNEL); | ||
234 | if (!fcoe) { | ||
235 | FCOE_NETDEV_DBG(netdev, "Could not allocate fcoe structure\n"); | ||
236 | return NULL; | ||
237 | } | ||
238 | |||
239 | dev_hold(netdev); | ||
240 | kref_init(&fcoe->kref); | ||
241 | |||
242 | /* | ||
243 | * Initialize FIP. | ||
244 | */ | ||
245 | fcoe_ctlr_init(&fcoe->ctlr); | ||
246 | fcoe->ctlr.send = fcoe_fip_send; | ||
247 | fcoe->ctlr.update_mac = fcoe_update_src_mac; | ||
248 | |||
249 | fcoe_interface_setup(fcoe, netdev); | ||
250 | |||
251 | return fcoe; | ||
252 | } | ||
253 | |||
254 | /** | ||
255 | * fcoe_interface_cleanup() - clean up netdev configurations | ||
256 | * @fcoe: | ||
257 | * | ||
258 | * Caller must be holding the RTNL mutex | ||
259 | */ | ||
260 | void fcoe_interface_cleanup(struct fcoe_interface *fcoe) | ||
261 | { | ||
262 | struct net_device *netdev = fcoe->netdev; | ||
263 | struct fcoe_ctlr *fip = &fcoe->ctlr; | ||
264 | u8 flogi_maddr[ETH_ALEN]; | ||
265 | |||
266 | /* | ||
267 | * Don't listen for Ethernet packets anymore. | ||
268 | * synchronize_net() ensures that the packet handlers are not running | ||
269 | * on another CPU. dev_remove_pack() would do that, this calls the | ||
270 | * unsyncronized version __dev_remove_pack() to avoid multiple delays. | ||
271 | */ | ||
272 | __dev_remove_pack(&fcoe->fcoe_packet_type); | ||
273 | __dev_remove_pack(&fcoe->fip_packet_type); | ||
274 | synchronize_net(); | ||
275 | |||
276 | /* Delete secondary MAC addresses */ | ||
277 | memcpy(flogi_maddr, (u8[6]) FC_FCOE_FLOGI_MAC, ETH_ALEN); | ||
278 | dev_unicast_delete(netdev, flogi_maddr); | ||
279 | if (!is_zero_ether_addr(fip->data_src_addr)) | ||
280 | dev_unicast_delete(netdev, fip->data_src_addr); | ||
281 | if (fip->spma) | ||
282 | dev_unicast_delete(netdev, fip->ctl_src_addr); | ||
283 | dev_mc_delete(netdev, FIP_ALL_ENODE_MACS, ETH_ALEN, 0); | ||
284 | } | ||
285 | |||
286 | /** | ||
287 | * fcoe_interface_release() - fcoe_port kref release function | ||
288 | * @kref: embedded reference count in an fcoe_interface struct | ||
289 | */ | ||
290 | static void fcoe_interface_release(struct kref *kref) | ||
291 | { | ||
292 | struct fcoe_interface *fcoe; | ||
293 | struct net_device *netdev; | ||
294 | |||
295 | fcoe = container_of(kref, struct fcoe_interface, kref); | ||
296 | netdev = fcoe->netdev; | ||
297 | /* tear-down the FCoE controller */ | ||
298 | fcoe_ctlr_destroy(&fcoe->ctlr); | ||
299 | kfree(fcoe); | ||
300 | dev_put(netdev); | ||
301 | } | ||
302 | |||
303 | /** | ||
304 | * fcoe_interface_get() | ||
305 | * @fcoe: | ||
306 | */ | ||
307 | static inline void fcoe_interface_get(struct fcoe_interface *fcoe) | ||
308 | { | ||
309 | kref_get(&fcoe->kref); | ||
310 | } | ||
311 | |||
312 | /** | ||
313 | * fcoe_interface_put() | ||
314 | * @fcoe: | ||
315 | */ | ||
316 | static inline void fcoe_interface_put(struct fcoe_interface *fcoe) | ||
317 | { | ||
318 | kref_put(&fcoe->kref, fcoe_interface_release); | ||
319 | } | ||
320 | |||
135 | /** | 321 | /** |
136 | * fcoe_fip_recv - handle a received FIP frame. | 322 | * fcoe_fip_recv - handle a received FIP frame. |
137 | * @skb: the receive skb | 323 | * @skb: the receive skb |
@@ -145,10 +331,10 @@ static int fcoe_fip_recv(struct sk_buff *skb, struct net_device *dev, | |||
145 | struct packet_type *ptype, | 331 | struct packet_type *ptype, |
146 | struct net_device *orig_dev) | 332 | struct net_device *orig_dev) |
147 | { | 333 | { |
148 | struct fcoe_softc *fc; | 334 | struct fcoe_interface *fcoe; |
149 | 335 | ||
150 | fc = container_of(ptype, struct fcoe_softc, fip_packet_type); | 336 | fcoe = container_of(ptype, struct fcoe_interface, fip_packet_type); |
151 | fcoe_ctlr_recv(&fc->ctlr, skb); | 337 | fcoe_ctlr_recv(&fcoe->ctlr, skb); |
152 | return 0; | 338 | return 0; |
153 | } | 339 | } |
154 | 340 | ||
@@ -159,7 +345,7 @@ static int fcoe_fip_recv(struct sk_buff *skb, struct net_device *dev, | |||
159 | */ | 345 | */ |
160 | static void fcoe_fip_send(struct fcoe_ctlr *fip, struct sk_buff *skb) | 346 | static void fcoe_fip_send(struct fcoe_ctlr *fip, struct sk_buff *skb) |
161 | { | 347 | { |
162 | skb->dev = fcoe_from_ctlr(fip)->real_dev; | 348 | skb->dev = fcoe_from_ctlr(fip)->netdev; |
163 | dev_queue_xmit(skb); | 349 | dev_queue_xmit(skb); |
164 | } | 350 | } |
165 | 351 | ||
@@ -174,13 +360,13 @@ static void fcoe_fip_send(struct fcoe_ctlr *fip, struct sk_buff *skb) | |||
174 | */ | 360 | */ |
175 | static void fcoe_update_src_mac(struct fcoe_ctlr *fip, u8 *old, u8 *new) | 361 | static void fcoe_update_src_mac(struct fcoe_ctlr *fip, u8 *old, u8 *new) |
176 | { | 362 | { |
177 | struct fcoe_softc *fc; | 363 | struct fcoe_interface *fcoe; |
178 | 364 | ||
179 | fc = fcoe_from_ctlr(fip); | 365 | fcoe = fcoe_from_ctlr(fip); |
180 | rtnl_lock(); | 366 | rtnl_lock(); |
181 | if (!is_zero_ether_addr(old)) | 367 | if (!is_zero_ether_addr(old)) |
182 | dev_unicast_delete(fc->real_dev, old); | 368 | dev_unicast_delete(fcoe->netdev, old); |
183 | dev_unicast_add(fc->real_dev, new); | 369 | dev_unicast_add(fcoe->netdev, new); |
184 | rtnl_unlock(); | 370 | rtnl_unlock(); |
185 | } | 371 | } |
186 | 372 | ||
@@ -217,30 +403,6 @@ static int fcoe_lport_config(struct fc_lport *lp) | |||
217 | } | 403 | } |
218 | 404 | ||
219 | /** | 405 | /** |
220 | * fcoe_netdev_cleanup() - clean up netdev configurations | ||
221 | * @fc: ptr to the fcoe_softc | ||
222 | */ | ||
223 | void fcoe_netdev_cleanup(struct fcoe_softc *fc) | ||
224 | { | ||
225 | u8 flogi_maddr[ETH_ALEN]; | ||
226 | |||
227 | /* Don't listen for Ethernet packets anymore */ | ||
228 | dev_remove_pack(&fc->fcoe_packet_type); | ||
229 | dev_remove_pack(&fc->fip_packet_type); | ||
230 | |||
231 | /* Delete secondary MAC addresses */ | ||
232 | rtnl_lock(); | ||
233 | memcpy(flogi_maddr, (u8[6]) FC_FCOE_FLOGI_MAC, ETH_ALEN); | ||
234 | dev_unicast_delete(fc->real_dev, flogi_maddr); | ||
235 | if (!is_zero_ether_addr(fc->ctlr.data_src_addr)) | ||
236 | dev_unicast_delete(fc->real_dev, fc->ctlr.data_src_addr); | ||
237 | if (fc->ctlr.spma) | ||
238 | dev_unicast_delete(fc->real_dev, fc->ctlr.ctl_src_addr); | ||
239 | dev_mc_delete(fc->real_dev, FIP_ALL_ENODE_MACS, ETH_ALEN, 0); | ||
240 | rtnl_unlock(); | ||
241 | } | ||
242 | |||
243 | /** | ||
244 | * fcoe_queue_timer() - fcoe queue timer | 406 | * fcoe_queue_timer() - fcoe queue timer |
245 | * @lp: the fc_lport pointer | 407 | * @lp: the fc_lport pointer |
246 | * | 408 | * |
@@ -265,116 +427,53 @@ static int fcoe_netdev_config(struct fc_lport *lp, struct net_device *netdev) | |||
265 | { | 427 | { |
266 | u32 mfs; | 428 | u32 mfs; |
267 | u64 wwnn, wwpn; | 429 | u64 wwnn, wwpn; |
268 | struct fcoe_softc *fc; | 430 | struct fcoe_interface *fcoe; |
269 | u8 flogi_maddr[ETH_ALEN]; | 431 | struct fcoe_port *port; |
270 | struct netdev_hw_addr *ha; | ||
271 | 432 | ||
272 | /* Setup lport private data to point to fcoe softc */ | 433 | /* Setup lport private data to point to fcoe softc */ |
273 | fc = lport_priv(lp); | 434 | port = lport_priv(lp); |
274 | fc->ctlr.lp = lp; | 435 | fcoe = port->fcoe; |
275 | fc->real_dev = netdev; | ||
276 | fc->phys_dev = netdev; | ||
277 | |||
278 | /* Require support for get_pauseparam ethtool op. */ | ||
279 | if (netdev->priv_flags & IFF_802_1Q_VLAN) | ||
280 | fc->phys_dev = vlan_dev_real_dev(netdev); | ||
281 | |||
282 | /* Do not support for bonding device */ | ||
283 | if ((fc->real_dev->priv_flags & IFF_MASTER_ALB) || | ||
284 | (fc->real_dev->priv_flags & IFF_SLAVE_INACTIVE) || | ||
285 | (fc->real_dev->priv_flags & IFF_MASTER_8023AD)) { | ||
286 | return -EOPNOTSUPP; | ||
287 | } | ||
288 | 436 | ||
289 | /* | 437 | /* |
290 | * Determine max frame size based on underlying device and optional | 438 | * Determine max frame size based on underlying device and optional |
291 | * user-configured limit. If the MFS is too low, fcoe_link_ok() | 439 | * user-configured limit. If the MFS is too low, fcoe_link_ok() |
292 | * will return 0, so do this first. | 440 | * will return 0, so do this first. |
293 | */ | 441 | */ |
294 | mfs = fc->real_dev->mtu - (sizeof(struct fcoe_hdr) + | 442 | mfs = netdev->mtu - (sizeof(struct fcoe_hdr) + |
295 | sizeof(struct fcoe_crc_eof)); | 443 | sizeof(struct fcoe_crc_eof)); |
296 | if (fc_set_mfs(lp, mfs)) | 444 | if (fc_set_mfs(lp, mfs)) |
297 | return -EINVAL; | 445 | return -EINVAL; |
298 | 446 | ||
299 | /* offload features support */ | 447 | /* offload features support */ |
300 | if (fc->real_dev->features & NETIF_F_SG) | 448 | if (netdev->features & NETIF_F_SG) |
301 | lp->sg_supp = 1; | 449 | lp->sg_supp = 1; |
302 | 450 | ||
303 | #ifdef NETIF_F_FCOE_CRC | ||
304 | if (netdev->features & NETIF_F_FCOE_CRC) { | 451 | if (netdev->features & NETIF_F_FCOE_CRC) { |
305 | lp->crc_offload = 1; | 452 | lp->crc_offload = 1; |
306 | FCOE_NETDEV_DBG(netdev, "Supports FCCRC offload\n"); | 453 | FCOE_NETDEV_DBG(netdev, "Supports FCCRC offload\n"); |
307 | } | 454 | } |
308 | #endif | ||
309 | #ifdef NETIF_F_FSO | ||
310 | if (netdev->features & NETIF_F_FSO) { | 455 | if (netdev->features & NETIF_F_FSO) { |
311 | lp->seq_offload = 1; | 456 | lp->seq_offload = 1; |
312 | lp->lso_max = netdev->gso_max_size; | 457 | lp->lso_max = netdev->gso_max_size; |
313 | FCOE_NETDEV_DBG(netdev, "Supports LSO for max len 0x%x\n", | 458 | FCOE_NETDEV_DBG(netdev, "Supports LSO for max len 0x%x\n", |
314 | lp->lso_max); | 459 | lp->lso_max); |
315 | } | 460 | } |
316 | #endif | ||
317 | if (netdev->fcoe_ddp_xid) { | 461 | if (netdev->fcoe_ddp_xid) { |
318 | lp->lro_enabled = 1; | 462 | lp->lro_enabled = 1; |
319 | lp->lro_xid = netdev->fcoe_ddp_xid; | 463 | lp->lro_xid = netdev->fcoe_ddp_xid; |
320 | FCOE_NETDEV_DBG(netdev, "Supports LRO for max xid 0x%x\n", | 464 | FCOE_NETDEV_DBG(netdev, "Supports LRO for max xid 0x%x\n", |
321 | lp->lro_xid); | 465 | lp->lro_xid); |
322 | } | 466 | } |
323 | skb_queue_head_init(&fc->fcoe_pending_queue); | 467 | skb_queue_head_init(&port->fcoe_pending_queue); |
324 | fc->fcoe_pending_queue_active = 0; | 468 | port->fcoe_pending_queue_active = 0; |
325 | setup_timer(&fc->timer, fcoe_queue_timer, (unsigned long)lp); | 469 | setup_timer(&port->timer, fcoe_queue_timer, (unsigned long)lp); |
326 | |||
327 | /* look for SAN MAC address, if multiple SAN MACs exist, only | ||
328 | * use the first one for SPMA */ | ||
329 | rcu_read_lock(); | ||
330 | for_each_dev_addr(netdev, ha) { | ||
331 | if ((ha->type == NETDEV_HW_ADDR_T_SAN) && | ||
332 | (is_valid_ether_addr(fc->ctlr.ctl_src_addr))) { | ||
333 | memcpy(fc->ctlr.ctl_src_addr, ha->addr, ETH_ALEN); | ||
334 | fc->ctlr.spma = 1; | ||
335 | break; | ||
336 | } | ||
337 | } | ||
338 | rcu_read_unlock(); | ||
339 | |||
340 | /* setup Source Mac Address */ | ||
341 | if (!fc->ctlr.spma) | ||
342 | memcpy(fc->ctlr.ctl_src_addr, fc->real_dev->dev_addr, | ||
343 | fc->real_dev->addr_len); | ||
344 | 470 | ||
345 | wwnn = fcoe_wwn_from_mac(fc->real_dev->dev_addr, 1, 0); | 471 | wwnn = fcoe_wwn_from_mac(netdev->dev_addr, 1, 0); |
346 | fc_set_wwnn(lp, wwnn); | 472 | fc_set_wwnn(lp, wwnn); |
347 | /* XXX - 3rd arg needs to be vlan id */ | 473 | /* XXX - 3rd arg needs to be vlan id */ |
348 | wwpn = fcoe_wwn_from_mac(fc->real_dev->dev_addr, 2, 0); | 474 | wwpn = fcoe_wwn_from_mac(netdev->dev_addr, 2, 0); |
349 | fc_set_wwpn(lp, wwpn); | 475 | fc_set_wwpn(lp, wwpn); |
350 | 476 | ||
351 | /* | ||
352 | * Add FCoE MAC address as second unicast MAC address | ||
353 | * or enter promiscuous mode if not capable of listening | ||
354 | * for multiple unicast MACs. | ||
355 | */ | ||
356 | rtnl_lock(); | ||
357 | memcpy(flogi_maddr, (u8[6]) FC_FCOE_FLOGI_MAC, ETH_ALEN); | ||
358 | dev_unicast_add(fc->real_dev, flogi_maddr); | ||
359 | if (fc->ctlr.spma) | ||
360 | dev_unicast_add(fc->real_dev, fc->ctlr.ctl_src_addr); | ||
361 | dev_mc_add(fc->real_dev, FIP_ALL_ENODE_MACS, ETH_ALEN, 0); | ||
362 | rtnl_unlock(); | ||
363 | |||
364 | /* | ||
365 | * setup the receive function from ethernet driver | ||
366 | * on the ethertype for the given device | ||
367 | */ | ||
368 | fc->fcoe_packet_type.func = fcoe_rcv; | ||
369 | fc->fcoe_packet_type.type = __constant_htons(ETH_P_FCOE); | ||
370 | fc->fcoe_packet_type.dev = fc->real_dev; | ||
371 | dev_add_pack(&fc->fcoe_packet_type); | ||
372 | |||
373 | fc->fip_packet_type.func = fcoe_fip_recv; | ||
374 | fc->fip_packet_type.type = htons(ETH_P_FIP); | ||
375 | fc->fip_packet_type.dev = fc->real_dev; | ||
376 | dev_add_pack(&fc->fip_packet_type); | ||
377 | |||
378 | return 0; | 477 | return 0; |
379 | } | 478 | } |
380 | 479 | ||
@@ -415,86 +514,140 @@ static int fcoe_shost_config(struct fc_lport *lp, struct Scsi_Host *shost, | |||
415 | return 0; | 514 | return 0; |
416 | } | 515 | } |
417 | 516 | ||
517 | /* | ||
518 | * fcoe_oem_match() - match for read types IO | ||
519 | * @fp: the fc_frame for new IO. | ||
520 | * | ||
521 | * Returns : true for read types IO, otherwise returns false. | ||
522 | */ | ||
523 | bool fcoe_oem_match(struct fc_frame *fp) | ||
524 | { | ||
525 | return fc_fcp_is_read(fr_fsp(fp)) && | ||
526 | (fr_fsp(fp)->data_len > fcoe_ddp_min); | ||
527 | } | ||
528 | |||
418 | /** | 529 | /** |
419 | * fcoe_em_config() - allocates em for this lport | 530 | * fcoe_em_config() - allocates em for this lport |
420 | * @lp: the port that em is to allocated for | 531 | * @lp: the fcoe that em is to allocated for |
421 | * | 532 | * |
422 | * Returns : 0 on success | 533 | * Returns : 0 on success |
423 | */ | 534 | */ |
424 | static inline int fcoe_em_config(struct fc_lport *lp) | 535 | static inline int fcoe_em_config(struct fc_lport *lp) |
425 | { | 536 | { |
426 | BUG_ON(lp->emp); | 537 | struct fcoe_port *port = lport_priv(lp); |
538 | struct fcoe_interface *fcoe = port->fcoe; | ||
539 | struct fcoe_interface *oldfcoe = NULL; | ||
540 | struct net_device *old_real_dev, *cur_real_dev; | ||
541 | u16 min_xid = FCOE_MIN_XID; | ||
542 | u16 max_xid = FCOE_MAX_XID; | ||
427 | 543 | ||
428 | lp->emp = fc_exch_mgr_alloc(lp, FC_CLASS_3, | 544 | /* |
429 | FCOE_MIN_XID, FCOE_MAX_XID); | 545 | * Check if need to allocate an em instance for |
430 | if (!lp->emp) | 546 | * offload exchange ids to be shared across all VN_PORTs/lport. |
547 | */ | ||
548 | if (!lp->lro_enabled || !lp->lro_xid || (lp->lro_xid >= max_xid)) { | ||
549 | lp->lro_xid = 0; | ||
550 | goto skip_oem; | ||
551 | } | ||
552 | |||
553 | /* | ||
554 | * Reuse existing offload em instance in case | ||
555 | * it is already allocated on real eth device | ||
556 | */ | ||
557 | if (fcoe->netdev->priv_flags & IFF_802_1Q_VLAN) | ||
558 | cur_real_dev = vlan_dev_real_dev(fcoe->netdev); | ||
559 | else | ||
560 | cur_real_dev = fcoe->netdev; | ||
561 | |||
562 | list_for_each_entry(oldfcoe, &fcoe_hostlist, list) { | ||
563 | if (oldfcoe->netdev->priv_flags & IFF_802_1Q_VLAN) | ||
564 | old_real_dev = vlan_dev_real_dev(oldfcoe->netdev); | ||
565 | else | ||
566 | old_real_dev = oldfcoe->netdev; | ||
567 | |||
568 | if (cur_real_dev == old_real_dev) { | ||
569 | fcoe->oem = oldfcoe->oem; | ||
570 | break; | ||
571 | } | ||
572 | } | ||
573 | |||
574 | if (fcoe->oem) { | ||
575 | if (!fc_exch_mgr_add(lp, fcoe->oem, fcoe_oem_match)) { | ||
576 | printk(KERN_ERR "fcoe_em_config: failed to add " | ||
577 | "offload em:%p on interface:%s\n", | ||
578 | fcoe->oem, fcoe->netdev->name); | ||
579 | return -ENOMEM; | ||
580 | } | ||
581 | } else { | ||
582 | fcoe->oem = fc_exch_mgr_alloc(lp, FC_CLASS_3, | ||
583 | FCOE_MIN_XID, lp->lro_xid, | ||
584 | fcoe_oem_match); | ||
585 | if (!fcoe->oem) { | ||
586 | printk(KERN_ERR "fcoe_em_config: failed to allocate " | ||
587 | "em for offload exches on interface:%s\n", | ||
588 | fcoe->netdev->name); | ||
589 | return -ENOMEM; | ||
590 | } | ||
591 | } | ||
592 | |||
593 | /* | ||
594 | * Exclude offload EM xid range from next EM xid range. | ||
595 | */ | ||
596 | min_xid += lp->lro_xid + 1; | ||
597 | |||
598 | skip_oem: | ||
599 | if (!fc_exch_mgr_alloc(lp, FC_CLASS_3, min_xid, max_xid, NULL)) { | ||
600 | printk(KERN_ERR "fcoe_em_config: failed to " | ||
601 | "allocate em on interface %s\n", fcoe->netdev->name); | ||
431 | return -ENOMEM; | 602 | return -ENOMEM; |
603 | } | ||
432 | 604 | ||
433 | return 0; | 605 | return 0; |
434 | } | 606 | } |
435 | 607 | ||
436 | /** | 608 | /** |
437 | * fcoe_if_destroy() - FCoE software HBA tear-down function | 609 | * fcoe_if_destroy() - FCoE software HBA tear-down function |
438 | * @netdev: ptr to the associated net_device | 610 | * @lport: fc_lport to destroy |
439 | * | ||
440 | * Returns: 0 if link is OK for use by FCoE. | ||
441 | */ | 611 | */ |
442 | static int fcoe_if_destroy(struct net_device *netdev) | 612 | static void fcoe_if_destroy(struct fc_lport *lport) |
443 | { | 613 | { |
444 | struct fc_lport *lp = NULL; | 614 | struct fcoe_port *port = lport_priv(lport); |
445 | struct fcoe_softc *fc; | 615 | struct fcoe_interface *fcoe = port->fcoe; |
446 | 616 | struct net_device *netdev = fcoe->netdev; | |
447 | BUG_ON(!netdev); | ||
448 | 617 | ||
449 | FCOE_NETDEV_DBG(netdev, "Destroying interface\n"); | 618 | FCOE_NETDEV_DBG(netdev, "Destroying interface\n"); |
450 | 619 | ||
451 | lp = fcoe_hostlist_lookup(netdev); | ||
452 | if (!lp) | ||
453 | return -ENODEV; | ||
454 | |||
455 | fc = lport_priv(lp); | ||
456 | |||
457 | /* Logout of the fabric */ | 620 | /* Logout of the fabric */ |
458 | fc_fabric_logoff(lp); | 621 | fc_fabric_logoff(lport); |
459 | 622 | ||
460 | /* Remove the instance from fcoe's list */ | 623 | /* Cleanup the fc_lport */ |
461 | fcoe_hostlist_remove(lp); | 624 | fc_lport_destroy(lport); |
625 | fc_fcp_destroy(lport); | ||
462 | 626 | ||
463 | /* clean up netdev configurations */ | 627 | /* Stop the transmit retry timer */ |
464 | fcoe_netdev_cleanup(fc); | 628 | del_timer_sync(&port->timer); |
465 | 629 | ||
466 | /* tear-down the FCoE controller */ | 630 | /* Free existing transmit skbs */ |
467 | fcoe_ctlr_destroy(&fc->ctlr); | 631 | fcoe_clean_pending_queue(lport); |
468 | 632 | ||
469 | /* Cleanup the fc_lport */ | 633 | /* receives may not be stopped until after this */ |
470 | fc_lport_destroy(lp); | 634 | fcoe_interface_put(fcoe); |
471 | fc_fcp_destroy(lp); | 635 | |
636 | /* Free queued packets for the per-CPU receive threads */ | ||
637 | fcoe_percpu_clean(lport); | ||
472 | 638 | ||
473 | /* Detach from the scsi-ml */ | 639 | /* Detach from the scsi-ml */ |
474 | fc_remove_host(lp->host); | 640 | fc_remove_host(lport->host); |
475 | scsi_remove_host(lp->host); | 641 | scsi_remove_host(lport->host); |
476 | 642 | ||
477 | /* There are no more rports or I/O, free the EM */ | 643 | /* There are no more rports or I/O, free the EM */ |
478 | if (lp->emp) | 644 | fc_exch_mgr_free(lport); |
479 | fc_exch_mgr_free(lp->emp); | ||
480 | |||
481 | /* Free the per-CPU receive threads */ | ||
482 | fcoe_percpu_clean(lp); | ||
483 | |||
484 | /* Free existing skbs */ | ||
485 | fcoe_clean_pending_queue(lp); | ||
486 | |||
487 | /* Stop the timer */ | ||
488 | del_timer_sync(&fc->timer); | ||
489 | 645 | ||
490 | /* Free memory used by statistical counters */ | 646 | /* Free memory used by statistical counters */ |
491 | fc_lport_free_stats(lp); | 647 | fc_lport_free_stats(lport); |
492 | |||
493 | /* Release the net_device and Scsi_Host */ | ||
494 | dev_put(fc->real_dev); | ||
495 | scsi_host_put(lp->host); | ||
496 | 648 | ||
497 | return 0; | 649 | /* Release the Scsi_Host */ |
650 | scsi_host_put(lport->host); | ||
498 | } | 651 | } |
499 | 652 | ||
500 | /* | 653 | /* |
@@ -540,106 +693,96 @@ static struct libfc_function_template fcoe_libfc_fcn_templ = { | |||
540 | }; | 693 | }; |
541 | 694 | ||
542 | /** | 695 | /** |
543 | * fcoe_if_create() - this function creates the fcoe interface | 696 | * fcoe_if_create() - this function creates the fcoe port |
544 | * @netdev: pointer the associated netdevice | 697 | * @fcoe: fcoe_interface structure to create an fc_lport instance on |
698 | * @parent: device pointer to be the parent in sysfs for the SCSI host | ||
545 | * | 699 | * |
546 | * Creates fc_lport struct and scsi_host for lport, configures lport | 700 | * Creates fc_lport struct and scsi_host for lport, configures lport. |
547 | * and starts fabric login. | ||
548 | * | 701 | * |
549 | * Returns : 0 on success | 702 | * Returns : The allocated fc_lport or an error pointer |
550 | */ | 703 | */ |
551 | static int fcoe_if_create(struct net_device *netdev) | 704 | static struct fc_lport *fcoe_if_create(struct fcoe_interface *fcoe, |
705 | struct device *parent) | ||
552 | { | 706 | { |
553 | int rc; | 707 | int rc; |
554 | struct fc_lport *lp = NULL; | 708 | struct fc_lport *lport = NULL; |
555 | struct fcoe_softc *fc; | 709 | struct fcoe_port *port; |
556 | struct Scsi_Host *shost; | 710 | struct Scsi_Host *shost; |
557 | 711 | struct net_device *netdev = fcoe->netdev; | |
558 | BUG_ON(!netdev); | ||
559 | 712 | ||
560 | FCOE_NETDEV_DBG(netdev, "Create Interface\n"); | 713 | FCOE_NETDEV_DBG(netdev, "Create Interface\n"); |
561 | 714 | ||
562 | lp = fcoe_hostlist_lookup(netdev); | ||
563 | if (lp) | ||
564 | return -EEXIST; | ||
565 | |||
566 | shost = libfc_host_alloc(&fcoe_shost_template, | 715 | shost = libfc_host_alloc(&fcoe_shost_template, |
567 | sizeof(struct fcoe_softc)); | 716 | sizeof(struct fcoe_port)); |
568 | if (!shost) { | 717 | if (!shost) { |
569 | FCOE_NETDEV_DBG(netdev, "Could not allocate host structure\n"); | 718 | FCOE_NETDEV_DBG(netdev, "Could not allocate host structure\n"); |
570 | return -ENOMEM; | 719 | rc = -ENOMEM; |
720 | goto out; | ||
571 | } | 721 | } |
572 | lp = shost_priv(shost); | 722 | lport = shost_priv(shost); |
573 | fc = lport_priv(lp); | 723 | port = lport_priv(lport); |
724 | port->lport = lport; | ||
725 | port->fcoe = fcoe; | ||
726 | INIT_WORK(&port->destroy_work, fcoe_destroy_work); | ||
574 | 727 | ||
575 | /* configure fc_lport, e.g., em */ | 728 | /* configure fc_lport, e.g., em */ |
576 | rc = fcoe_lport_config(lp); | 729 | rc = fcoe_lport_config(lport); |
577 | if (rc) { | 730 | if (rc) { |
578 | FCOE_NETDEV_DBG(netdev, "Could not configure lport for the " | 731 | FCOE_NETDEV_DBG(netdev, "Could not configure lport for the " |
579 | "interface\n"); | 732 | "interface\n"); |
580 | goto out_host_put; | 733 | goto out_host_put; |
581 | } | 734 | } |
582 | 735 | ||
583 | /* | ||
584 | * Initialize FIP. | ||
585 | */ | ||
586 | fcoe_ctlr_init(&fc->ctlr); | ||
587 | fc->ctlr.send = fcoe_fip_send; | ||
588 | fc->ctlr.update_mac = fcoe_update_src_mac; | ||
589 | |||
590 | /* configure lport network properties */ | 736 | /* configure lport network properties */ |
591 | rc = fcoe_netdev_config(lp, netdev); | 737 | rc = fcoe_netdev_config(lport, netdev); |
592 | if (rc) { | 738 | if (rc) { |
593 | FCOE_NETDEV_DBG(netdev, "Could not configure netdev for the " | 739 | FCOE_NETDEV_DBG(netdev, "Could not configure netdev for the " |
594 | "interface\n"); | 740 | "interface\n"); |
595 | goto out_netdev_cleanup; | 741 | goto out_lp_destroy; |
596 | } | 742 | } |
597 | 743 | ||
598 | /* configure lport scsi host properties */ | 744 | /* configure lport scsi host properties */ |
599 | rc = fcoe_shost_config(lp, shost, &netdev->dev); | 745 | rc = fcoe_shost_config(lport, shost, parent); |
600 | if (rc) { | 746 | if (rc) { |
601 | FCOE_NETDEV_DBG(netdev, "Could not configure shost for the " | 747 | FCOE_NETDEV_DBG(netdev, "Could not configure shost for the " |
602 | "interface\n"); | 748 | "interface\n"); |
603 | goto out_netdev_cleanup; | 749 | goto out_lp_destroy; |
604 | } | ||
605 | |||
606 | /* lport exch manager allocation */ | ||
607 | rc = fcoe_em_config(lp); | ||
608 | if (rc) { | ||
609 | FCOE_NETDEV_DBG(netdev, "Could not configure the EM for the " | ||
610 | "interface\n"); | ||
611 | goto out_netdev_cleanup; | ||
612 | } | 750 | } |
613 | 751 | ||
614 | /* Initialize the library */ | 752 | /* Initialize the library */ |
615 | rc = fcoe_libfc_config(lp, &fcoe_libfc_fcn_templ); | 753 | rc = fcoe_libfc_config(lport, &fcoe_libfc_fcn_templ); |
616 | if (rc) { | 754 | if (rc) { |
617 | FCOE_NETDEV_DBG(netdev, "Could not configure libfc for the " | 755 | FCOE_NETDEV_DBG(netdev, "Could not configure libfc for the " |
618 | "interface\n"); | 756 | "interface\n"); |
619 | goto out_lp_destroy; | 757 | goto out_lp_destroy; |
620 | } | 758 | } |
621 | 759 | ||
622 | /* add to lports list */ | 760 | /* |
623 | fcoe_hostlist_add(lp); | 761 | * fcoe_em_alloc() and fcoe_hostlist_add() both |
624 | 762 | * need to be atomic with respect to other changes to the hostlist | |
625 | lp->boot_time = jiffies; | 763 | * since fcoe_em_alloc() looks for an existing EM |
626 | 764 | * instance on host list updated by fcoe_hostlist_add(). | |
627 | fc_fabric_login(lp); | 765 | * |
628 | 766 | * This is currently handled through the fcoe_config_mutex begin held. | |
629 | if (!fcoe_link_ok(lp)) | 767 | */ |
630 | fcoe_ctlr_link_up(&fc->ctlr); | ||
631 | 768 | ||
632 | dev_hold(netdev); | 769 | /* lport exch manager allocation */ |
770 | rc = fcoe_em_config(lport); | ||
771 | if (rc) { | ||
772 | FCOE_NETDEV_DBG(netdev, "Could not configure the EM for the " | ||
773 | "interface\n"); | ||
774 | goto out_lp_destroy; | ||
775 | } | ||
633 | 776 | ||
634 | return rc; | 777 | fcoe_interface_get(fcoe); |
778 | return lport; | ||
635 | 779 | ||
636 | out_lp_destroy: | 780 | out_lp_destroy: |
637 | fc_exch_mgr_free(lp->emp); /* Free the EM */ | 781 | fc_exch_mgr_free(lport); |
638 | out_netdev_cleanup: | ||
639 | fcoe_netdev_cleanup(fc); | ||
640 | out_host_put: | 782 | out_host_put: |
641 | scsi_host_put(lp->host); | 783 | scsi_host_put(lport->host); |
642 | return rc; | 784 | out: |
785 | return ERR_PTR(rc); | ||
643 | } | 786 | } |
644 | 787 | ||
645 | /** | 788 | /** |
@@ -669,6 +812,7 @@ static int __init fcoe_if_init(void) | |||
669 | int __exit fcoe_if_exit(void) | 812 | int __exit fcoe_if_exit(void) |
670 | { | 813 | { |
671 | fc_release_transport(scsi_transport_fcoe_sw); | 814 | fc_release_transport(scsi_transport_fcoe_sw); |
815 | scsi_transport_fcoe_sw = NULL; | ||
672 | return 0; | 816 | return 0; |
673 | } | 817 | } |
674 | 818 | ||
@@ -686,7 +830,7 @@ static void fcoe_percpu_thread_create(unsigned int cpu) | |||
686 | thread = kthread_create(fcoe_percpu_receive_thread, | 830 | thread = kthread_create(fcoe_percpu_receive_thread, |
687 | (void *)p, "fcoethread/%d", cpu); | 831 | (void *)p, "fcoethread/%d", cpu); |
688 | 832 | ||
689 | if (likely(!IS_ERR(p->thread))) { | 833 | if (likely(!IS_ERR(thread))) { |
690 | kthread_bind(thread, cpu); | 834 | kthread_bind(thread, cpu); |
691 | wake_up_process(thread); | 835 | wake_up_process(thread); |
692 | 836 | ||
@@ -838,14 +982,13 @@ int fcoe_rcv(struct sk_buff *skb, struct net_device *dev, | |||
838 | { | 982 | { |
839 | struct fc_lport *lp; | 983 | struct fc_lport *lp; |
840 | struct fcoe_rcv_info *fr; | 984 | struct fcoe_rcv_info *fr; |
841 | struct fcoe_softc *fc; | 985 | struct fcoe_interface *fcoe; |
842 | struct fc_frame_header *fh; | 986 | struct fc_frame_header *fh; |
843 | struct fcoe_percpu_s *fps; | 987 | struct fcoe_percpu_s *fps; |
844 | unsigned short oxid; | 988 | unsigned int cpu; |
845 | unsigned int cpu = 0; | ||
846 | 989 | ||
847 | fc = container_of(ptype, struct fcoe_softc, fcoe_packet_type); | 990 | fcoe = container_of(ptype, struct fcoe_interface, fcoe_packet_type); |
848 | lp = fc->ctlr.lp; | 991 | lp = fcoe->ctlr.lp; |
849 | if (unlikely(lp == NULL)) { | 992 | if (unlikely(lp == NULL)) { |
850 | FCOE_NETDEV_DBG(dev, "Cannot find hba structure"); | 993 | FCOE_NETDEV_DBG(dev, "Cannot find hba structure"); |
851 | goto err2; | 994 | goto err2; |
@@ -876,20 +1019,20 @@ int fcoe_rcv(struct sk_buff *skb, struct net_device *dev, | |||
876 | skb_set_transport_header(skb, sizeof(struct fcoe_hdr)); | 1019 | skb_set_transport_header(skb, sizeof(struct fcoe_hdr)); |
877 | fh = (struct fc_frame_header *) skb_transport_header(skb); | 1020 | fh = (struct fc_frame_header *) skb_transport_header(skb); |
878 | 1021 | ||
879 | oxid = ntohs(fh->fh_ox_id); | ||
880 | |||
881 | fr = fcoe_dev_from_skb(skb); | 1022 | fr = fcoe_dev_from_skb(skb); |
882 | fr->fr_dev = lp; | 1023 | fr->fr_dev = lp; |
883 | fr->ptype = ptype; | 1024 | fr->ptype = ptype; |
884 | 1025 | ||
885 | #ifdef CONFIG_SMP | ||
886 | /* | 1026 | /* |
887 | * The incoming frame exchange id(oxid) is ANDed with num of online | 1027 | * In case the incoming frame's exchange is originated from |
888 | * cpu bits to get cpu and then this cpu is used for selecting | 1028 | * the initiator, then received frame's exchange id is ANDed |
889 | * a per cpu kernel thread from fcoe_percpu. | 1029 | * with fc_cpu_mask bits to get the same cpu on which exchange |
1030 | * was originated, otherwise just use the current cpu. | ||
890 | */ | 1031 | */ |
891 | cpu = oxid & (num_online_cpus() - 1); | 1032 | if (ntoh24(fh->fh_f_ctl) & FC_FC_EX_CTX) |
892 | #endif | 1033 | cpu = ntohs(fh->fh_ox_id) & fc_cpu_mask; |
1034 | else | ||
1035 | cpu = smp_processor_id(); | ||
893 | 1036 | ||
894 | fps = &per_cpu(fcoe_percpu, cpu); | 1037 | fps = &per_cpu(fcoe_percpu, cpu); |
895 | spin_lock_bh(&fps->fcoe_rx_list.lock); | 1038 | spin_lock_bh(&fps->fcoe_rx_list.lock); |
@@ -996,7 +1139,7 @@ static int fcoe_get_paged_crc_eof(struct sk_buff *skb, int tlen) | |||
996 | * fcoe_fc_crc() - calculates FC CRC in this fcoe skb | 1139 | * fcoe_fc_crc() - calculates FC CRC in this fcoe skb |
997 | * @fp: the fc_frame containing data to be checksummed | 1140 | * @fp: the fc_frame containing data to be checksummed |
998 | * | 1141 | * |
999 | * This uses crc32() to calculate the crc for fc frame | 1142 | * This uses crc32() to calculate the crc for port frame |
1000 | * Return : 32 bit crc | 1143 | * Return : 32 bit crc |
1001 | */ | 1144 | */ |
1002 | u32 fcoe_fc_crc(struct fc_frame *fp) | 1145 | u32 fcoe_fc_crc(struct fc_frame *fp) |
@@ -1029,7 +1172,7 @@ u32 fcoe_fc_crc(struct fc_frame *fp) | |||
1029 | 1172 | ||
1030 | /** | 1173 | /** |
1031 | * fcoe_xmit() - FCoE frame transmit function | 1174 | * fcoe_xmit() - FCoE frame transmit function |
1032 | * @lp: the associated local port | 1175 | * @lp: the associated local fcoe |
1033 | * @fp: the fc_frame to be transmitted | 1176 | * @fp: the fc_frame to be transmitted |
1034 | * | 1177 | * |
1035 | * Return : 0 for success | 1178 | * Return : 0 for success |
@@ -1046,13 +1189,13 @@ int fcoe_xmit(struct fc_lport *lp, struct fc_frame *fp) | |||
1046 | unsigned int hlen; /* header length implies the version */ | 1189 | unsigned int hlen; /* header length implies the version */ |
1047 | unsigned int tlen; /* trailer length */ | 1190 | unsigned int tlen; /* trailer length */ |
1048 | unsigned int elen; /* eth header, may include vlan */ | 1191 | unsigned int elen; /* eth header, may include vlan */ |
1049 | struct fcoe_softc *fc; | 1192 | struct fcoe_port *port = lport_priv(lp); |
1193 | struct fcoe_interface *fcoe = port->fcoe; | ||
1050 | u8 sof, eof; | 1194 | u8 sof, eof; |
1051 | struct fcoe_hdr *hp; | 1195 | struct fcoe_hdr *hp; |
1052 | 1196 | ||
1053 | WARN_ON((fr_len(fp) % sizeof(u32)) != 0); | 1197 | WARN_ON((fr_len(fp) % sizeof(u32)) != 0); |
1054 | 1198 | ||
1055 | fc = lport_priv(lp); | ||
1056 | fh = fc_frame_header_get(fp); | 1199 | fh = fc_frame_header_get(fp); |
1057 | skb = fp_skb(fp); | 1200 | skb = fp_skb(fp); |
1058 | wlen = skb->len / FCOE_WORD_TO_BYTE; | 1201 | wlen = skb->len / FCOE_WORD_TO_BYTE; |
@@ -1063,7 +1206,7 @@ int fcoe_xmit(struct fc_lport *lp, struct fc_frame *fp) | |||
1063 | } | 1206 | } |
1064 | 1207 | ||
1065 | if (unlikely(fh->fh_r_ctl == FC_RCTL_ELS_REQ) && | 1208 | if (unlikely(fh->fh_r_ctl == FC_RCTL_ELS_REQ) && |
1066 | fcoe_ctlr_els_send(&fc->ctlr, skb)) | 1209 | fcoe_ctlr_els_send(&fcoe->ctlr, skb)) |
1067 | return 0; | 1210 | return 0; |
1068 | 1211 | ||
1069 | sof = fr_sof(fp); | 1212 | sof = fr_sof(fp); |
@@ -1085,7 +1228,7 @@ int fcoe_xmit(struct fc_lport *lp, struct fc_frame *fp) | |||
1085 | crc = fcoe_fc_crc(fp); | 1228 | crc = fcoe_fc_crc(fp); |
1086 | } | 1229 | } |
1087 | 1230 | ||
1088 | /* copy fc crc and eof to the skb buff */ | 1231 | /* copy port crc and eof to the skb buff */ |
1089 | if (skb_is_nonlinear(skb)) { | 1232 | if (skb_is_nonlinear(skb)) { |
1090 | skb_frag_t *frag; | 1233 | skb_frag_t *frag; |
1091 | if (fcoe_get_paged_crc_eof(skb, tlen)) { | 1234 | if (fcoe_get_paged_crc_eof(skb, tlen)) { |
@@ -1108,27 +1251,27 @@ int fcoe_xmit(struct fc_lport *lp, struct fc_frame *fp) | |||
1108 | cp = NULL; | 1251 | cp = NULL; |
1109 | } | 1252 | } |
1110 | 1253 | ||
1111 | /* adjust skb network/transport offsets to match mac/fcoe/fc */ | 1254 | /* adjust skb network/transport offsets to match mac/fcoe/port */ |
1112 | skb_push(skb, elen + hlen); | 1255 | skb_push(skb, elen + hlen); |
1113 | skb_reset_mac_header(skb); | 1256 | skb_reset_mac_header(skb); |
1114 | skb_reset_network_header(skb); | 1257 | skb_reset_network_header(skb); |
1115 | skb->mac_len = elen; | 1258 | skb->mac_len = elen; |
1116 | skb->protocol = htons(ETH_P_FCOE); | 1259 | skb->protocol = htons(ETH_P_FCOE); |
1117 | skb->dev = fc->real_dev; | 1260 | skb->dev = fcoe->netdev; |
1118 | 1261 | ||
1119 | /* fill up mac and fcoe headers */ | 1262 | /* fill up mac and fcoe headers */ |
1120 | eh = eth_hdr(skb); | 1263 | eh = eth_hdr(skb); |
1121 | eh->h_proto = htons(ETH_P_FCOE); | 1264 | eh->h_proto = htons(ETH_P_FCOE); |
1122 | if (fc->ctlr.map_dest) | 1265 | if (fcoe->ctlr.map_dest) |
1123 | fc_fcoe_set_mac(eh->h_dest, fh->fh_d_id); | 1266 | fc_fcoe_set_mac(eh->h_dest, fh->fh_d_id); |
1124 | else | 1267 | else |
1125 | /* insert GW address */ | 1268 | /* insert GW address */ |
1126 | memcpy(eh->h_dest, fc->ctlr.dest_addr, ETH_ALEN); | 1269 | memcpy(eh->h_dest, fcoe->ctlr.dest_addr, ETH_ALEN); |
1127 | 1270 | ||
1128 | if (unlikely(fc->ctlr.flogi_oxid != FC_XID_UNKNOWN)) | 1271 | if (unlikely(fcoe->ctlr.flogi_oxid != FC_XID_UNKNOWN)) |
1129 | memcpy(eh->h_source, fc->ctlr.ctl_src_addr, ETH_ALEN); | 1272 | memcpy(eh->h_source, fcoe->ctlr.ctl_src_addr, ETH_ALEN); |
1130 | else | 1273 | else |
1131 | memcpy(eh->h_source, fc->ctlr.data_src_addr, ETH_ALEN); | 1274 | memcpy(eh->h_source, fcoe->ctlr.data_src_addr, ETH_ALEN); |
1132 | 1275 | ||
1133 | hp = (struct fcoe_hdr *)(eh + 1); | 1276 | hp = (struct fcoe_hdr *)(eh + 1); |
1134 | memset(hp, 0, sizeof(*hp)); | 1277 | memset(hp, 0, sizeof(*hp)); |
@@ -1136,7 +1279,6 @@ int fcoe_xmit(struct fc_lport *lp, struct fc_frame *fp) | |||
1136 | FC_FCOE_ENCAPS_VER(hp, FC_FCOE_VER); | 1279 | FC_FCOE_ENCAPS_VER(hp, FC_FCOE_VER); |
1137 | hp->fcoe_sof = sof; | 1280 | hp->fcoe_sof = sof; |
1138 | 1281 | ||
1139 | #ifdef NETIF_F_FSO | ||
1140 | /* fcoe lso, mss is in max_payload which is non-zero for FCP data */ | 1282 | /* fcoe lso, mss is in max_payload which is non-zero for FCP data */ |
1141 | if (lp->seq_offload && fr_max_payload(fp)) { | 1283 | if (lp->seq_offload && fr_max_payload(fp)) { |
1142 | skb_shinfo(skb)->gso_type = SKB_GSO_FCOE; | 1284 | skb_shinfo(skb)->gso_type = SKB_GSO_FCOE; |
@@ -1145,7 +1287,6 @@ int fcoe_xmit(struct fc_lport *lp, struct fc_frame *fp) | |||
1145 | skb_shinfo(skb)->gso_type = 0; | 1287 | skb_shinfo(skb)->gso_type = 0; |
1146 | skb_shinfo(skb)->gso_size = 0; | 1288 | skb_shinfo(skb)->gso_size = 0; |
1147 | } | 1289 | } |
1148 | #endif | ||
1149 | /* update tx stats: regardless if LLD fails */ | 1290 | /* update tx stats: regardless if LLD fails */ |
1150 | stats = fc_lport_get_stats(lp); | 1291 | stats = fc_lport_get_stats(lp); |
1151 | stats->TxFrames++; | 1292 | stats->TxFrames++; |
@@ -1153,7 +1294,7 @@ int fcoe_xmit(struct fc_lport *lp, struct fc_frame *fp) | |||
1153 | 1294 | ||
1154 | /* send down to lld */ | 1295 | /* send down to lld */ |
1155 | fr_dev(fp) = lp; | 1296 | fr_dev(fp) = lp; |
1156 | if (fc->fcoe_pending_queue.qlen) | 1297 | if (port->fcoe_pending_queue.qlen) |
1157 | fcoe_check_wait_queue(lp, skb); | 1298 | fcoe_check_wait_queue(lp, skb); |
1158 | else if (fcoe_start_io(skb)) | 1299 | else if (fcoe_start_io(skb)) |
1159 | fcoe_check_wait_queue(lp, skb); | 1300 | fcoe_check_wait_queue(lp, skb); |
@@ -1162,6 +1303,15 @@ int fcoe_xmit(struct fc_lport *lp, struct fc_frame *fp) | |||
1162 | } | 1303 | } |
1163 | 1304 | ||
1164 | /** | 1305 | /** |
1306 | * fcoe_percpu_flush_done() - Indicate percpu queue flush completion. | ||
1307 | * @skb: the skb being completed. | ||
1308 | */ | ||
1309 | static void fcoe_percpu_flush_done(struct sk_buff *skb) | ||
1310 | { | ||
1311 | complete(&fcoe_flush_completion); | ||
1312 | } | ||
1313 | |||
1314 | /** | ||
1165 | * fcoe_percpu_receive_thread() - recv thread per cpu | 1315 | * fcoe_percpu_receive_thread() - recv thread per cpu |
1166 | * @arg: ptr to the fcoe per cpu struct | 1316 | * @arg: ptr to the fcoe per cpu struct |
1167 | * | 1317 | * |
@@ -1179,7 +1329,7 @@ int fcoe_percpu_receive_thread(void *arg) | |||
1179 | struct fcoe_crc_eof crc_eof; | 1329 | struct fcoe_crc_eof crc_eof; |
1180 | struct fc_frame *fp; | 1330 | struct fc_frame *fp; |
1181 | u8 *mac = NULL; | 1331 | u8 *mac = NULL; |
1182 | struct fcoe_softc *fc; | 1332 | struct fcoe_port *port; |
1183 | struct fcoe_hdr *hp; | 1333 | struct fcoe_hdr *hp; |
1184 | 1334 | ||
1185 | set_user_nice(current, -20); | 1335 | set_user_nice(current, -20); |
@@ -1200,7 +1350,8 @@ int fcoe_percpu_receive_thread(void *arg) | |||
1200 | fr = fcoe_dev_from_skb(skb); | 1350 | fr = fcoe_dev_from_skb(skb); |
1201 | lp = fr->fr_dev; | 1351 | lp = fr->fr_dev; |
1202 | if (unlikely(lp == NULL)) { | 1352 | if (unlikely(lp == NULL)) { |
1203 | FCOE_NETDEV_DBG(skb->dev, "Invalid HBA Structure"); | 1353 | if (skb->destructor != fcoe_percpu_flush_done) |
1354 | FCOE_NETDEV_DBG(skb->dev, "NULL lport in skb"); | ||
1204 | kfree_skb(skb); | 1355 | kfree_skb(skb); |
1205 | continue; | 1356 | continue; |
1206 | } | 1357 | } |
@@ -1215,7 +1366,7 @@ int fcoe_percpu_receive_thread(void *arg) | |||
1215 | /* | 1366 | /* |
1216 | * Save source MAC address before discarding header. | 1367 | * Save source MAC address before discarding header. |
1217 | */ | 1368 | */ |
1218 | fc = lport_priv(lp); | 1369 | port = lport_priv(lp); |
1219 | if (skb_is_nonlinear(skb)) | 1370 | if (skb_is_nonlinear(skb)) |
1220 | skb_linearize(skb); /* not ideal */ | 1371 | skb_linearize(skb); /* not ideal */ |
1221 | mac = eth_hdr(skb)->h_source; | 1372 | mac = eth_hdr(skb)->h_source; |
@@ -1277,7 +1428,7 @@ int fcoe_percpu_receive_thread(void *arg) | |||
1277 | fh = fc_frame_header_get(fp); | 1428 | fh = fc_frame_header_get(fp); |
1278 | if (fh->fh_r_ctl == FC_RCTL_DD_SOL_DATA && | 1429 | if (fh->fh_r_ctl == FC_RCTL_DD_SOL_DATA && |
1279 | fh->fh_type == FC_TYPE_FCP) { | 1430 | fh->fh_type == FC_TYPE_FCP) { |
1280 | fc_exch_recv(lp, lp->emp, fp); | 1431 | fc_exch_recv(lp, fp); |
1281 | continue; | 1432 | continue; |
1282 | } | 1433 | } |
1283 | if (fr_flags(fp) & FCPHF_CRC_UNCHECKED) { | 1434 | if (fr_flags(fp) & FCPHF_CRC_UNCHECKED) { |
@@ -1293,12 +1444,12 @@ int fcoe_percpu_receive_thread(void *arg) | |||
1293 | } | 1444 | } |
1294 | fr_flags(fp) &= ~FCPHF_CRC_UNCHECKED; | 1445 | fr_flags(fp) &= ~FCPHF_CRC_UNCHECKED; |
1295 | } | 1446 | } |
1296 | if (unlikely(fc->ctlr.flogi_oxid != FC_XID_UNKNOWN) && | 1447 | if (unlikely(port->fcoe->ctlr.flogi_oxid != FC_XID_UNKNOWN) && |
1297 | fcoe_ctlr_recv_flogi(&fc->ctlr, fp, mac)) { | 1448 | fcoe_ctlr_recv_flogi(&port->fcoe->ctlr, fp, mac)) { |
1298 | fc_frame_free(fp); | 1449 | fc_frame_free(fp); |
1299 | continue; | 1450 | continue; |
1300 | } | 1451 | } |
1301 | fc_exch_recv(lp, lp->emp, fp); | 1452 | fc_exch_recv(lp, fp); |
1302 | } | 1453 | } |
1303 | return 0; | 1454 | return 0; |
1304 | } | 1455 | } |
@@ -1318,46 +1469,46 @@ int fcoe_percpu_receive_thread(void *arg) | |||
1318 | */ | 1469 | */ |
1319 | static void fcoe_check_wait_queue(struct fc_lport *lp, struct sk_buff *skb) | 1470 | static void fcoe_check_wait_queue(struct fc_lport *lp, struct sk_buff *skb) |
1320 | { | 1471 | { |
1321 | struct fcoe_softc *fc = lport_priv(lp); | 1472 | struct fcoe_port *port = lport_priv(lp); |
1322 | int rc; | 1473 | int rc; |
1323 | 1474 | ||
1324 | spin_lock_bh(&fc->fcoe_pending_queue.lock); | 1475 | spin_lock_bh(&port->fcoe_pending_queue.lock); |
1325 | 1476 | ||
1326 | if (skb) | 1477 | if (skb) |
1327 | __skb_queue_tail(&fc->fcoe_pending_queue, skb); | 1478 | __skb_queue_tail(&port->fcoe_pending_queue, skb); |
1328 | 1479 | ||
1329 | if (fc->fcoe_pending_queue_active) | 1480 | if (port->fcoe_pending_queue_active) |
1330 | goto out; | 1481 | goto out; |
1331 | fc->fcoe_pending_queue_active = 1; | 1482 | port->fcoe_pending_queue_active = 1; |
1332 | 1483 | ||
1333 | while (fc->fcoe_pending_queue.qlen) { | 1484 | while (port->fcoe_pending_queue.qlen) { |
1334 | /* keep qlen > 0 until fcoe_start_io succeeds */ | 1485 | /* keep qlen > 0 until fcoe_start_io succeeds */ |
1335 | fc->fcoe_pending_queue.qlen++; | 1486 | port->fcoe_pending_queue.qlen++; |
1336 | skb = __skb_dequeue(&fc->fcoe_pending_queue); | 1487 | skb = __skb_dequeue(&port->fcoe_pending_queue); |
1337 | 1488 | ||
1338 | spin_unlock_bh(&fc->fcoe_pending_queue.lock); | 1489 | spin_unlock_bh(&port->fcoe_pending_queue.lock); |
1339 | rc = fcoe_start_io(skb); | 1490 | rc = fcoe_start_io(skb); |
1340 | spin_lock_bh(&fc->fcoe_pending_queue.lock); | 1491 | spin_lock_bh(&port->fcoe_pending_queue.lock); |
1341 | 1492 | ||
1342 | if (rc) { | 1493 | if (rc) { |
1343 | __skb_queue_head(&fc->fcoe_pending_queue, skb); | 1494 | __skb_queue_head(&port->fcoe_pending_queue, skb); |
1344 | /* undo temporary increment above */ | 1495 | /* undo temporary increment above */ |
1345 | fc->fcoe_pending_queue.qlen--; | 1496 | port->fcoe_pending_queue.qlen--; |
1346 | break; | 1497 | break; |
1347 | } | 1498 | } |
1348 | /* undo temporary increment above */ | 1499 | /* undo temporary increment above */ |
1349 | fc->fcoe_pending_queue.qlen--; | 1500 | port->fcoe_pending_queue.qlen--; |
1350 | } | 1501 | } |
1351 | 1502 | ||
1352 | if (fc->fcoe_pending_queue.qlen < FCOE_LOW_QUEUE_DEPTH) | 1503 | if (port->fcoe_pending_queue.qlen < FCOE_LOW_QUEUE_DEPTH) |
1353 | lp->qfull = 0; | 1504 | lp->qfull = 0; |
1354 | if (fc->fcoe_pending_queue.qlen && !timer_pending(&fc->timer)) | 1505 | if (port->fcoe_pending_queue.qlen && !timer_pending(&port->timer)) |
1355 | mod_timer(&fc->timer, jiffies + 2); | 1506 | mod_timer(&port->timer, jiffies + 2); |
1356 | fc->fcoe_pending_queue_active = 0; | 1507 | port->fcoe_pending_queue_active = 0; |
1357 | out: | 1508 | out: |
1358 | if (fc->fcoe_pending_queue.qlen > FCOE_MAX_QUEUE_DEPTH) | 1509 | if (port->fcoe_pending_queue.qlen > FCOE_MAX_QUEUE_DEPTH) |
1359 | lp->qfull = 1; | 1510 | lp->qfull = 1; |
1360 | spin_unlock_bh(&fc->fcoe_pending_queue.lock); | 1511 | spin_unlock_bh(&port->fcoe_pending_queue.lock); |
1361 | return; | 1512 | return; |
1362 | } | 1513 | } |
1363 | 1514 | ||
@@ -1391,21 +1542,20 @@ static int fcoe_device_notification(struct notifier_block *notifier, | |||
1391 | ulong event, void *ptr) | 1542 | ulong event, void *ptr) |
1392 | { | 1543 | { |
1393 | struct fc_lport *lp = NULL; | 1544 | struct fc_lport *lp = NULL; |
1394 | struct net_device *real_dev = ptr; | 1545 | struct net_device *netdev = ptr; |
1395 | struct fcoe_softc *fc; | 1546 | struct fcoe_interface *fcoe; |
1547 | struct fcoe_port *port; | ||
1396 | struct fcoe_dev_stats *stats; | 1548 | struct fcoe_dev_stats *stats; |
1397 | u32 link_possible = 1; | 1549 | u32 link_possible = 1; |
1398 | u32 mfs; | 1550 | u32 mfs; |
1399 | int rc = NOTIFY_OK; | 1551 | int rc = NOTIFY_OK; |
1400 | 1552 | ||
1401 | read_lock(&fcoe_hostlist_lock); | 1553 | list_for_each_entry(fcoe, &fcoe_hostlist, list) { |
1402 | list_for_each_entry(fc, &fcoe_hostlist, list) { | 1554 | if (fcoe->netdev == netdev) { |
1403 | if (fc->real_dev == real_dev) { | 1555 | lp = fcoe->ctlr.lp; |
1404 | lp = fc->ctlr.lp; | ||
1405 | break; | 1556 | break; |
1406 | } | 1557 | } |
1407 | } | 1558 | } |
1408 | read_unlock(&fcoe_hostlist_lock); | ||
1409 | if (lp == NULL) { | 1559 | if (lp == NULL) { |
1410 | rc = NOTIFY_DONE; | 1560 | rc = NOTIFY_DONE; |
1411 | goto out; | 1561 | goto out; |
@@ -1420,21 +1570,27 @@ static int fcoe_device_notification(struct notifier_block *notifier, | |||
1420 | case NETDEV_CHANGE: | 1570 | case NETDEV_CHANGE: |
1421 | break; | 1571 | break; |
1422 | case NETDEV_CHANGEMTU: | 1572 | case NETDEV_CHANGEMTU: |
1423 | mfs = fc->real_dev->mtu - | 1573 | mfs = netdev->mtu - (sizeof(struct fcoe_hdr) + |
1424 | (sizeof(struct fcoe_hdr) + | 1574 | sizeof(struct fcoe_crc_eof)); |
1425 | sizeof(struct fcoe_crc_eof)); | ||
1426 | if (mfs >= FC_MIN_MAX_FRAME) | 1575 | if (mfs >= FC_MIN_MAX_FRAME) |
1427 | fc_set_mfs(lp, mfs); | 1576 | fc_set_mfs(lp, mfs); |
1428 | break; | 1577 | break; |
1429 | case NETDEV_REGISTER: | 1578 | case NETDEV_REGISTER: |
1430 | break; | 1579 | break; |
1580 | case NETDEV_UNREGISTER: | ||
1581 | list_del(&fcoe->list); | ||
1582 | port = lport_priv(fcoe->ctlr.lp); | ||
1583 | fcoe_interface_cleanup(fcoe); | ||
1584 | schedule_work(&port->destroy_work); | ||
1585 | goto out; | ||
1586 | break; | ||
1431 | default: | 1587 | default: |
1432 | FCOE_NETDEV_DBG(real_dev, "Unknown event %ld " | 1588 | FCOE_NETDEV_DBG(netdev, "Unknown event %ld " |
1433 | "from netdev netlink\n", event); | 1589 | "from netdev netlink\n", event); |
1434 | } | 1590 | } |
1435 | if (link_possible && !fcoe_link_ok(lp)) | 1591 | if (link_possible && !fcoe_link_ok(lp)) |
1436 | fcoe_ctlr_link_up(&fc->ctlr); | 1592 | fcoe_ctlr_link_up(&fcoe->ctlr); |
1437 | else if (fcoe_ctlr_link_down(&fc->ctlr)) { | 1593 | else if (fcoe_ctlr_link_down(&fcoe->ctlr)) { |
1438 | stats = fc_lport_get_stats(lp); | 1594 | stats = fc_lport_get_stats(lp); |
1439 | stats->LinkFailureCount++; | 1595 | stats->LinkFailureCount++; |
1440 | fcoe_clean_pending_queue(lp); | 1596 | fcoe_clean_pending_queue(lp); |
@@ -1465,75 +1621,6 @@ static struct net_device *fcoe_if_to_netdev(const char *buffer) | |||
1465 | } | 1621 | } |
1466 | 1622 | ||
1467 | /** | 1623 | /** |
1468 | * fcoe_netdev_to_module_owner() - finds out the driver module of the netdev | ||
1469 | * @netdev: the target netdev | ||
1470 | * | ||
1471 | * Returns: ptr to the struct module, NULL for failure | ||
1472 | */ | ||
1473 | static struct module * | ||
1474 | fcoe_netdev_to_module_owner(const struct net_device *netdev) | ||
1475 | { | ||
1476 | struct device *dev; | ||
1477 | |||
1478 | if (!netdev) | ||
1479 | return NULL; | ||
1480 | |||
1481 | dev = netdev->dev.parent; | ||
1482 | if (!dev) | ||
1483 | return NULL; | ||
1484 | |||
1485 | if (!dev->driver) | ||
1486 | return NULL; | ||
1487 | |||
1488 | return dev->driver->owner; | ||
1489 | } | ||
1490 | |||
1491 | /** | ||
1492 | * fcoe_ethdrv_get() - Hold the Ethernet driver | ||
1493 | * @netdev: the target netdev | ||
1494 | * | ||
1495 | * Holds the Ethernet driver module by try_module_get() for | ||
1496 | * the corresponding netdev. | ||
1497 | * | ||
1498 | * Returns: 0 for success | ||
1499 | */ | ||
1500 | static int fcoe_ethdrv_get(const struct net_device *netdev) | ||
1501 | { | ||
1502 | struct module *owner; | ||
1503 | |||
1504 | owner = fcoe_netdev_to_module_owner(netdev); | ||
1505 | if (owner) { | ||
1506 | FCOE_NETDEV_DBG(netdev, "Hold driver module %s\n", | ||
1507 | module_name(owner)); | ||
1508 | return try_module_get(owner); | ||
1509 | } | ||
1510 | return -ENODEV; | ||
1511 | } | ||
1512 | |||
1513 | /** | ||
1514 | * fcoe_ethdrv_put() - Release the Ethernet driver | ||
1515 | * @netdev: the target netdev | ||
1516 | * | ||
1517 | * Releases the Ethernet driver module by module_put for | ||
1518 | * the corresponding netdev. | ||
1519 | * | ||
1520 | * Returns: 0 for success | ||
1521 | */ | ||
1522 | static int fcoe_ethdrv_put(const struct net_device *netdev) | ||
1523 | { | ||
1524 | struct module *owner; | ||
1525 | |||
1526 | owner = fcoe_netdev_to_module_owner(netdev); | ||
1527 | if (owner) { | ||
1528 | FCOE_NETDEV_DBG(netdev, "Release driver module %s\n", | ||
1529 | module_name(owner)); | ||
1530 | module_put(owner); | ||
1531 | return 0; | ||
1532 | } | ||
1533 | return -ENODEV; | ||
1534 | } | ||
1535 | |||
1536 | /** | ||
1537 | * fcoe_destroy() - handles the destroy from sysfs | 1624 | * fcoe_destroy() - handles the destroy from sysfs |
1538 | * @buffer: expected to be an eth if name | 1625 | * @buffer: expected to be an eth if name |
1539 | * @kp: associated kernel param | 1626 | * @kp: associated kernel param |
@@ -1542,34 +1629,57 @@ static int fcoe_ethdrv_put(const struct net_device *netdev) | |||
1542 | */ | 1629 | */ |
1543 | static int fcoe_destroy(const char *buffer, struct kernel_param *kp) | 1630 | static int fcoe_destroy(const char *buffer, struct kernel_param *kp) |
1544 | { | 1631 | { |
1545 | int rc; | 1632 | struct fcoe_interface *fcoe; |
1546 | struct net_device *netdev; | 1633 | struct net_device *netdev; |
1634 | int rc; | ||
1635 | |||
1636 | mutex_lock(&fcoe_config_mutex); | ||
1637 | #ifdef CONFIG_FCOE_MODULE | ||
1638 | /* | ||
1639 | * Make sure the module has been initialized, and is not about to be | ||
1640 | * removed. Module paramter sysfs files are writable before the | ||
1641 | * module_init function is called and after module_exit. | ||
1642 | */ | ||
1643 | if (THIS_MODULE->state != MODULE_STATE_LIVE) { | ||
1644 | rc = -ENODEV; | ||
1645 | goto out_nodev; | ||
1646 | } | ||
1647 | #endif | ||
1547 | 1648 | ||
1548 | netdev = fcoe_if_to_netdev(buffer); | 1649 | netdev = fcoe_if_to_netdev(buffer); |
1549 | if (!netdev) { | 1650 | if (!netdev) { |
1550 | rc = -ENODEV; | 1651 | rc = -ENODEV; |
1551 | goto out_nodev; | 1652 | goto out_nodev; |
1552 | } | 1653 | } |
1553 | /* look for existing lport */ | 1654 | |
1554 | if (!fcoe_hostlist_lookup(netdev)) { | 1655 | rtnl_lock(); |
1656 | fcoe = fcoe_hostlist_lookup_port(netdev); | ||
1657 | if (!fcoe) { | ||
1658 | rtnl_unlock(); | ||
1555 | rc = -ENODEV; | 1659 | rc = -ENODEV; |
1556 | goto out_putdev; | 1660 | goto out_putdev; |
1557 | } | 1661 | } |
1558 | rc = fcoe_if_destroy(netdev); | 1662 | list_del(&fcoe->list); |
1559 | if (rc) { | 1663 | fcoe_interface_cleanup(fcoe); |
1560 | printk(KERN_ERR "fcoe: Failed to destroy interface (%s)\n", | 1664 | rtnl_unlock(); |
1561 | netdev->name); | 1665 | fcoe_if_destroy(fcoe->ctlr.lp); |
1562 | rc = -EIO; | ||
1563 | goto out_putdev; | ||
1564 | } | ||
1565 | fcoe_ethdrv_put(netdev); | ||
1566 | rc = 0; | ||
1567 | out_putdev: | 1666 | out_putdev: |
1568 | dev_put(netdev); | 1667 | dev_put(netdev); |
1569 | out_nodev: | 1668 | out_nodev: |
1669 | mutex_unlock(&fcoe_config_mutex); | ||
1570 | return rc; | 1670 | return rc; |
1571 | } | 1671 | } |
1572 | 1672 | ||
1673 | static void fcoe_destroy_work(struct work_struct *work) | ||
1674 | { | ||
1675 | struct fcoe_port *port; | ||
1676 | |||
1677 | port = container_of(work, struct fcoe_port, destroy_work); | ||
1678 | mutex_lock(&fcoe_config_mutex); | ||
1679 | fcoe_if_destroy(port->lport); | ||
1680 | mutex_unlock(&fcoe_config_mutex); | ||
1681 | } | ||
1682 | |||
1573 | /** | 1683 | /** |
1574 | * fcoe_create() - Handles the create call from sysfs | 1684 | * fcoe_create() - Handles the create call from sysfs |
1575 | * @buffer: expected to be an eth if name | 1685 | * @buffer: expected to be an eth if name |
@@ -1580,41 +1690,84 @@ out_nodev: | |||
1580 | static int fcoe_create(const char *buffer, struct kernel_param *kp) | 1690 | static int fcoe_create(const char *buffer, struct kernel_param *kp) |
1581 | { | 1691 | { |
1582 | int rc; | 1692 | int rc; |
1693 | struct fcoe_interface *fcoe; | ||
1694 | struct fc_lport *lport; | ||
1583 | struct net_device *netdev; | 1695 | struct net_device *netdev; |
1584 | 1696 | ||
1697 | mutex_lock(&fcoe_config_mutex); | ||
1698 | #ifdef CONFIG_FCOE_MODULE | ||
1699 | /* | ||
1700 | * Make sure the module has been initialized, and is not about to be | ||
1701 | * removed. Module paramter sysfs files are writable before the | ||
1702 | * module_init function is called and after module_exit. | ||
1703 | */ | ||
1704 | if (THIS_MODULE->state != MODULE_STATE_LIVE) { | ||
1705 | rc = -ENODEV; | ||
1706 | goto out_nodev; | ||
1707 | } | ||
1708 | #endif | ||
1709 | |||
1710 | rtnl_lock(); | ||
1585 | netdev = fcoe_if_to_netdev(buffer); | 1711 | netdev = fcoe_if_to_netdev(buffer); |
1586 | if (!netdev) { | 1712 | if (!netdev) { |
1587 | rc = -ENODEV; | 1713 | rc = -ENODEV; |
1588 | goto out_nodev; | 1714 | goto out_nodev; |
1589 | } | 1715 | } |
1716 | |||
1590 | /* look for existing lport */ | 1717 | /* look for existing lport */ |
1591 | if (fcoe_hostlist_lookup(netdev)) { | 1718 | if (fcoe_hostlist_lookup(netdev)) { |
1592 | rc = -EEXIST; | 1719 | rc = -EEXIST; |
1593 | goto out_putdev; | 1720 | goto out_putdev; |
1594 | } | 1721 | } |
1595 | fcoe_ethdrv_get(netdev); | ||
1596 | 1722 | ||
1597 | rc = fcoe_if_create(netdev); | 1723 | fcoe = fcoe_interface_create(netdev); |
1598 | if (rc) { | 1724 | if (!fcoe) { |
1725 | rc = -ENOMEM; | ||
1726 | goto out_putdev; | ||
1727 | } | ||
1728 | |||
1729 | lport = fcoe_if_create(fcoe, &netdev->dev); | ||
1730 | if (IS_ERR(lport)) { | ||
1599 | printk(KERN_ERR "fcoe: Failed to create interface (%s)\n", | 1731 | printk(KERN_ERR "fcoe: Failed to create interface (%s)\n", |
1600 | netdev->name); | 1732 | netdev->name); |
1601 | fcoe_ethdrv_put(netdev); | ||
1602 | rc = -EIO; | 1733 | rc = -EIO; |
1603 | goto out_putdev; | 1734 | fcoe_interface_cleanup(fcoe); |
1735 | goto out_free; | ||
1604 | } | 1736 | } |
1737 | |||
1738 | /* Make this the "master" N_Port */ | ||
1739 | fcoe->ctlr.lp = lport; | ||
1740 | |||
1741 | /* add to lports list */ | ||
1742 | fcoe_hostlist_add(lport); | ||
1743 | |||
1744 | /* start FIP Discovery and FLOGI */ | ||
1745 | lport->boot_time = jiffies; | ||
1746 | fc_fabric_login(lport); | ||
1747 | if (!fcoe_link_ok(lport)) | ||
1748 | fcoe_ctlr_link_up(&fcoe->ctlr); | ||
1749 | |||
1605 | rc = 0; | 1750 | rc = 0; |
1751 | out_free: | ||
1752 | /* | ||
1753 | * Release from init in fcoe_interface_create(), on success lport | ||
1754 | * should be holding a reference taken in fcoe_if_create(). | ||
1755 | */ | ||
1756 | fcoe_interface_put(fcoe); | ||
1606 | out_putdev: | 1757 | out_putdev: |
1607 | dev_put(netdev); | 1758 | dev_put(netdev); |
1608 | out_nodev: | 1759 | out_nodev: |
1760 | rtnl_unlock(); | ||
1761 | mutex_unlock(&fcoe_config_mutex); | ||
1609 | return rc; | 1762 | return rc; |
1610 | } | 1763 | } |
1611 | 1764 | ||
1612 | module_param_call(create, fcoe_create, NULL, NULL, S_IWUSR); | 1765 | module_param_call(create, fcoe_create, NULL, NULL, S_IWUSR); |
1613 | __MODULE_PARM_TYPE(create, "string"); | 1766 | __MODULE_PARM_TYPE(create, "string"); |
1614 | MODULE_PARM_DESC(create, "Create fcoe port using net device passed in."); | 1767 | MODULE_PARM_DESC(create, "Create fcoe fcoe using net device passed in."); |
1615 | module_param_call(destroy, fcoe_destroy, NULL, NULL, S_IWUSR); | 1768 | module_param_call(destroy, fcoe_destroy, NULL, NULL, S_IWUSR); |
1616 | __MODULE_PARM_TYPE(destroy, "string"); | 1769 | __MODULE_PARM_TYPE(destroy, "string"); |
1617 | MODULE_PARM_DESC(destroy, "Destroy fcoe port"); | 1770 | MODULE_PARM_DESC(destroy, "Destroy fcoe fcoe"); |
1618 | 1771 | ||
1619 | /** | 1772 | /** |
1620 | * fcoe_link_ok() - Check if link is ok for the fc_lport | 1773 | * fcoe_link_ok() - Check if link is ok for the fc_lport |
@@ -1632,37 +1785,40 @@ MODULE_PARM_DESC(destroy, "Destroy fcoe port"); | |||
1632 | */ | 1785 | */ |
1633 | int fcoe_link_ok(struct fc_lport *lp) | 1786 | int fcoe_link_ok(struct fc_lport *lp) |
1634 | { | 1787 | { |
1635 | struct fcoe_softc *fc = lport_priv(lp); | 1788 | struct fcoe_port *port = lport_priv(lp); |
1636 | struct net_device *dev = fc->real_dev; | 1789 | struct net_device *dev = port->fcoe->netdev; |
1637 | struct ethtool_cmd ecmd = { ETHTOOL_GSET }; | 1790 | struct ethtool_cmd ecmd = { ETHTOOL_GSET }; |
1638 | int rc = 0; | ||
1639 | 1791 | ||
1640 | if ((dev->flags & IFF_UP) && netif_carrier_ok(dev)) { | 1792 | if ((dev->flags & IFF_UP) && netif_carrier_ok(dev) && |
1641 | dev = fc->phys_dev; | 1793 | (!dev_ethtool_get_settings(dev, &ecmd))) { |
1642 | if (dev->ethtool_ops->get_settings) { | 1794 | lp->link_supported_speeds &= |
1643 | dev->ethtool_ops->get_settings(dev, &ecmd); | 1795 | ~(FC_PORTSPEED_1GBIT | FC_PORTSPEED_10GBIT); |
1644 | lp->link_supported_speeds &= | 1796 | if (ecmd.supported & (SUPPORTED_1000baseT_Half | |
1645 | ~(FC_PORTSPEED_1GBIT | FC_PORTSPEED_10GBIT); | 1797 | SUPPORTED_1000baseT_Full)) |
1646 | if (ecmd.supported & (SUPPORTED_1000baseT_Half | | 1798 | lp->link_supported_speeds |= FC_PORTSPEED_1GBIT; |
1647 | SUPPORTED_1000baseT_Full)) | 1799 | if (ecmd.supported & SUPPORTED_10000baseT_Full) |
1648 | lp->link_supported_speeds |= FC_PORTSPEED_1GBIT; | 1800 | lp->link_supported_speeds |= |
1649 | if (ecmd.supported & SUPPORTED_10000baseT_Full) | 1801 | FC_PORTSPEED_10GBIT; |
1650 | lp->link_supported_speeds |= | 1802 | if (ecmd.speed == SPEED_1000) |
1651 | FC_PORTSPEED_10GBIT; | 1803 | lp->link_speed = FC_PORTSPEED_1GBIT; |
1652 | if (ecmd.speed == SPEED_1000) | 1804 | if (ecmd.speed == SPEED_10000) |
1653 | lp->link_speed = FC_PORTSPEED_1GBIT; | 1805 | lp->link_speed = FC_PORTSPEED_10GBIT; |
1654 | if (ecmd.speed == SPEED_10000) | ||
1655 | lp->link_speed = FC_PORTSPEED_10GBIT; | ||
1656 | } | ||
1657 | } else | ||
1658 | rc = -1; | ||
1659 | 1806 | ||
1660 | return rc; | 1807 | return 0; |
1808 | } | ||
1809 | return -1; | ||
1661 | } | 1810 | } |
1662 | 1811 | ||
1663 | /** | 1812 | /** |
1664 | * fcoe_percpu_clean() - Clear the pending skbs for an lport | 1813 | * fcoe_percpu_clean() - Clear the pending skbs for an lport |
1665 | * @lp: the fc_lport | 1814 | * @lp: the fc_lport |
1815 | * | ||
1816 | * Must be called with fcoe_create_mutex held to single-thread completion. | ||
1817 | * | ||
1818 | * This flushes the pending skbs by adding a new skb to each queue and | ||
1819 | * waiting until they are all freed. This assures us that not only are | ||
1820 | * there no packets that will be handled by the lport, but also that any | ||
1821 | * threads already handling packet have returned. | ||
1666 | */ | 1822 | */ |
1667 | void fcoe_percpu_clean(struct fc_lport *lp) | 1823 | void fcoe_percpu_clean(struct fc_lport *lp) |
1668 | { | 1824 | { |
@@ -1687,7 +1843,25 @@ void fcoe_percpu_clean(struct fc_lport *lp) | |||
1687 | kfree_skb(skb); | 1843 | kfree_skb(skb); |
1688 | } | 1844 | } |
1689 | } | 1845 | } |
1846 | |||
1847 | if (!pp->thread || !cpu_online(cpu)) { | ||
1848 | spin_unlock_bh(&pp->fcoe_rx_list.lock); | ||
1849 | continue; | ||
1850 | } | ||
1851 | |||
1852 | skb = dev_alloc_skb(0); | ||
1853 | if (!skb) { | ||
1854 | spin_unlock_bh(&pp->fcoe_rx_list.lock); | ||
1855 | continue; | ||
1856 | } | ||
1857 | skb->destructor = fcoe_percpu_flush_done; | ||
1858 | |||
1859 | __skb_queue_tail(&pp->fcoe_rx_list, skb); | ||
1860 | if (pp->fcoe_rx_list.qlen == 1) | ||
1861 | wake_up_process(pp->thread); | ||
1690 | spin_unlock_bh(&pp->fcoe_rx_list.lock); | 1862 | spin_unlock_bh(&pp->fcoe_rx_list.lock); |
1863 | |||
1864 | wait_for_completion(&fcoe_flush_completion); | ||
1691 | } | 1865 | } |
1692 | } | 1866 | } |
1693 | 1867 | ||
@@ -1699,16 +1873,16 @@ void fcoe_percpu_clean(struct fc_lport *lp) | |||
1699 | */ | 1873 | */ |
1700 | void fcoe_clean_pending_queue(struct fc_lport *lp) | 1874 | void fcoe_clean_pending_queue(struct fc_lport *lp) |
1701 | { | 1875 | { |
1702 | struct fcoe_softc *fc = lport_priv(lp); | 1876 | struct fcoe_port *port = lport_priv(lp); |
1703 | struct sk_buff *skb; | 1877 | struct sk_buff *skb; |
1704 | 1878 | ||
1705 | spin_lock_bh(&fc->fcoe_pending_queue.lock); | 1879 | spin_lock_bh(&port->fcoe_pending_queue.lock); |
1706 | while ((skb = __skb_dequeue(&fc->fcoe_pending_queue)) != NULL) { | 1880 | while ((skb = __skb_dequeue(&port->fcoe_pending_queue)) != NULL) { |
1707 | spin_unlock_bh(&fc->fcoe_pending_queue.lock); | 1881 | spin_unlock_bh(&port->fcoe_pending_queue.lock); |
1708 | kfree_skb(skb); | 1882 | kfree_skb(skb); |
1709 | spin_lock_bh(&fc->fcoe_pending_queue.lock); | 1883 | spin_lock_bh(&port->fcoe_pending_queue.lock); |
1710 | } | 1884 | } |
1711 | spin_unlock_bh(&fc->fcoe_pending_queue.lock); | 1885 | spin_unlock_bh(&port->fcoe_pending_queue.lock); |
1712 | } | 1886 | } |
1713 | 1887 | ||
1714 | /** | 1888 | /** |
@@ -1725,24 +1899,21 @@ int fcoe_reset(struct Scsi_Host *shost) | |||
1725 | } | 1899 | } |
1726 | 1900 | ||
1727 | /** | 1901 | /** |
1728 | * fcoe_hostlist_lookup_softc() - find the corresponding lport by a given device | 1902 | * fcoe_hostlist_lookup_port() - find the corresponding lport by a given device |
1729 | * @dev: this is currently ptr to net_device | 1903 | * @dev: this is currently ptr to net_device |
1730 | * | 1904 | * |
1731 | * Returns: NULL or the located fcoe_softc | 1905 | * Returns: NULL or the located fcoe_port |
1906 | * Locking: must be called with the RNL mutex held | ||
1732 | */ | 1907 | */ |
1733 | static struct fcoe_softc * | 1908 | static struct fcoe_interface * |
1734 | fcoe_hostlist_lookup_softc(const struct net_device *dev) | 1909 | fcoe_hostlist_lookup_port(const struct net_device *dev) |
1735 | { | 1910 | { |
1736 | struct fcoe_softc *fc; | 1911 | struct fcoe_interface *fcoe; |
1737 | 1912 | ||
1738 | read_lock(&fcoe_hostlist_lock); | 1913 | list_for_each_entry(fcoe, &fcoe_hostlist, list) { |
1739 | list_for_each_entry(fc, &fcoe_hostlist, list) { | 1914 | if (fcoe->netdev == dev) |
1740 | if (fc->real_dev == dev) { | 1915 | return fcoe; |
1741 | read_unlock(&fcoe_hostlist_lock); | ||
1742 | return fc; | ||
1743 | } | ||
1744 | } | 1916 | } |
1745 | read_unlock(&fcoe_hostlist_lock); | ||
1746 | return NULL; | 1917 | return NULL; |
1747 | } | 1918 | } |
1748 | 1919 | ||
@@ -1751,14 +1922,14 @@ fcoe_hostlist_lookup_softc(const struct net_device *dev) | |||
1751 | * @netdev: ptr to net_device | 1922 | * @netdev: ptr to net_device |
1752 | * | 1923 | * |
1753 | * Returns: 0 for success | 1924 | * Returns: 0 for success |
1925 | * Locking: must be called with the RTNL mutex held | ||
1754 | */ | 1926 | */ |
1755 | struct fc_lport *fcoe_hostlist_lookup(const struct net_device *netdev) | 1927 | static struct fc_lport *fcoe_hostlist_lookup(const struct net_device *netdev) |
1756 | { | 1928 | { |
1757 | struct fcoe_softc *fc; | 1929 | struct fcoe_interface *fcoe; |
1758 | |||
1759 | fc = fcoe_hostlist_lookup_softc(netdev); | ||
1760 | 1930 | ||
1761 | return (fc) ? fc->ctlr.lp : NULL; | 1931 | fcoe = fcoe_hostlist_lookup_port(netdev); |
1932 | return (fcoe) ? fcoe->ctlr.lp : NULL; | ||
1762 | } | 1933 | } |
1763 | 1934 | ||
1764 | /** | 1935 | /** |
@@ -1766,41 +1937,23 @@ struct fc_lport *fcoe_hostlist_lookup(const struct net_device *netdev) | |||
1766 | * @lp: ptr to the fc_lport to be added | 1937 | * @lp: ptr to the fc_lport to be added |
1767 | * | 1938 | * |
1768 | * Returns: 0 for success | 1939 | * Returns: 0 for success |
1940 | * Locking: must be called with the RTNL mutex held | ||
1769 | */ | 1941 | */ |
1770 | int fcoe_hostlist_add(const struct fc_lport *lp) | 1942 | static int fcoe_hostlist_add(const struct fc_lport *lport) |
1771 | { | 1943 | { |
1772 | struct fcoe_softc *fc; | 1944 | struct fcoe_interface *fcoe; |
1773 | 1945 | struct fcoe_port *port; | |
1774 | fc = fcoe_hostlist_lookup_softc(fcoe_netdev(lp)); | 1946 | |
1775 | if (!fc) { | 1947 | fcoe = fcoe_hostlist_lookup_port(fcoe_netdev(lport)); |
1776 | fc = lport_priv(lp); | 1948 | if (!fcoe) { |
1777 | write_lock_bh(&fcoe_hostlist_lock); | 1949 | port = lport_priv(lport); |
1778 | list_add_tail(&fc->list, &fcoe_hostlist); | 1950 | fcoe = port->fcoe; |
1779 | write_unlock_bh(&fcoe_hostlist_lock); | 1951 | list_add_tail(&fcoe->list, &fcoe_hostlist); |
1780 | } | 1952 | } |
1781 | return 0; | 1953 | return 0; |
1782 | } | 1954 | } |
1783 | 1955 | ||
1784 | /** | 1956 | /** |
1785 | * fcoe_hostlist_remove() - remove a lport from lports list | ||
1786 | * @lp: ptr to the fc_lport to be removed | ||
1787 | * | ||
1788 | * Returns: 0 for success | ||
1789 | */ | ||
1790 | int fcoe_hostlist_remove(const struct fc_lport *lp) | ||
1791 | { | ||
1792 | struct fcoe_softc *fc; | ||
1793 | |||
1794 | fc = fcoe_hostlist_lookup_softc(fcoe_netdev(lp)); | ||
1795 | BUG_ON(!fc); | ||
1796 | write_lock_bh(&fcoe_hostlist_lock); | ||
1797 | list_del(&fc->list); | ||
1798 | write_unlock_bh(&fcoe_hostlist_lock); | ||
1799 | |||
1800 | return 0; | ||
1801 | } | ||
1802 | |||
1803 | /** | ||
1804 | * fcoe_init() - fcoe module loading initialization | 1957 | * fcoe_init() - fcoe module loading initialization |
1805 | * | 1958 | * |
1806 | * Returns 0 on success, negative on failure | 1959 | * Returns 0 on success, negative on failure |
@@ -1811,8 +1964,7 @@ static int __init fcoe_init(void) | |||
1811 | int rc = 0; | 1964 | int rc = 0; |
1812 | struct fcoe_percpu_s *p; | 1965 | struct fcoe_percpu_s *p; |
1813 | 1966 | ||
1814 | INIT_LIST_HEAD(&fcoe_hostlist); | 1967 | mutex_lock(&fcoe_config_mutex); |
1815 | rwlock_init(&fcoe_hostlist_lock); | ||
1816 | 1968 | ||
1817 | for_each_possible_cpu(cpu) { | 1969 | for_each_possible_cpu(cpu) { |
1818 | p = &per_cpu(fcoe_percpu, cpu); | 1970 | p = &per_cpu(fcoe_percpu, cpu); |
@@ -1830,15 +1982,18 @@ static int __init fcoe_init(void) | |||
1830 | /* Setup link change notification */ | 1982 | /* Setup link change notification */ |
1831 | fcoe_dev_setup(); | 1983 | fcoe_dev_setup(); |
1832 | 1984 | ||
1833 | fcoe_if_init(); | 1985 | rc = fcoe_if_init(); |
1986 | if (rc) | ||
1987 | goto out_free; | ||
1834 | 1988 | ||
1989 | mutex_unlock(&fcoe_config_mutex); | ||
1835 | return 0; | 1990 | return 0; |
1836 | 1991 | ||
1837 | out_free: | 1992 | out_free: |
1838 | for_each_online_cpu(cpu) { | 1993 | for_each_online_cpu(cpu) { |
1839 | fcoe_percpu_thread_destroy(cpu); | 1994 | fcoe_percpu_thread_destroy(cpu); |
1840 | } | 1995 | } |
1841 | 1996 | mutex_unlock(&fcoe_config_mutex); | |
1842 | return rc; | 1997 | return rc; |
1843 | } | 1998 | } |
1844 | module_init(fcoe_init); | 1999 | module_init(fcoe_init); |
@@ -1851,21 +2006,36 @@ module_init(fcoe_init); | |||
1851 | static void __exit fcoe_exit(void) | 2006 | static void __exit fcoe_exit(void) |
1852 | { | 2007 | { |
1853 | unsigned int cpu; | 2008 | unsigned int cpu; |
1854 | struct fcoe_softc *fc, *tmp; | 2009 | struct fcoe_interface *fcoe, *tmp; |
2010 | struct fcoe_port *port; | ||
2011 | |||
2012 | mutex_lock(&fcoe_config_mutex); | ||
1855 | 2013 | ||
1856 | fcoe_dev_cleanup(); | 2014 | fcoe_dev_cleanup(); |
1857 | 2015 | ||
1858 | /* releases the associated fcoe hosts */ | 2016 | /* releases the associated fcoe hosts */ |
1859 | list_for_each_entry_safe(fc, tmp, &fcoe_hostlist, list) | 2017 | rtnl_lock(); |
1860 | fcoe_if_destroy(fc->real_dev); | 2018 | list_for_each_entry_safe(fcoe, tmp, &fcoe_hostlist, list) { |
2019 | list_del(&fcoe->list); | ||
2020 | port = lport_priv(fcoe->ctlr.lp); | ||
2021 | fcoe_interface_cleanup(fcoe); | ||
2022 | schedule_work(&port->destroy_work); | ||
2023 | } | ||
2024 | rtnl_unlock(); | ||
1861 | 2025 | ||
1862 | unregister_hotcpu_notifier(&fcoe_cpu_notifier); | 2026 | unregister_hotcpu_notifier(&fcoe_cpu_notifier); |
1863 | 2027 | ||
1864 | for_each_online_cpu(cpu) { | 2028 | for_each_online_cpu(cpu) |
1865 | fcoe_percpu_thread_destroy(cpu); | 2029 | fcoe_percpu_thread_destroy(cpu); |
1866 | } | ||
1867 | 2030 | ||
1868 | /* detach from scsi transport */ | 2031 | mutex_unlock(&fcoe_config_mutex); |
2032 | |||
2033 | /* flush any asyncronous interface destroys, | ||
2034 | * this should happen after the netdev notifier is unregistered */ | ||
2035 | flush_scheduled_work(); | ||
2036 | |||
2037 | /* detach from scsi transport | ||
2038 | * must happen after all destroys are done, therefor after the flush */ | ||
1869 | fcoe_if_exit(); | 2039 | fcoe_if_exit(); |
1870 | } | 2040 | } |
1871 | module_exit(fcoe_exit); | 2041 | module_exit(fcoe_exit); |
diff --git a/drivers/scsi/fcoe/fcoe.h b/drivers/scsi/fcoe/fcoe.h index 0d724fa0898f..ce7f60fb1bc0 100644 --- a/drivers/scsi/fcoe/fcoe.h +++ b/drivers/scsi/fcoe/fcoe.h | |||
@@ -37,8 +37,8 @@ | |||
37 | 37 | ||
38 | #define FCOE_MAX_OUTSTANDING_COMMANDS 1024 | 38 | #define FCOE_MAX_OUTSTANDING_COMMANDS 1024 |
39 | 39 | ||
40 | #define FCOE_MIN_XID 0x0001 /* the min xid supported by fcoe_sw */ | 40 | #define FCOE_MIN_XID 0x0000 /* the min xid supported by fcoe_sw */ |
41 | #define FCOE_MAX_XID 0x07ef /* the max xid supported by fcoe_sw */ | 41 | #define FCOE_MAX_XID 0x0FFF /* the max xid supported by fcoe_sw */ |
42 | 42 | ||
43 | unsigned int fcoe_debug_logging; | 43 | unsigned int fcoe_debug_logging; |
44 | module_param_named(debug_logging, fcoe_debug_logging, int, S_IRUGO|S_IWUSR); | 44 | module_param_named(debug_logging, fcoe_debug_logging, int, S_IRUGO|S_IWUSR); |
@@ -53,7 +53,7 @@ do { \ | |||
53 | do { \ | 53 | do { \ |
54 | CMD; \ | 54 | CMD; \ |
55 | } while (0); \ | 55 | } while (0); \ |
56 | } while (0); | 56 | } while (0) |
57 | 57 | ||
58 | #define FCOE_DBG(fmt, args...) \ | 58 | #define FCOE_DBG(fmt, args...) \ |
59 | FCOE_CHECK_LOGGING(FCOE_LOGGING, \ | 59 | FCOE_CHECK_LOGGING(FCOE_LOGGING, \ |
@@ -61,7 +61,7 @@ do { \ | |||
61 | 61 | ||
62 | #define FCOE_NETDEV_DBG(netdev, fmt, args...) \ | 62 | #define FCOE_NETDEV_DBG(netdev, fmt, args...) \ |
63 | FCOE_CHECK_LOGGING(FCOE_NETDEV_LOGGING, \ | 63 | FCOE_CHECK_LOGGING(FCOE_NETDEV_LOGGING, \ |
64 | printk(KERN_INFO "fcoe: %s" fmt, \ | 64 | printk(KERN_INFO "fcoe: %s: " fmt, \ |
65 | netdev->name, ##args);) | 65 | netdev->name, ##args);) |
66 | 66 | ||
67 | /* | 67 | /* |
@@ -75,26 +75,36 @@ struct fcoe_percpu_s { | |||
75 | }; | 75 | }; |
76 | 76 | ||
77 | /* | 77 | /* |
78 | * the fcoe sw transport private data | 78 | * an FCoE interface, 1:1 with netdev |
79 | */ | 79 | */ |
80 | struct fcoe_softc { | 80 | struct fcoe_interface { |
81 | struct list_head list; | 81 | struct list_head list; |
82 | struct net_device *real_dev; | 82 | struct net_device *netdev; |
83 | struct net_device *phys_dev; /* device with ethtool_ops */ | ||
84 | struct packet_type fcoe_packet_type; | 83 | struct packet_type fcoe_packet_type; |
85 | struct packet_type fip_packet_type; | 84 | struct packet_type fip_packet_type; |
85 | struct fcoe_ctlr ctlr; | ||
86 | struct fc_exch_mgr *oem; /* offload exchange manager */ | ||
87 | struct kref kref; | ||
88 | }; | ||
89 | |||
90 | /* | ||
91 | * the FCoE private structure that's allocated along with the | ||
92 | * Scsi_Host and libfc fc_lport structures | ||
93 | */ | ||
94 | struct fcoe_port { | ||
95 | struct fcoe_interface *fcoe; | ||
96 | struct fc_lport *lport; | ||
86 | struct sk_buff_head fcoe_pending_queue; | 97 | struct sk_buff_head fcoe_pending_queue; |
87 | u8 fcoe_pending_queue_active; | 98 | u8 fcoe_pending_queue_active; |
88 | struct timer_list timer; /* queue timer */ | 99 | struct timer_list timer; /* queue timer */ |
89 | struct fcoe_ctlr ctlr; | 100 | struct work_struct destroy_work; /* to prevent rtnl deadlocks */ |
90 | }; | 101 | }; |
91 | 102 | ||
92 | #define fcoe_from_ctlr(fc) container_of(fc, struct fcoe_softc, ctlr) | 103 | #define fcoe_from_ctlr(fip) container_of(fip, struct fcoe_interface, ctlr) |
93 | 104 | ||
94 | static inline struct net_device *fcoe_netdev( | 105 | static inline struct net_device *fcoe_netdev(const struct fc_lport *lp) |
95 | const struct fc_lport *lp) | ||
96 | { | 106 | { |
97 | return ((struct fcoe_softc *)lport_priv(lp))->real_dev; | 107 | return ((struct fcoe_port *)lport_priv(lp))->fcoe->netdev; |
98 | } | 108 | } |
99 | 109 | ||
100 | #endif /* _FCOE_H_ */ | 110 | #endif /* _FCOE_H_ */ |
diff --git a/drivers/scsi/fcoe/libfcoe.c b/drivers/scsi/fcoe/libfcoe.c index f544340d318b..62a4c2026072 100644 --- a/drivers/scsi/fcoe/libfcoe.c +++ b/drivers/scsi/fcoe/libfcoe.c | |||
@@ -69,7 +69,7 @@ do { \ | |||
69 | do { \ | 69 | do { \ |
70 | CMD; \ | 70 | CMD; \ |
71 | } while (0); \ | 71 | } while (0); \ |
72 | } while (0); | 72 | } while (0) |
73 | 73 | ||
74 | #define LIBFCOE_DBG(fmt, args...) \ | 74 | #define LIBFCOE_DBG(fmt, args...) \ |
75 | LIBFCOE_CHECK_LOGGING(LIBFCOE_LOGGING, \ | 75 | LIBFCOE_CHECK_LOGGING(LIBFCOE_LOGGING, \ |
@@ -148,13 +148,17 @@ static void fcoe_ctlr_reset_fcfs(struct fcoe_ctlr *fip) | |||
148 | */ | 148 | */ |
149 | void fcoe_ctlr_destroy(struct fcoe_ctlr *fip) | 149 | void fcoe_ctlr_destroy(struct fcoe_ctlr *fip) |
150 | { | 150 | { |
151 | flush_work(&fip->recv_work); | 151 | cancel_work_sync(&fip->recv_work); |
152 | spin_lock_bh(&fip->fip_recv_list.lock); | ||
153 | __skb_queue_purge(&fip->fip_recv_list); | ||
154 | spin_unlock_bh(&fip->fip_recv_list.lock); | ||
155 | |||
152 | spin_lock_bh(&fip->lock); | 156 | spin_lock_bh(&fip->lock); |
153 | fip->state = FIP_ST_DISABLED; | 157 | fip->state = FIP_ST_DISABLED; |
154 | fcoe_ctlr_reset_fcfs(fip); | 158 | fcoe_ctlr_reset_fcfs(fip); |
155 | spin_unlock_bh(&fip->lock); | 159 | spin_unlock_bh(&fip->lock); |
156 | del_timer_sync(&fip->timer); | 160 | del_timer_sync(&fip->timer); |
157 | flush_work(&fip->link_work); | 161 | cancel_work_sync(&fip->link_work); |
158 | } | 162 | } |
159 | EXPORT_SYMBOL(fcoe_ctlr_destroy); | 163 | EXPORT_SYMBOL(fcoe_ctlr_destroy); |
160 | 164 | ||
@@ -413,10 +417,18 @@ static int fcoe_ctlr_encaps(struct fcoe_ctlr *fip, | |||
413 | struct fip_mac_desc *mac; | 417 | struct fip_mac_desc *mac; |
414 | struct fcoe_fcf *fcf; | 418 | struct fcoe_fcf *fcf; |
415 | size_t dlen; | 419 | size_t dlen; |
420 | u16 fip_flags; | ||
416 | 421 | ||
417 | fcf = fip->sel_fcf; | 422 | fcf = fip->sel_fcf; |
418 | if (!fcf) | 423 | if (!fcf) |
419 | return -ENODEV; | 424 | return -ENODEV; |
425 | |||
426 | /* set flags according to both FCF and lport's capability on SPMA */ | ||
427 | fip_flags = fcf->flags; | ||
428 | fip_flags &= fip->spma ? FIP_FL_SPMA | FIP_FL_FPMA : FIP_FL_FPMA; | ||
429 | if (!fip_flags) | ||
430 | return -ENODEV; | ||
431 | |||
420 | dlen = sizeof(struct fip_encaps) + skb->len; /* len before push */ | 432 | dlen = sizeof(struct fip_encaps) + skb->len; /* len before push */ |
421 | cap = (struct fip_encaps_head *)skb_push(skb, sizeof(*cap)); | 433 | cap = (struct fip_encaps_head *)skb_push(skb, sizeof(*cap)); |
422 | 434 | ||
@@ -429,9 +441,7 @@ static int fcoe_ctlr_encaps(struct fcoe_ctlr *fip, | |||
429 | cap->fip.fip_op = htons(FIP_OP_LS); | 441 | cap->fip.fip_op = htons(FIP_OP_LS); |
430 | cap->fip.fip_subcode = FIP_SC_REQ; | 442 | cap->fip.fip_subcode = FIP_SC_REQ; |
431 | cap->fip.fip_dl_len = htons((dlen + sizeof(*mac)) / FIP_BPW); | 443 | cap->fip.fip_dl_len = htons((dlen + sizeof(*mac)) / FIP_BPW); |
432 | cap->fip.fip_flags = htons(FIP_FL_FPMA); | 444 | cap->fip.fip_flags = htons(fip_flags); |
433 | if (fip->spma) | ||
434 | cap->fip.fip_flags |= htons(FIP_FL_SPMA); | ||
435 | 445 | ||
436 | cap->encaps.fd_desc.fip_dtype = dtype; | 446 | cap->encaps.fd_desc.fip_dtype = dtype; |
437 | cap->encaps.fd_desc.fip_dlen = dlen / FIP_BPW; | 447 | cap->encaps.fd_desc.fip_dlen = dlen / FIP_BPW; |
@@ -879,7 +889,7 @@ static void fcoe_ctlr_recv_els(struct fcoe_ctlr *fip, struct sk_buff *skb) | |||
879 | stats->RxFrames++; | 889 | stats->RxFrames++; |
880 | stats->RxWords += skb->len / FIP_BPW; | 890 | stats->RxWords += skb->len / FIP_BPW; |
881 | 891 | ||
882 | fc_exch_recv(lp, lp->emp, fp); | 892 | fc_exch_recv(lp, fp); |
883 | return; | 893 | return; |
884 | 894 | ||
885 | len_err: | 895 | len_err: |
@@ -1104,7 +1114,6 @@ static void fcoe_ctlr_timeout(unsigned long arg) | |||
1104 | struct fcoe_fcf *sel; | 1114 | struct fcoe_fcf *sel; |
1105 | struct fcoe_fcf *fcf; | 1115 | struct fcoe_fcf *fcf; |
1106 | unsigned long next_timer = jiffies + msecs_to_jiffies(FIP_VN_KA_PERIOD); | 1116 | unsigned long next_timer = jiffies + msecs_to_jiffies(FIP_VN_KA_PERIOD); |
1107 | DECLARE_MAC_BUF(buf); | ||
1108 | u8 send_ctlr_ka; | 1117 | u8 send_ctlr_ka; |
1109 | u8 send_port_ka; | 1118 | u8 send_port_ka; |
1110 | 1119 | ||
@@ -1128,9 +1137,8 @@ static void fcoe_ctlr_timeout(unsigned long arg) | |||
1128 | fcf = sel; /* the old FCF may have been freed */ | 1137 | fcf = sel; /* the old FCF may have been freed */ |
1129 | if (sel) { | 1138 | if (sel) { |
1130 | printk(KERN_INFO "libfcoe: host%d: FIP selected " | 1139 | printk(KERN_INFO "libfcoe: host%d: FIP selected " |
1131 | "Fibre-Channel Forwarder MAC %s\n", | 1140 | "Fibre-Channel Forwarder MAC %pM\n", |
1132 | fip->lp->host->host_no, | 1141 | fip->lp->host->host_no, sel->fcf_mac); |
1133 | print_mac(buf, sel->fcf_mac)); | ||
1134 | memcpy(fip->dest_addr, sel->fcf_mac, ETH_ALEN); | 1142 | memcpy(fip->dest_addr, sel->fcf_mac, ETH_ALEN); |
1135 | fip->port_ka_time = jiffies + | 1143 | fip->port_ka_time = jiffies + |
1136 | msecs_to_jiffies(FIP_VN_KA_PERIOD); | 1144 | msecs_to_jiffies(FIP_VN_KA_PERIOD); |
diff --git a/drivers/scsi/fnic/fnic_fcs.c b/drivers/scsi/fnic/fnic_fcs.c index 07e6eedb83ce..50db3e36a619 100644 --- a/drivers/scsi/fnic/fnic_fcs.c +++ b/drivers/scsi/fnic/fnic_fcs.c | |||
@@ -115,7 +115,7 @@ void fnic_handle_frame(struct work_struct *work) | |||
115 | } | 115 | } |
116 | spin_unlock_irqrestore(&fnic->fnic_lock, flags); | 116 | spin_unlock_irqrestore(&fnic->fnic_lock, flags); |
117 | 117 | ||
118 | fc_exch_recv(lp, lp->emp, fp); | 118 | fc_exch_recv(lp, fp); |
119 | } | 119 | } |
120 | 120 | ||
121 | } | 121 | } |
diff --git a/drivers/scsi/fnic/fnic_main.c b/drivers/scsi/fnic/fnic_main.c index 2c266c01dc5a..71c7bbe26d05 100644 --- a/drivers/scsi/fnic/fnic_main.c +++ b/drivers/scsi/fnic/fnic_main.c | |||
@@ -671,14 +671,6 @@ static int __devinit fnic_probe(struct pci_dev *pdev, | |||
671 | lp->link_up = 0; | 671 | lp->link_up = 0; |
672 | lp->tt = fnic_transport_template; | 672 | lp->tt = fnic_transport_template; |
673 | 673 | ||
674 | lp->emp = fc_exch_mgr_alloc(lp, FC_CLASS_3, | ||
675 | FCPIO_HOST_EXCH_RANGE_START, | ||
676 | FCPIO_HOST_EXCH_RANGE_END); | ||
677 | if (!lp->emp) { | ||
678 | err = -ENOMEM; | ||
679 | goto err_out_remove_scsi_host; | ||
680 | } | ||
681 | |||
682 | lp->max_retry_count = fnic->config.flogi_retries; | 674 | lp->max_retry_count = fnic->config.flogi_retries; |
683 | lp->max_rport_retry_count = fnic->config.plogi_retries; | 675 | lp->max_rport_retry_count = fnic->config.plogi_retries; |
684 | lp->service_params = (FCP_SPPF_INIT_FCN | FCP_SPPF_RD_XRDY_DIS | | 676 | lp->service_params = (FCP_SPPF_INIT_FCN | FCP_SPPF_RD_XRDY_DIS | |
@@ -693,12 +685,18 @@ static int __devinit fnic_probe(struct pci_dev *pdev, | |||
693 | fc_set_wwnn(lp, fnic->config.node_wwn); | 685 | fc_set_wwnn(lp, fnic->config.node_wwn); |
694 | fc_set_wwpn(lp, fnic->config.port_wwn); | 686 | fc_set_wwpn(lp, fnic->config.port_wwn); |
695 | 687 | ||
696 | fc_exch_init(lp); | ||
697 | fc_lport_init(lp); | 688 | fc_lport_init(lp); |
689 | fc_exch_init(lp); | ||
698 | fc_elsct_init(lp); | 690 | fc_elsct_init(lp); |
699 | fc_rport_init(lp); | 691 | fc_rport_init(lp); |
700 | fc_disc_init(lp); | 692 | fc_disc_init(lp); |
701 | 693 | ||
694 | if (!fc_exch_mgr_alloc(lp, FC_CLASS_3, FCPIO_HOST_EXCH_RANGE_START, | ||
695 | FCPIO_HOST_EXCH_RANGE_END, NULL)) { | ||
696 | err = -ENOMEM; | ||
697 | goto err_out_remove_scsi_host; | ||
698 | } | ||
699 | |||
702 | fc_lport_config(lp); | 700 | fc_lport_config(lp); |
703 | 701 | ||
704 | if (fc_set_mfs(lp, fnic->config.maxdatafieldsize + | 702 | if (fc_set_mfs(lp, fnic->config.maxdatafieldsize + |
@@ -738,7 +736,7 @@ static int __devinit fnic_probe(struct pci_dev *pdev, | |||
738 | return 0; | 736 | return 0; |
739 | 737 | ||
740 | err_out_free_exch_mgr: | 738 | err_out_free_exch_mgr: |
741 | fc_exch_mgr_free(lp->emp); | 739 | fc_exch_mgr_free(lp); |
742 | err_out_remove_scsi_host: | 740 | err_out_remove_scsi_host: |
743 | fc_remove_host(fnic->lport->host); | 741 | fc_remove_host(fnic->lport->host); |
744 | scsi_remove_host(fnic->lport->host); | 742 | scsi_remove_host(fnic->lport->host); |
@@ -827,7 +825,7 @@ static void __devexit fnic_remove(struct pci_dev *pdev) | |||
827 | 825 | ||
828 | fc_remove_host(fnic->lport->host); | 826 | fc_remove_host(fnic->lport->host); |
829 | scsi_remove_host(fnic->lport->host); | 827 | scsi_remove_host(fnic->lport->host); |
830 | fc_exch_mgr_free(fnic->lport->emp); | 828 | fc_exch_mgr_free(fnic->lport); |
831 | vnic_dev_notify_unset(fnic->vdev); | 829 | vnic_dev_notify_unset(fnic->vdev); |
832 | fnic_free_vnic_resources(fnic); | 830 | fnic_free_vnic_resources(fnic); |
833 | fnic_free_intr(fnic); | 831 | fnic_free_intr(fnic); |
diff --git a/drivers/scsi/ibmvscsi/ibmvfc.c b/drivers/scsi/ibmvscsi/ibmvfc.c index 166d96450a0e..bb2c696c006a 100644 --- a/drivers/scsi/ibmvscsi/ibmvfc.c +++ b/drivers/scsi/ibmvscsi/ibmvfc.c | |||
@@ -4217,7 +4217,7 @@ static int ibmvfc_alloc_mem(struct ibmvfc_host *vhost) | |||
4217 | if (!vhost->trace) | 4217 | if (!vhost->trace) |
4218 | goto free_disc_buffer; | 4218 | goto free_disc_buffer; |
4219 | 4219 | ||
4220 | vhost->tgt_pool = mempool_create_kzalloc_pool(IBMVFC_TGT_MEMPOOL_SZ, | 4220 | vhost->tgt_pool = mempool_create_kmalloc_pool(IBMVFC_TGT_MEMPOOL_SZ, |
4221 | sizeof(struct ibmvfc_target)); | 4221 | sizeof(struct ibmvfc_target)); |
4222 | 4222 | ||
4223 | if (!vhost->tgt_pool) { | 4223 | if (!vhost->tgt_pool) { |
diff --git a/drivers/scsi/ipr.h b/drivers/scsi/ipr.h index 4b63dd6b1c81..163245a1c3e5 100644 --- a/drivers/scsi/ipr.h +++ b/drivers/scsi/ipr.h | |||
@@ -1199,7 +1199,7 @@ struct ipr_ioa_cfg { | |||
1199 | 1199 | ||
1200 | struct ata_host ata_host; | 1200 | struct ata_host ata_host; |
1201 | char ipr_cmd_label[8]; | 1201 | char ipr_cmd_label[8]; |
1202 | #define IPR_CMD_LABEL "ipr_cmnd" | 1202 | #define IPR_CMD_LABEL "ipr_cmd" |
1203 | struct ipr_cmnd *ipr_cmnd_list[IPR_NUM_CMD_BLKS]; | 1203 | struct ipr_cmnd *ipr_cmnd_list[IPR_NUM_CMD_BLKS]; |
1204 | u32 ipr_cmnd_list_dma[IPR_NUM_CMD_BLKS]; | 1204 | u32 ipr_cmnd_list_dma[IPR_NUM_CMD_BLKS]; |
1205 | }; | 1205 | }; |
diff --git a/drivers/scsi/iscsi_tcp.c b/drivers/scsi/iscsi_tcp.c index 518dbd91df85..2b1b834a098b 100644 --- a/drivers/scsi/iscsi_tcp.c +++ b/drivers/scsi/iscsi_tcp.c | |||
@@ -99,6 +99,27 @@ static int iscsi_sw_tcp_recv(read_descriptor_t *rd_desc, struct sk_buff *skb, | |||
99 | return total_consumed; | 99 | return total_consumed; |
100 | } | 100 | } |
101 | 101 | ||
102 | /** | ||
103 | * iscsi_sw_sk_state_check - check socket state | ||
104 | * @sk: socket | ||
105 | * | ||
106 | * If the socket is in CLOSE or CLOSE_WAIT we should | ||
107 | * not close the connection if there is still some | ||
108 | * data pending. | ||
109 | */ | ||
110 | static inline int iscsi_sw_sk_state_check(struct sock *sk) | ||
111 | { | ||
112 | struct iscsi_conn *conn = (struct iscsi_conn*)sk->sk_user_data; | ||
113 | |||
114 | if ((sk->sk_state == TCP_CLOSE_WAIT || sk->sk_state == TCP_CLOSE) && | ||
115 | !atomic_read(&sk->sk_rmem_alloc)) { | ||
116 | ISCSI_SW_TCP_DBG(conn, "TCP_CLOSE|TCP_CLOSE_WAIT\n"); | ||
117 | iscsi_conn_failure(conn, ISCSI_ERR_TCP_CONN_CLOSE); | ||
118 | return -ECONNRESET; | ||
119 | } | ||
120 | return 0; | ||
121 | } | ||
122 | |||
102 | static void iscsi_sw_tcp_data_ready(struct sock *sk, int flag) | 123 | static void iscsi_sw_tcp_data_ready(struct sock *sk, int flag) |
103 | { | 124 | { |
104 | struct iscsi_conn *conn = sk->sk_user_data; | 125 | struct iscsi_conn *conn = sk->sk_user_data; |
@@ -117,6 +138,8 @@ static void iscsi_sw_tcp_data_ready(struct sock *sk, int flag) | |||
117 | rd_desc.count = 1; | 138 | rd_desc.count = 1; |
118 | tcp_read_sock(sk, &rd_desc, iscsi_sw_tcp_recv); | 139 | tcp_read_sock(sk, &rd_desc, iscsi_sw_tcp_recv); |
119 | 140 | ||
141 | iscsi_sw_sk_state_check(sk); | ||
142 | |||
120 | read_unlock(&sk->sk_callback_lock); | 143 | read_unlock(&sk->sk_callback_lock); |
121 | 144 | ||
122 | /* If we had to (atomically) map a highmem page, | 145 | /* If we had to (atomically) map a highmem page, |
@@ -137,13 +160,7 @@ static void iscsi_sw_tcp_state_change(struct sock *sk) | |||
137 | conn = (struct iscsi_conn*)sk->sk_user_data; | 160 | conn = (struct iscsi_conn*)sk->sk_user_data; |
138 | session = conn->session; | 161 | session = conn->session; |
139 | 162 | ||
140 | if ((sk->sk_state == TCP_CLOSE_WAIT || | 163 | iscsi_sw_sk_state_check(sk); |
141 | sk->sk_state == TCP_CLOSE) && | ||
142 | !atomic_read(&sk->sk_rmem_alloc)) { | ||
143 | ISCSI_SW_TCP_DBG(conn, "iscsi_tcp_state_change: " | ||
144 | "TCP_CLOSE|TCP_CLOSE_WAIT\n"); | ||
145 | iscsi_conn_failure(conn, ISCSI_ERR_CONN_FAILED); | ||
146 | } | ||
147 | 164 | ||
148 | tcp_conn = conn->dd_data; | 165 | tcp_conn = conn->dd_data; |
149 | tcp_sw_conn = tcp_conn->dd_data; | 166 | tcp_sw_conn = tcp_conn->dd_data; |
diff --git a/drivers/scsi/libfc/fc_disc.c b/drivers/scsi/libfc/fc_disc.c index 6fabf66972b9..c48799e9dd8e 100644 --- a/drivers/scsi/libfc/fc_disc.c +++ b/drivers/scsi/libfc/fc_disc.c | |||
@@ -43,47 +43,14 @@ | |||
43 | #define FC_DISC_RETRY_LIMIT 3 /* max retries */ | 43 | #define FC_DISC_RETRY_LIMIT 3 /* max retries */ |
44 | #define FC_DISC_RETRY_DELAY 500UL /* (msecs) delay */ | 44 | #define FC_DISC_RETRY_DELAY 500UL /* (msecs) delay */ |
45 | 45 | ||
46 | #define FC_DISC_DELAY 3 | ||
47 | |||
48 | static void fc_disc_gpn_ft_req(struct fc_disc *); | 46 | static void fc_disc_gpn_ft_req(struct fc_disc *); |
49 | static void fc_disc_gpn_ft_resp(struct fc_seq *, struct fc_frame *, void *); | 47 | static void fc_disc_gpn_ft_resp(struct fc_seq *, struct fc_frame *, void *); |
50 | static int fc_disc_new_target(struct fc_disc *, struct fc_rport *, | 48 | static void fc_disc_done(struct fc_disc *, enum fc_disc_event); |
51 | struct fc_rport_identifiers *); | ||
52 | static void fc_disc_del_target(struct fc_disc *, struct fc_rport *); | ||
53 | static void fc_disc_done(struct fc_disc *); | ||
54 | static void fc_disc_timeout(struct work_struct *); | 49 | static void fc_disc_timeout(struct work_struct *); |
55 | static void fc_disc_single(struct fc_disc *, struct fc_disc_port *); | 50 | static int fc_disc_single(struct fc_lport *, struct fc_disc_port *); |
56 | static void fc_disc_restart(struct fc_disc *); | 51 | static void fc_disc_restart(struct fc_disc *); |
57 | 52 | ||
58 | /** | 53 | /** |
59 | * fc_disc_lookup_rport() - lookup a remote port by port_id | ||
60 | * @lport: Fibre Channel host port instance | ||
61 | * @port_id: remote port port_id to match | ||
62 | */ | ||
63 | struct fc_rport *fc_disc_lookup_rport(const struct fc_lport *lport, | ||
64 | u32 port_id) | ||
65 | { | ||
66 | const struct fc_disc *disc = &lport->disc; | ||
67 | struct fc_rport *rport, *found = NULL; | ||
68 | struct fc_rport_libfc_priv *rdata; | ||
69 | int disc_found = 0; | ||
70 | |||
71 | list_for_each_entry(rdata, &disc->rports, peers) { | ||
72 | rport = PRIV_TO_RPORT(rdata); | ||
73 | if (rport->port_id == port_id) { | ||
74 | disc_found = 1; | ||
75 | found = rport; | ||
76 | break; | ||
77 | } | ||
78 | } | ||
79 | |||
80 | if (!disc_found) | ||
81 | found = NULL; | ||
82 | |||
83 | return found; | ||
84 | } | ||
85 | |||
86 | /** | ||
87 | * fc_disc_stop_rports() - delete all the remote ports associated with the lport | 54 | * fc_disc_stop_rports() - delete all the remote ports associated with the lport |
88 | * @disc: The discovery job to stop rports on | 55 | * @disc: The discovery job to stop rports on |
89 | * | 56 | * |
@@ -93,70 +60,17 @@ struct fc_rport *fc_disc_lookup_rport(const struct fc_lport *lport, | |||
93 | void fc_disc_stop_rports(struct fc_disc *disc) | 60 | void fc_disc_stop_rports(struct fc_disc *disc) |
94 | { | 61 | { |
95 | struct fc_lport *lport; | 62 | struct fc_lport *lport; |
96 | struct fc_rport *rport; | 63 | struct fc_rport_priv *rdata, *next; |
97 | struct fc_rport_libfc_priv *rdata, *next; | ||
98 | 64 | ||
99 | lport = disc->lport; | 65 | lport = disc->lport; |
100 | 66 | ||
101 | mutex_lock(&disc->disc_mutex); | 67 | mutex_lock(&disc->disc_mutex); |
102 | list_for_each_entry_safe(rdata, next, &disc->rports, peers) { | 68 | list_for_each_entry_safe(rdata, next, &disc->rports, peers) |
103 | rport = PRIV_TO_RPORT(rdata); | 69 | lport->tt.rport_logoff(rdata); |
104 | list_del(&rdata->peers); | ||
105 | lport->tt.rport_logoff(rport); | ||
106 | } | ||
107 | |||
108 | list_for_each_entry_safe(rdata, next, &disc->rogue_rports, peers) { | ||
109 | rport = PRIV_TO_RPORT(rdata); | ||
110 | lport->tt.rport_logoff(rport); | ||
111 | } | ||
112 | |||
113 | mutex_unlock(&disc->disc_mutex); | 70 | mutex_unlock(&disc->disc_mutex); |
114 | } | 71 | } |
115 | 72 | ||
116 | /** | 73 | /** |
117 | * fc_disc_rport_callback() - Event handler for rport events | ||
118 | * @lport: The lport which is receiving the event | ||
119 | * @rport: The rport which the event has occured on | ||
120 | * @event: The event that occured | ||
121 | * | ||
122 | * Locking Note: The rport lock should not be held when calling | ||
123 | * this function. | ||
124 | */ | ||
125 | static void fc_disc_rport_callback(struct fc_lport *lport, | ||
126 | struct fc_rport *rport, | ||
127 | enum fc_rport_event event) | ||
128 | { | ||
129 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
130 | struct fc_disc *disc = &lport->disc; | ||
131 | |||
132 | FC_DISC_DBG(disc, "Received a %d event for port (%6x)\n", event, | ||
133 | rport->port_id); | ||
134 | |||
135 | switch (event) { | ||
136 | case RPORT_EV_CREATED: | ||
137 | if (disc) { | ||
138 | mutex_lock(&disc->disc_mutex); | ||
139 | list_add_tail(&rdata->peers, &disc->rports); | ||
140 | mutex_unlock(&disc->disc_mutex); | ||
141 | } | ||
142 | break; | ||
143 | case RPORT_EV_LOGO: | ||
144 | case RPORT_EV_FAILED: | ||
145 | case RPORT_EV_STOP: | ||
146 | mutex_lock(&disc->disc_mutex); | ||
147 | mutex_lock(&rdata->rp_mutex); | ||
148 | if (rdata->trans_state == FC_PORTSTATE_ROGUE) | ||
149 | list_del(&rdata->peers); | ||
150 | mutex_unlock(&rdata->rp_mutex); | ||
151 | mutex_unlock(&disc->disc_mutex); | ||
152 | break; | ||
153 | default: | ||
154 | break; | ||
155 | } | ||
156 | |||
157 | } | ||
158 | |||
159 | /** | ||
160 | * fc_disc_recv_rscn_req() - Handle Registered State Change Notification (RSCN) | 74 | * fc_disc_recv_rscn_req() - Handle Registered State Change Notification (RSCN) |
161 | * @sp: Current sequence of the RSCN exchange | 75 | * @sp: Current sequence of the RSCN exchange |
162 | * @fp: RSCN Frame | 76 | * @fp: RSCN Frame |
@@ -169,8 +83,6 @@ static void fc_disc_recv_rscn_req(struct fc_seq *sp, struct fc_frame *fp, | |||
169 | struct fc_disc *disc) | 83 | struct fc_disc *disc) |
170 | { | 84 | { |
171 | struct fc_lport *lport; | 85 | struct fc_lport *lport; |
172 | struct fc_rport *rport; | ||
173 | struct fc_rport_libfc_priv *rdata; | ||
174 | struct fc_els_rscn *rp; | 86 | struct fc_els_rscn *rp; |
175 | struct fc_els_rscn_page *pp; | 87 | struct fc_els_rscn_page *pp; |
176 | struct fc_seq_els_data rjt_data; | 88 | struct fc_seq_els_data rjt_data; |
@@ -224,10 +136,7 @@ static void fc_disc_recv_rscn_req(struct fc_seq *sp, struct fc_frame *fp, | |||
224 | break; | 136 | break; |
225 | } | 137 | } |
226 | dp->lp = lport; | 138 | dp->lp = lport; |
227 | dp->ids.port_id = ntoh24(pp->rscn_fid); | 139 | dp->port_id = ntoh24(pp->rscn_fid); |
228 | dp->ids.port_name = -1; | ||
229 | dp->ids.node_name = -1; | ||
230 | dp->ids.roles = FC_RPORT_ROLE_UNKNOWN; | ||
231 | list_add_tail(&dp->peers, &disc_ports); | 140 | list_add_tail(&dp->peers, &disc_ports); |
232 | break; | 141 | break; |
233 | case ELS_ADDR_FMT_AREA: | 142 | case ELS_ADDR_FMT_AREA: |
@@ -240,6 +149,19 @@ static void fc_disc_recv_rscn_req(struct fc_seq *sp, struct fc_frame *fp, | |||
240 | } | 149 | } |
241 | } | 150 | } |
242 | lport->tt.seq_els_rsp_send(sp, ELS_LS_ACC, NULL); | 151 | lport->tt.seq_els_rsp_send(sp, ELS_LS_ACC, NULL); |
152 | |||
153 | /* | ||
154 | * If not doing a complete rediscovery, do GPN_ID on | ||
155 | * the individual ports mentioned in the list. | ||
156 | * If any of these get an error, do a full rediscovery. | ||
157 | * In any case, go through the list and free the entries. | ||
158 | */ | ||
159 | list_for_each_entry_safe(dp, next, &disc_ports, peers) { | ||
160 | list_del(&dp->peers); | ||
161 | if (!redisc) | ||
162 | redisc = fc_disc_single(lport, dp); | ||
163 | kfree(dp); | ||
164 | } | ||
243 | if (redisc) { | 165 | if (redisc) { |
244 | FC_DISC_DBG(disc, "RSCN received: rediscovering\n"); | 166 | FC_DISC_DBG(disc, "RSCN received: rediscovering\n"); |
245 | fc_disc_restart(disc); | 167 | fc_disc_restart(disc); |
@@ -247,16 +169,6 @@ static void fc_disc_recv_rscn_req(struct fc_seq *sp, struct fc_frame *fp, | |||
247 | FC_DISC_DBG(disc, "RSCN received: not rediscovering. " | 169 | FC_DISC_DBG(disc, "RSCN received: not rediscovering. " |
248 | "redisc %d state %d in_prog %d\n", | 170 | "redisc %d state %d in_prog %d\n", |
249 | redisc, lport->state, disc->pending); | 171 | redisc, lport->state, disc->pending); |
250 | list_for_each_entry_safe(dp, next, &disc_ports, peers) { | ||
251 | list_del(&dp->peers); | ||
252 | rport = lport->tt.rport_lookup(lport, dp->ids.port_id); | ||
253 | if (rport) { | ||
254 | rdata = rport->dd_data; | ||
255 | list_del(&rdata->peers); | ||
256 | lport->tt.rport_logoff(rport); | ||
257 | } | ||
258 | fc_disc_single(disc, dp); | ||
259 | } | ||
260 | } | 172 | } |
261 | fc_frame_free(fp); | 173 | fc_frame_free(fp); |
262 | return; | 174 | return; |
@@ -308,35 +220,34 @@ static void fc_disc_recv_req(struct fc_seq *sp, struct fc_frame *fp, | |||
308 | */ | 220 | */ |
309 | static void fc_disc_restart(struct fc_disc *disc) | 221 | static void fc_disc_restart(struct fc_disc *disc) |
310 | { | 222 | { |
311 | struct fc_rport *rport; | 223 | if (!disc->disc_callback) |
312 | struct fc_rport_libfc_priv *rdata, *next; | 224 | return; |
313 | struct fc_lport *lport = disc->lport; | ||
314 | 225 | ||
315 | FC_DISC_DBG(disc, "Restarting discovery\n"); | 226 | FC_DISC_DBG(disc, "Restarting discovery\n"); |
316 | 227 | ||
317 | list_for_each_entry_safe(rdata, next, &disc->rports, peers) { | ||
318 | rport = PRIV_TO_RPORT(rdata); | ||
319 | list_del(&rdata->peers); | ||
320 | lport->tt.rport_logoff(rport); | ||
321 | } | ||
322 | |||
323 | disc->requested = 1; | 228 | disc->requested = 1; |
324 | if (!disc->pending) | 229 | if (disc->pending) |
325 | fc_disc_gpn_ft_req(disc); | 230 | return; |
231 | |||
232 | /* | ||
233 | * Advance disc_id. This is an arbitrary non-zero number that will | ||
234 | * match the value in the fc_rport_priv after discovery for all | ||
235 | * freshly-discovered remote ports. Avoid wrapping to zero. | ||
236 | */ | ||
237 | disc->disc_id = (disc->disc_id + 2) | 1; | ||
238 | disc->retry_count = 0; | ||
239 | fc_disc_gpn_ft_req(disc); | ||
326 | } | 240 | } |
327 | 241 | ||
328 | /** | 242 | /** |
329 | * fc_disc_start() - Fibre Channel Target discovery | 243 | * fc_disc_start() - Fibre Channel Target discovery |
330 | * @lport: FC local port | 244 | * @lport: FC local port |
331 | * | 245 | * @disc_callback: function to be called when discovery is complete |
332 | * Returns non-zero if discovery cannot be started. | ||
333 | */ | 246 | */ |
334 | static void fc_disc_start(void (*disc_callback)(struct fc_lport *, | 247 | static void fc_disc_start(void (*disc_callback)(struct fc_lport *, |
335 | enum fc_disc_event), | 248 | enum fc_disc_event), |
336 | struct fc_lport *lport) | 249 | struct fc_lport *lport) |
337 | { | 250 | { |
338 | struct fc_rport *rport; | ||
339 | struct fc_rport_identifiers ids; | ||
340 | struct fc_disc *disc = &lport->disc; | 251 | struct fc_disc *disc = &lport->disc; |
341 | 252 | ||
342 | /* | 253 | /* |
@@ -345,145 +256,47 @@ static void fc_disc_start(void (*disc_callback)(struct fc_lport *, | |||
345 | * and send the GPN_FT request. | 256 | * and send the GPN_FT request. |
346 | */ | 257 | */ |
347 | mutex_lock(&disc->disc_mutex); | 258 | mutex_lock(&disc->disc_mutex); |
348 | |||
349 | disc->disc_callback = disc_callback; | 259 | disc->disc_callback = disc_callback; |
350 | 260 | fc_disc_restart(disc); | |
351 | /* | ||
352 | * If not ready, or already running discovery, just set request flag. | ||
353 | */ | ||
354 | disc->requested = 1; | ||
355 | |||
356 | if (disc->pending) { | ||
357 | mutex_unlock(&disc->disc_mutex); | ||
358 | return; | ||
359 | } | ||
360 | |||
361 | /* | ||
362 | * Handle point-to-point mode as a simple discovery | ||
363 | * of the remote port. Yucky, yucky, yuck, yuck! | ||
364 | */ | ||
365 | rport = disc->lport->ptp_rp; | ||
366 | if (rport) { | ||
367 | ids.port_id = rport->port_id; | ||
368 | ids.port_name = rport->port_name; | ||
369 | ids.node_name = rport->node_name; | ||
370 | ids.roles = FC_RPORT_ROLE_UNKNOWN; | ||
371 | get_device(&rport->dev); | ||
372 | |||
373 | if (!fc_disc_new_target(disc, rport, &ids)) { | ||
374 | disc->event = DISC_EV_SUCCESS; | ||
375 | fc_disc_done(disc); | ||
376 | } | ||
377 | put_device(&rport->dev); | ||
378 | } else { | ||
379 | fc_disc_gpn_ft_req(disc); /* get ports by FC-4 type */ | ||
380 | } | ||
381 | |||
382 | mutex_unlock(&disc->disc_mutex); | 261 | mutex_unlock(&disc->disc_mutex); |
383 | } | 262 | } |
384 | 263 | ||
385 | static struct fc_rport_operations fc_disc_rport_ops = { | ||
386 | .event_callback = fc_disc_rport_callback, | ||
387 | }; | ||
388 | |||
389 | /** | ||
390 | * fc_disc_new_target() - Handle new target found by discovery | ||
391 | * @lport: FC local port | ||
392 | * @rport: The previous FC remote port (NULL if new remote port) | ||
393 | * @ids: Identifiers for the new FC remote port | ||
394 | * | ||
395 | * Locking Note: This function expects that the disc_mutex is locked | ||
396 | * before it is called. | ||
397 | */ | ||
398 | static int fc_disc_new_target(struct fc_disc *disc, | ||
399 | struct fc_rport *rport, | ||
400 | struct fc_rport_identifiers *ids) | ||
401 | { | ||
402 | struct fc_lport *lport = disc->lport; | ||
403 | struct fc_rport_libfc_priv *rdata; | ||
404 | int error = 0; | ||
405 | |||
406 | if (rport && ids->port_name) { | ||
407 | if (rport->port_name == -1) { | ||
408 | /* | ||
409 | * Set WWN and fall through to notify of create. | ||
410 | */ | ||
411 | fc_rport_set_name(rport, ids->port_name, | ||
412 | rport->node_name); | ||
413 | } else if (rport->port_name != ids->port_name) { | ||
414 | /* | ||
415 | * This is a new port with the same FCID as | ||
416 | * a previously-discovered port. Presumably the old | ||
417 | * port logged out and a new port logged in and was | ||
418 | * assigned the same FCID. This should be rare. | ||
419 | * Delete the old one and fall thru to re-create. | ||
420 | */ | ||
421 | fc_disc_del_target(disc, rport); | ||
422 | rport = NULL; | ||
423 | } | ||
424 | } | ||
425 | if (((ids->port_name != -1) || (ids->port_id != -1)) && | ||
426 | ids->port_id != fc_host_port_id(lport->host) && | ||
427 | ids->port_name != lport->wwpn) { | ||
428 | if (!rport) { | ||
429 | rport = lport->tt.rport_lookup(lport, ids->port_id); | ||
430 | if (!rport) { | ||
431 | struct fc_disc_port dp; | ||
432 | dp.lp = lport; | ||
433 | dp.ids.port_id = ids->port_id; | ||
434 | dp.ids.port_name = ids->port_name; | ||
435 | dp.ids.node_name = ids->node_name; | ||
436 | dp.ids.roles = ids->roles; | ||
437 | rport = lport->tt.rport_create(&dp); | ||
438 | } | ||
439 | if (!rport) | ||
440 | error = -ENOMEM; | ||
441 | } | ||
442 | if (rport) { | ||
443 | rdata = rport->dd_data; | ||
444 | rdata->ops = &fc_disc_rport_ops; | ||
445 | rdata->rp_state = RPORT_ST_INIT; | ||
446 | list_add_tail(&rdata->peers, &disc->rogue_rports); | ||
447 | lport->tt.rport_login(rport); | ||
448 | } | ||
449 | } | ||
450 | return error; | ||
451 | } | ||
452 | |||
453 | /** | ||
454 | * fc_disc_del_target() - Delete a target | ||
455 | * @disc: FC discovery context | ||
456 | * @rport: The remote port to be removed | ||
457 | */ | ||
458 | static void fc_disc_del_target(struct fc_disc *disc, struct fc_rport *rport) | ||
459 | { | ||
460 | struct fc_lport *lport = disc->lport; | ||
461 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
462 | list_del(&rdata->peers); | ||
463 | lport->tt.rport_logoff(rport); | ||
464 | } | ||
465 | |||
466 | /** | 264 | /** |
467 | * fc_disc_done() - Discovery has been completed | 265 | * fc_disc_done() - Discovery has been completed |
468 | * @disc: FC discovery context | 266 | * @disc: FC discovery context |
267 | * @event: discovery completion status | ||
268 | * | ||
469 | * Locking Note: This function expects that the disc mutex is locked before | 269 | * Locking Note: This function expects that the disc mutex is locked before |
470 | * it is called. The discovery callback is then made with the lock released, | 270 | * it is called. The discovery callback is then made with the lock released, |
471 | * and the lock is re-taken before returning from this function | 271 | * and the lock is re-taken before returning from this function |
472 | */ | 272 | */ |
473 | static void fc_disc_done(struct fc_disc *disc) | 273 | static void fc_disc_done(struct fc_disc *disc, enum fc_disc_event event) |
474 | { | 274 | { |
475 | struct fc_lport *lport = disc->lport; | 275 | struct fc_lport *lport = disc->lport; |
476 | enum fc_disc_event event; | 276 | struct fc_rport_priv *rdata; |
477 | 277 | ||
478 | FC_DISC_DBG(disc, "Discovery complete\n"); | 278 | FC_DISC_DBG(disc, "Discovery complete\n"); |
479 | 279 | ||
480 | event = disc->event; | 280 | disc->pending = 0; |
481 | disc->event = DISC_EV_NONE; | 281 | if (disc->requested) { |
282 | fc_disc_restart(disc); | ||
283 | return; | ||
284 | } | ||
482 | 285 | ||
483 | if (disc->requested) | 286 | /* |
484 | fc_disc_gpn_ft_req(disc); | 287 | * Go through all remote ports. If they were found in the latest |
485 | else | 288 | * discovery, reverify or log them in. Otherwise, log them out. |
486 | disc->pending = 0; | 289 | * Skip ports which were never discovered. These are the dNS port |
290 | * and ports which were created by PLOGI. | ||
291 | */ | ||
292 | list_for_each_entry(rdata, &disc->rports, peers) { | ||
293 | if (!rdata->disc_id) | ||
294 | continue; | ||
295 | if (rdata->disc_id == disc->disc_id) | ||
296 | lport->tt.rport_login(rdata); | ||
297 | else | ||
298 | lport->tt.rport_logoff(rdata); | ||
299 | } | ||
487 | 300 | ||
488 | mutex_unlock(&disc->disc_mutex); | 301 | mutex_unlock(&disc->disc_mutex); |
489 | disc->disc_callback(lport, event); | 302 | disc->disc_callback(lport, event); |
@@ -522,11 +335,8 @@ static void fc_disc_error(struct fc_disc *disc, struct fc_frame *fp) | |||
522 | } | 335 | } |
523 | disc->retry_count++; | 336 | disc->retry_count++; |
524 | schedule_delayed_work(&disc->disc_work, delay); | 337 | schedule_delayed_work(&disc->disc_work, delay); |
525 | } else { | 338 | } else |
526 | /* exceeded retries */ | 339 | fc_disc_done(disc, DISC_EV_FAILED); |
527 | disc->event = DISC_EV_FAILED; | ||
528 | fc_disc_done(disc); | ||
529 | } | ||
530 | } | 340 | } |
531 | } | 341 | } |
532 | 342 | ||
@@ -555,7 +365,7 @@ static void fc_disc_gpn_ft_req(struct fc_disc *disc) | |||
555 | if (!fp) | 365 | if (!fp) |
556 | goto err; | 366 | goto err; |
557 | 367 | ||
558 | if (lport->tt.elsct_send(lport, NULL, fp, | 368 | if (lport->tt.elsct_send(lport, 0, fp, |
559 | FC_NS_GPN_FT, | 369 | FC_NS_GPN_FT, |
560 | fc_disc_gpn_ft_resp, | 370 | fc_disc_gpn_ft_resp, |
561 | disc, lport->e_d_tov)) | 371 | disc, lport->e_d_tov)) |
@@ -565,10 +375,12 @@ err: | |||
565 | } | 375 | } |
566 | 376 | ||
567 | /** | 377 | /** |
568 | * fc_disc_gpn_ft_parse() - Parse the list of IDs and names resulting from a request | 378 | * fc_disc_gpn_ft_parse() - Parse the body of the dNS GPN_FT response. |
569 | * @lport: Fibre Channel host port instance | 379 | * @lport: Fibre Channel host port instance |
570 | * @buf: GPN_FT response buffer | 380 | * @buf: GPN_FT response buffer |
571 | * @len: size of response buffer | 381 | * @len: size of response buffer |
382 | * | ||
383 | * Goes through the list of IDs and names resulting from a request. | ||
572 | */ | 384 | */ |
573 | static int fc_disc_gpn_ft_parse(struct fc_disc *disc, void *buf, size_t len) | 385 | static int fc_disc_gpn_ft_parse(struct fc_disc *disc, void *buf, size_t len) |
574 | { | 386 | { |
@@ -578,11 +390,11 @@ static int fc_disc_gpn_ft_parse(struct fc_disc *disc, void *buf, size_t len) | |||
578 | size_t plen; | 390 | size_t plen; |
579 | size_t tlen; | 391 | size_t tlen; |
580 | int error = 0; | 392 | int error = 0; |
581 | struct fc_disc_port dp; | 393 | struct fc_rport_identifiers ids; |
582 | struct fc_rport *rport; | 394 | struct fc_rport_priv *rdata; |
583 | struct fc_rport_libfc_priv *rdata; | ||
584 | 395 | ||
585 | lport = disc->lport; | 396 | lport = disc->lport; |
397 | disc->seq_count++; | ||
586 | 398 | ||
587 | /* | 399 | /* |
588 | * Handle partial name record left over from previous call. | 400 | * Handle partial name record left over from previous call. |
@@ -591,6 +403,7 @@ static int fc_disc_gpn_ft_parse(struct fc_disc *disc, void *buf, size_t len) | |||
591 | plen = len; | 403 | plen = len; |
592 | np = (struct fc_gpn_ft_resp *)bp; | 404 | np = (struct fc_gpn_ft_resp *)bp; |
593 | tlen = disc->buf_len; | 405 | tlen = disc->buf_len; |
406 | disc->buf_len = 0; | ||
594 | if (tlen) { | 407 | if (tlen) { |
595 | WARN_ON(tlen >= sizeof(*np)); | 408 | WARN_ON(tlen >= sizeof(*np)); |
596 | plen = sizeof(*np) - tlen; | 409 | plen = sizeof(*np) - tlen; |
@@ -621,31 +434,25 @@ static int fc_disc_gpn_ft_parse(struct fc_disc *disc, void *buf, size_t len) | |||
621 | * After the first time through the loop, things return to "normal". | 434 | * After the first time through the loop, things return to "normal". |
622 | */ | 435 | */ |
623 | while (plen >= sizeof(*np)) { | 436 | while (plen >= sizeof(*np)) { |
624 | dp.lp = lport; | 437 | ids.port_id = ntoh24(np->fp_fid); |
625 | dp.ids.port_id = ntoh24(np->fp_fid); | 438 | ids.port_name = ntohll(np->fp_wwpn); |
626 | dp.ids.port_name = ntohll(np->fp_wwpn); | 439 | |
627 | dp.ids.node_name = -1; | 440 | if (ids.port_id != fc_host_port_id(lport->host) && |
628 | dp.ids.roles = FC_RPORT_ROLE_UNKNOWN; | 441 | ids.port_name != lport->wwpn) { |
629 | 442 | rdata = lport->tt.rport_create(lport, ids.port_id); | |
630 | if ((dp.ids.port_id != fc_host_port_id(lport->host)) && | 443 | if (rdata) { |
631 | (dp.ids.port_name != lport->wwpn)) { | 444 | rdata->ids.port_name = ids.port_name; |
632 | rport = lport->tt.rport_create(&dp); | 445 | rdata->disc_id = disc->disc_id; |
633 | if (rport) { | 446 | } else { |
634 | rdata = rport->dd_data; | ||
635 | rdata->ops = &fc_disc_rport_ops; | ||
636 | rdata->local_port = lport; | ||
637 | list_add_tail(&rdata->peers, | ||
638 | &disc->rogue_rports); | ||
639 | lport->tt.rport_login(rport); | ||
640 | } else | ||
641 | printk(KERN_WARNING "libfc: Failed to allocate " | 447 | printk(KERN_WARNING "libfc: Failed to allocate " |
642 | "memory for the newly discovered port " | 448 | "memory for the newly discovered port " |
643 | "(%6x)\n", dp.ids.port_id); | 449 | "(%6x)\n", ids.port_id); |
450 | error = -ENOMEM; | ||
451 | } | ||
644 | } | 452 | } |
645 | 453 | ||
646 | if (np->fp_flags & FC_NS_FID_LAST) { | 454 | if (np->fp_flags & FC_NS_FID_LAST) { |
647 | disc->event = DISC_EV_SUCCESS; | 455 | fc_disc_done(disc, DISC_EV_SUCCESS); |
648 | fc_disc_done(disc); | ||
649 | len = 0; | 456 | len = 0; |
650 | break; | 457 | break; |
651 | } | 458 | } |
@@ -665,8 +472,6 @@ static int fc_disc_gpn_ft_parse(struct fc_disc *disc, void *buf, size_t len) | |||
665 | memcpy(&disc->partial_buf, np, len); | 472 | memcpy(&disc->partial_buf, np, len); |
666 | } | 473 | } |
667 | disc->buf_len = (unsigned char) len; | 474 | disc->buf_len = (unsigned char) len; |
668 | } else { | ||
669 | disc->buf_len = 0; | ||
670 | } | 475 | } |
671 | return error; | 476 | return error; |
672 | } | 477 | } |
@@ -683,8 +488,7 @@ static void fc_disc_timeout(struct work_struct *work) | |||
683 | struct fc_disc, | 488 | struct fc_disc, |
684 | disc_work.work); | 489 | disc_work.work); |
685 | mutex_lock(&disc->disc_mutex); | 490 | mutex_lock(&disc->disc_mutex); |
686 | if (disc->requested && !disc->pending) | 491 | fc_disc_gpn_ft_req(disc); |
687 | fc_disc_gpn_ft_req(disc); | ||
688 | mutex_unlock(&disc->disc_mutex); | 492 | mutex_unlock(&disc->disc_mutex); |
689 | } | 493 | } |
690 | 494 | ||
@@ -703,10 +507,10 @@ static void fc_disc_gpn_ft_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
703 | struct fc_disc *disc = disc_arg; | 507 | struct fc_disc *disc = disc_arg; |
704 | struct fc_ct_hdr *cp; | 508 | struct fc_ct_hdr *cp; |
705 | struct fc_frame_header *fh; | 509 | struct fc_frame_header *fh; |
510 | enum fc_disc_event event = DISC_EV_NONE; | ||
706 | unsigned int seq_cnt; | 511 | unsigned int seq_cnt; |
707 | void *buf = NULL; | ||
708 | unsigned int len; | 512 | unsigned int len; |
709 | int error; | 513 | int error = 0; |
710 | 514 | ||
711 | mutex_lock(&disc->disc_mutex); | 515 | mutex_lock(&disc->disc_mutex); |
712 | FC_DISC_DBG(disc, "Received a GPN_FT response\n"); | 516 | FC_DISC_DBG(disc, "Received a GPN_FT response\n"); |
@@ -721,77 +525,158 @@ static void fc_disc_gpn_ft_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
721 | fh = fc_frame_header_get(fp); | 525 | fh = fc_frame_header_get(fp); |
722 | len = fr_len(fp) - sizeof(*fh); | 526 | len = fr_len(fp) - sizeof(*fh); |
723 | seq_cnt = ntohs(fh->fh_seq_cnt); | 527 | seq_cnt = ntohs(fh->fh_seq_cnt); |
724 | if (fr_sof(fp) == FC_SOF_I3 && seq_cnt == 0 && | 528 | if (fr_sof(fp) == FC_SOF_I3 && seq_cnt == 0 && disc->seq_count == 0) { |
725 | disc->seq_count == 0) { | ||
726 | cp = fc_frame_payload_get(fp, sizeof(*cp)); | 529 | cp = fc_frame_payload_get(fp, sizeof(*cp)); |
727 | if (!cp) { | 530 | if (!cp) { |
728 | FC_DISC_DBG(disc, "GPN_FT response too short, len %d\n", | 531 | FC_DISC_DBG(disc, "GPN_FT response too short, len %d\n", |
729 | fr_len(fp)); | 532 | fr_len(fp)); |
533 | event = DISC_EV_FAILED; | ||
730 | } else if (ntohs(cp->ct_cmd) == FC_FS_ACC) { | 534 | } else if (ntohs(cp->ct_cmd) == FC_FS_ACC) { |
731 | 535 | ||
732 | /* Accepted, parse the response. */ | 536 | /* Accepted, parse the response. */ |
733 | buf = cp + 1; | ||
734 | len -= sizeof(*cp); | 537 | len -= sizeof(*cp); |
538 | error = fc_disc_gpn_ft_parse(disc, cp + 1, len); | ||
735 | } else if (ntohs(cp->ct_cmd) == FC_FS_RJT) { | 539 | } else if (ntohs(cp->ct_cmd) == FC_FS_RJT) { |
736 | FC_DISC_DBG(disc, "GPN_FT rejected reason %x exp %x " | 540 | FC_DISC_DBG(disc, "GPN_FT rejected reason %x exp %x " |
737 | "(check zoning)\n", cp->ct_reason, | 541 | "(check zoning)\n", cp->ct_reason, |
738 | cp->ct_explan); | 542 | cp->ct_explan); |
739 | disc->event = DISC_EV_FAILED; | 543 | event = DISC_EV_FAILED; |
740 | fc_disc_done(disc); | 544 | if (cp->ct_reason == FC_FS_RJT_UNABL && |
545 | cp->ct_explan == FC_FS_EXP_FTNR) | ||
546 | event = DISC_EV_SUCCESS; | ||
741 | } else { | 547 | } else { |
742 | FC_DISC_DBG(disc, "GPN_FT unexpected response code " | 548 | FC_DISC_DBG(disc, "GPN_FT unexpected response code " |
743 | "%x\n", ntohs(cp->ct_cmd)); | 549 | "%x\n", ntohs(cp->ct_cmd)); |
550 | event = DISC_EV_FAILED; | ||
744 | } | 551 | } |
745 | } else if (fr_sof(fp) == FC_SOF_N3 && | 552 | } else if (fr_sof(fp) == FC_SOF_N3 && seq_cnt == disc->seq_count) { |
746 | seq_cnt == disc->seq_count) { | 553 | error = fc_disc_gpn_ft_parse(disc, fh + 1, len); |
747 | buf = fh + 1; | ||
748 | } else { | 554 | } else { |
749 | FC_DISC_DBG(disc, "GPN_FT unexpected frame - out of sequence? " | 555 | FC_DISC_DBG(disc, "GPN_FT unexpected frame - out of sequence? " |
750 | "seq_cnt %x expected %x sof %x eof %x\n", | 556 | "seq_cnt %x expected %x sof %x eof %x\n", |
751 | seq_cnt, disc->seq_count, fr_sof(fp), fr_eof(fp)); | 557 | seq_cnt, disc->seq_count, fr_sof(fp), fr_eof(fp)); |
558 | event = DISC_EV_FAILED; | ||
752 | } | 559 | } |
753 | if (buf) { | 560 | if (error) |
754 | error = fc_disc_gpn_ft_parse(disc, buf, len); | 561 | fc_disc_error(disc, fp); |
755 | if (error) | 562 | else if (event != DISC_EV_NONE) |
756 | fc_disc_error(disc, fp); | 563 | fc_disc_done(disc, event); |
757 | else | ||
758 | disc->seq_count++; | ||
759 | } | ||
760 | fc_frame_free(fp); | 564 | fc_frame_free(fp); |
761 | |||
762 | mutex_unlock(&disc->disc_mutex); | 565 | mutex_unlock(&disc->disc_mutex); |
763 | } | 566 | } |
764 | 567 | ||
765 | /** | 568 | /** |
766 | * fc_disc_single() - Discover the directory information for a single target | 569 | * fc_disc_gpn_id_resp() - Handle a response frame from Get Port Names (GPN_ID) |
767 | * @lport: FC local port | 570 | * @sp: exchange sequence |
768 | * @dp: The port to rediscover | 571 | * @fp: response frame |
572 | * @rdata_arg: remote port private data | ||
769 | * | 573 | * |
770 | * Locking Note: This function expects that the disc_mutex is locked | 574 | * Locking Note: This function is called without disc mutex held. |
771 | * before it is called. | ||
772 | */ | 575 | */ |
773 | static void fc_disc_single(struct fc_disc *disc, struct fc_disc_port *dp) | 576 | static void fc_disc_gpn_id_resp(struct fc_seq *sp, struct fc_frame *fp, |
577 | void *rdata_arg) | ||
774 | { | 578 | { |
579 | struct fc_rport_priv *rdata = rdata_arg; | ||
580 | struct fc_rport_priv *new_rdata; | ||
775 | struct fc_lport *lport; | 581 | struct fc_lport *lport; |
776 | struct fc_rport *new_rport; | 582 | struct fc_disc *disc; |
777 | struct fc_rport_libfc_priv *rdata; | 583 | struct fc_ct_hdr *cp; |
584 | struct fc_ns_gid_pn *pn; | ||
585 | u64 port_name; | ||
778 | 586 | ||
779 | lport = disc->lport; | 587 | lport = rdata->local_port; |
588 | disc = &lport->disc; | ||
780 | 589 | ||
781 | if (dp->ids.port_id == fc_host_port_id(lport->host)) | 590 | mutex_lock(&disc->disc_mutex); |
591 | if (PTR_ERR(fp) == -FC_EX_CLOSED) | ||
782 | goto out; | 592 | goto out; |
783 | 593 | if (IS_ERR(fp)) | |
784 | new_rport = lport->tt.rport_create(dp); | 594 | goto redisc; |
785 | if (new_rport) { | 595 | |
786 | rdata = new_rport->dd_data; | 596 | cp = fc_frame_payload_get(fp, sizeof(*cp)); |
787 | rdata->ops = &fc_disc_rport_ops; | 597 | if (!cp) |
788 | kfree(dp); | 598 | goto redisc; |
789 | list_add_tail(&rdata->peers, &disc->rogue_rports); | 599 | if (ntohs(cp->ct_cmd) == FC_FS_ACC) { |
790 | lport->tt.rport_login(new_rport); | 600 | if (fr_len(fp) < sizeof(struct fc_frame_header) + |
601 | sizeof(*cp) + sizeof(*pn)) | ||
602 | goto redisc; | ||
603 | pn = (struct fc_ns_gid_pn *)(cp + 1); | ||
604 | port_name = get_unaligned_be64(&pn->fn_wwpn); | ||
605 | if (rdata->ids.port_name == -1) | ||
606 | rdata->ids.port_name = port_name; | ||
607 | else if (rdata->ids.port_name != port_name) { | ||
608 | FC_DISC_DBG(disc, "GPN_ID accepted. WWPN changed. " | ||
609 | "Port-id %x wwpn %llx\n", | ||
610 | rdata->ids.port_id, port_name); | ||
611 | lport->tt.rport_logoff(rdata); | ||
612 | |||
613 | new_rdata = lport->tt.rport_create(lport, | ||
614 | rdata->ids.port_id); | ||
615 | if (new_rdata) { | ||
616 | new_rdata->disc_id = disc->disc_id; | ||
617 | lport->tt.rport_login(new_rdata); | ||
618 | } | ||
619 | goto out; | ||
620 | } | ||
621 | rdata->disc_id = disc->disc_id; | ||
622 | lport->tt.rport_login(rdata); | ||
623 | } else if (ntohs(cp->ct_cmd) == FC_FS_RJT) { | ||
624 | FC_DISC_DBG(disc, "GPN_ID rejected reason %x exp %x\n", | ||
625 | cp->ct_reason, cp->ct_explan); | ||
626 | lport->tt.rport_logoff(rdata); | ||
627 | } else { | ||
628 | FC_DISC_DBG(disc, "GPN_ID unexpected response code %x\n", | ||
629 | ntohs(cp->ct_cmd)); | ||
630 | redisc: | ||
631 | fc_disc_restart(disc); | ||
791 | } | 632 | } |
792 | return; | ||
793 | out: | 633 | out: |
794 | kfree(dp); | 634 | mutex_unlock(&disc->disc_mutex); |
635 | kref_put(&rdata->kref, lport->tt.rport_destroy); | ||
636 | } | ||
637 | |||
638 | /** | ||
639 | * fc_disc_gpn_id_req() - Send Get Port Names by ID (GPN_ID) request | ||
640 | * @lport: local port | ||
641 | * @rdata: remote port private data | ||
642 | * | ||
643 | * Locking Note: This function expects that the disc_mutex is locked | ||
644 | * before it is called. | ||
645 | * On failure, an error code is returned. | ||
646 | */ | ||
647 | static int fc_disc_gpn_id_req(struct fc_lport *lport, | ||
648 | struct fc_rport_priv *rdata) | ||
649 | { | ||
650 | struct fc_frame *fp; | ||
651 | |||
652 | fp = fc_frame_alloc(lport, sizeof(struct fc_ct_hdr) + | ||
653 | sizeof(struct fc_ns_fid)); | ||
654 | if (!fp) | ||
655 | return -ENOMEM; | ||
656 | if (!lport->tt.elsct_send(lport, rdata->ids.port_id, fp, FC_NS_GPN_ID, | ||
657 | fc_disc_gpn_id_resp, rdata, lport->e_d_tov)) | ||
658 | return -ENOMEM; | ||
659 | kref_get(&rdata->kref); | ||
660 | return 0; | ||
661 | } | ||
662 | |||
663 | /** | ||
664 | * fc_disc_single() - Discover the directory information for a single target | ||
665 | * @lport: local port | ||
666 | * @dp: The port to rediscover | ||
667 | * | ||
668 | * Locking Note: This function expects that the disc_mutex is locked | ||
669 | * before it is called. | ||
670 | */ | ||
671 | static int fc_disc_single(struct fc_lport *lport, struct fc_disc_port *dp) | ||
672 | { | ||
673 | struct fc_rport_priv *rdata; | ||
674 | |||
675 | rdata = lport->tt.rport_create(lport, dp->port_id); | ||
676 | if (!rdata) | ||
677 | return -ENOMEM; | ||
678 | rdata->disc_id = 0; | ||
679 | return fc_disc_gpn_id_req(lport, rdata); | ||
795 | } | 680 | } |
796 | 681 | ||
797 | /** | 682 | /** |
@@ -841,18 +726,12 @@ int fc_disc_init(struct fc_lport *lport) | |||
841 | if (!lport->tt.disc_recv_req) | 726 | if (!lport->tt.disc_recv_req) |
842 | lport->tt.disc_recv_req = fc_disc_recv_req; | 727 | lport->tt.disc_recv_req = fc_disc_recv_req; |
843 | 728 | ||
844 | if (!lport->tt.rport_lookup) | ||
845 | lport->tt.rport_lookup = fc_disc_lookup_rport; | ||
846 | |||
847 | disc = &lport->disc; | 729 | disc = &lport->disc; |
848 | INIT_DELAYED_WORK(&disc->disc_work, fc_disc_timeout); | 730 | INIT_DELAYED_WORK(&disc->disc_work, fc_disc_timeout); |
849 | mutex_init(&disc->disc_mutex); | 731 | mutex_init(&disc->disc_mutex); |
850 | INIT_LIST_HEAD(&disc->rports); | 732 | INIT_LIST_HEAD(&disc->rports); |
851 | INIT_LIST_HEAD(&disc->rogue_rports); | ||
852 | 733 | ||
853 | disc->lport = lport; | 734 | disc->lport = lport; |
854 | disc->delay = FC_DISC_DELAY; | ||
855 | disc->event = DISC_EV_NONE; | ||
856 | 735 | ||
857 | return 0; | 736 | return 0; |
858 | } | 737 | } |
diff --git a/drivers/scsi/libfc/fc_elsct.c b/drivers/scsi/libfc/fc_elsct.c index 5878b34bff18..5cfa68732e9d 100644 --- a/drivers/scsi/libfc/fc_elsct.c +++ b/drivers/scsi/libfc/fc_elsct.c | |||
@@ -32,7 +32,7 @@ | |||
32 | * fc_elsct_send - sends ELS/CT frame | 32 | * fc_elsct_send - sends ELS/CT frame |
33 | */ | 33 | */ |
34 | static struct fc_seq *fc_elsct_send(struct fc_lport *lport, | 34 | static struct fc_seq *fc_elsct_send(struct fc_lport *lport, |
35 | struct fc_rport *rport, | 35 | u32 did, |
36 | struct fc_frame *fp, | 36 | struct fc_frame *fp, |
37 | unsigned int op, | 37 | unsigned int op, |
38 | void (*resp)(struct fc_seq *, | 38 | void (*resp)(struct fc_seq *, |
@@ -41,16 +41,17 @@ static struct fc_seq *fc_elsct_send(struct fc_lport *lport, | |||
41 | void *arg, u32 timer_msec) | 41 | void *arg, u32 timer_msec) |
42 | { | 42 | { |
43 | enum fc_rctl r_ctl; | 43 | enum fc_rctl r_ctl; |
44 | u32 did = FC_FID_NONE; | ||
45 | enum fc_fh_type fh_type; | 44 | enum fc_fh_type fh_type; |
46 | int rc; | 45 | int rc; |
47 | 46 | ||
48 | /* ELS requests */ | 47 | /* ELS requests */ |
49 | if ((op >= ELS_LS_RJT) && (op <= ELS_AUTH_ELS)) | 48 | if ((op >= ELS_LS_RJT) && (op <= ELS_AUTH_ELS)) |
50 | rc = fc_els_fill(lport, rport, fp, op, &r_ctl, &did, &fh_type); | 49 | rc = fc_els_fill(lport, did, fp, op, &r_ctl, &fh_type); |
51 | else | 50 | else { |
52 | /* CT requests */ | 51 | /* CT requests */ |
53 | rc = fc_ct_fill(lport, fp, op, &r_ctl, &did, &fh_type); | 52 | rc = fc_ct_fill(lport, did, fp, op, &r_ctl, &fh_type); |
53 | did = FC_FID_DIR_SERV; | ||
54 | } | ||
54 | 55 | ||
55 | if (rc) | 56 | if (rc) |
56 | return NULL; | 57 | return NULL; |
@@ -69,3 +70,41 @@ int fc_elsct_init(struct fc_lport *lport) | |||
69 | return 0; | 70 | return 0; |
70 | } | 71 | } |
71 | EXPORT_SYMBOL(fc_elsct_init); | 72 | EXPORT_SYMBOL(fc_elsct_init); |
73 | |||
74 | /** | ||
75 | * fc_els_resp_type() - return string describing ELS response for debug. | ||
76 | * @fp: frame pointer with possible error code. | ||
77 | */ | ||
78 | const char *fc_els_resp_type(struct fc_frame *fp) | ||
79 | { | ||
80 | const char *msg; | ||
81 | if (IS_ERR(fp)) { | ||
82 | switch (-PTR_ERR(fp)) { | ||
83 | case FC_NO_ERR: | ||
84 | msg = "response no error"; | ||
85 | break; | ||
86 | case FC_EX_TIMEOUT: | ||
87 | msg = "response timeout"; | ||
88 | break; | ||
89 | case FC_EX_CLOSED: | ||
90 | msg = "response closed"; | ||
91 | break; | ||
92 | default: | ||
93 | msg = "response unknown error"; | ||
94 | break; | ||
95 | } | ||
96 | } else { | ||
97 | switch (fc_frame_payload_op(fp)) { | ||
98 | case ELS_LS_ACC: | ||
99 | msg = "accept"; | ||
100 | break; | ||
101 | case ELS_LS_RJT: | ||
102 | msg = "reject"; | ||
103 | break; | ||
104 | default: | ||
105 | msg = "response unknown ELS"; | ||
106 | break; | ||
107 | } | ||
108 | } | ||
109 | return msg; | ||
110 | } | ||
diff --git a/drivers/scsi/libfc/fc_exch.c b/drivers/scsi/libfc/fc_exch.c index 145ab9ba55ea..c1c15748220c 100644 --- a/drivers/scsi/libfc/fc_exch.c +++ b/drivers/scsi/libfc/fc_exch.c | |||
@@ -32,6 +32,9 @@ | |||
32 | #include <scsi/libfc.h> | 32 | #include <scsi/libfc.h> |
33 | #include <scsi/fc_encode.h> | 33 | #include <scsi/fc_encode.h> |
34 | 34 | ||
35 | u16 fc_cpu_mask; /* cpu mask for possible cpus */ | ||
36 | EXPORT_SYMBOL(fc_cpu_mask); | ||
37 | static u16 fc_cpu_order; /* 2's power to represent total possible cpus */ | ||
35 | static struct kmem_cache *fc_em_cachep; /* cache for exchanges */ | 38 | static struct kmem_cache *fc_em_cachep; /* cache for exchanges */ |
36 | 39 | ||
37 | /* | 40 | /* |
@@ -48,6 +51,20 @@ static struct kmem_cache *fc_em_cachep; /* cache for exchanges */ | |||
48 | */ | 51 | */ |
49 | 52 | ||
50 | /* | 53 | /* |
54 | * Per cpu exchange pool | ||
55 | * | ||
56 | * This structure manages per cpu exchanges in array of exchange pointers. | ||
57 | * This array is allocated followed by struct fc_exch_pool memory for | ||
58 | * assigned range of exchanges to per cpu pool. | ||
59 | */ | ||
60 | struct fc_exch_pool { | ||
61 | u16 next_index; /* next possible free exchange index */ | ||
62 | u16 total_exches; /* total allocated exchanges */ | ||
63 | spinlock_t lock; /* exch pool lock */ | ||
64 | struct list_head ex_list; /* allocated exchanges list */ | ||
65 | }; | ||
66 | |||
67 | /* | ||
51 | * Exchange manager. | 68 | * Exchange manager. |
52 | * | 69 | * |
53 | * This structure is the center for creating exchanges and sequences. | 70 | * This structure is the center for creating exchanges and sequences. |
@@ -55,17 +72,13 @@ static struct kmem_cache *fc_em_cachep; /* cache for exchanges */ | |||
55 | */ | 72 | */ |
56 | struct fc_exch_mgr { | 73 | struct fc_exch_mgr { |
57 | enum fc_class class; /* default class for sequences */ | 74 | enum fc_class class; /* default class for sequences */ |
58 | spinlock_t em_lock; /* exchange manager lock, | 75 | struct kref kref; /* exchange mgr reference count */ |
59 | must be taken before ex_lock */ | ||
60 | u16 last_xid; /* last allocated exchange ID */ | ||
61 | u16 min_xid; /* min exchange ID */ | 76 | u16 min_xid; /* min exchange ID */ |
62 | u16 max_xid; /* max exchange ID */ | 77 | u16 max_xid; /* max exchange ID */ |
63 | u16 max_read; /* max exchange ID for read */ | ||
64 | u16 last_read; /* last xid allocated for read */ | ||
65 | u32 total_exches; /* total allocated exchanges */ | ||
66 | struct list_head ex_list; /* allocated exchanges list */ | 78 | struct list_head ex_list; /* allocated exchanges list */ |
67 | struct fc_lport *lp; /* fc device instance */ | ||
68 | mempool_t *ep_pool; /* reserve ep's */ | 79 | mempool_t *ep_pool; /* reserve ep's */ |
80 | u16 pool_max_index; /* max exch array index in exch pool */ | ||
81 | struct fc_exch_pool *pool; /* per cpu exch pool */ | ||
69 | 82 | ||
70 | /* | 83 | /* |
71 | * currently exchange mgr stats are updated but not used. | 84 | * currently exchange mgr stats are updated but not used. |
@@ -80,10 +93,15 @@ struct fc_exch_mgr { | |||
80 | atomic_t seq_not_found; | 93 | atomic_t seq_not_found; |
81 | atomic_t non_bls_resp; | 94 | atomic_t non_bls_resp; |
82 | } stats; | 95 | } stats; |
83 | struct fc_exch **exches; /* for exch pointers indexed by xid */ | ||
84 | }; | 96 | }; |
85 | #define fc_seq_exch(sp) container_of(sp, struct fc_exch, seq) | 97 | #define fc_seq_exch(sp) container_of(sp, struct fc_exch, seq) |
86 | 98 | ||
99 | struct fc_exch_mgr_anchor { | ||
100 | struct list_head ema_list; | ||
101 | struct fc_exch_mgr *mp; | ||
102 | bool (*match)(struct fc_frame *); | ||
103 | }; | ||
104 | |||
87 | static void fc_exch_rrq(struct fc_exch *); | 105 | static void fc_exch_rrq(struct fc_exch *); |
88 | static void fc_seq_ls_acc(struct fc_seq *); | 106 | static void fc_seq_ls_acc(struct fc_seq *); |
89 | static void fc_seq_ls_rjt(struct fc_seq *, enum fc_els_rjt_reason, | 107 | static void fc_seq_ls_rjt(struct fc_seq *, enum fc_els_rjt_reason, |
@@ -167,8 +185,8 @@ static struct fc_seq *fc_seq_start_next_locked(struct fc_seq *sp); | |||
167 | * sequence allocation and deallocation must be locked. | 185 | * sequence allocation and deallocation must be locked. |
168 | * - exchange refcnt can be done atomicly without locks. | 186 | * - exchange refcnt can be done atomicly without locks. |
169 | * - sequence allocation must be locked by exch lock. | 187 | * - sequence allocation must be locked by exch lock. |
170 | * - If the em_lock and ex_lock must be taken at the same time, then the | 188 | * - If the EM pool lock and ex_lock must be taken at the same time, then the |
171 | * em_lock must be taken before the ex_lock. | 189 | * EM pool lock must be taken before the ex_lock. |
172 | */ | 190 | */ |
173 | 191 | ||
174 | /* | 192 | /* |
@@ -268,8 +286,6 @@ static void fc_exch_release(struct fc_exch *ep) | |||
268 | mp = ep->em; | 286 | mp = ep->em; |
269 | if (ep->destructor) | 287 | if (ep->destructor) |
270 | ep->destructor(&ep->seq, ep->arg); | 288 | ep->destructor(&ep->seq, ep->arg); |
271 | if (ep->lp->tt.exch_put) | ||
272 | ep->lp->tt.exch_put(ep->lp, mp, ep->xid); | ||
273 | WARN_ON(!(ep->esb_stat & ESB_ST_COMPLETE)); | 289 | WARN_ON(!(ep->esb_stat & ESB_ST_COMPLETE)); |
274 | mempool_free(ep, mp->ep_pool); | 290 | mempool_free(ep, mp->ep_pool); |
275 | } | 291 | } |
@@ -299,17 +315,31 @@ static int fc_exch_done_locked(struct fc_exch *ep) | |||
299 | return rc; | 315 | return rc; |
300 | } | 316 | } |
301 | 317 | ||
302 | static void fc_exch_mgr_delete_ep(struct fc_exch *ep) | 318 | static inline struct fc_exch *fc_exch_ptr_get(struct fc_exch_pool *pool, |
319 | u16 index) | ||
303 | { | 320 | { |
304 | struct fc_exch_mgr *mp; | 321 | struct fc_exch **exches = (struct fc_exch **)(pool + 1); |
322 | return exches[index]; | ||
323 | } | ||
305 | 324 | ||
306 | mp = ep->em; | 325 | static inline void fc_exch_ptr_set(struct fc_exch_pool *pool, u16 index, |
307 | spin_lock_bh(&mp->em_lock); | 326 | struct fc_exch *ep) |
308 | WARN_ON(mp->total_exches <= 0); | 327 | { |
309 | mp->total_exches--; | 328 | ((struct fc_exch **)(pool + 1))[index] = ep; |
310 | mp->exches[ep->xid - mp->min_xid] = NULL; | 329 | } |
330 | |||
331 | static void fc_exch_delete(struct fc_exch *ep) | ||
332 | { | ||
333 | struct fc_exch_pool *pool; | ||
334 | |||
335 | pool = ep->pool; | ||
336 | spin_lock_bh(&pool->lock); | ||
337 | WARN_ON(pool->total_exches <= 0); | ||
338 | pool->total_exches--; | ||
339 | fc_exch_ptr_set(pool, (ep->xid - ep->em->min_xid) >> fc_cpu_order, | ||
340 | NULL); | ||
311 | list_del(&ep->ex_list); | 341 | list_del(&ep->ex_list); |
312 | spin_unlock_bh(&mp->em_lock); | 342 | spin_unlock_bh(&pool->lock); |
313 | fc_exch_release(ep); /* drop hold for exch in mp */ | 343 | fc_exch_release(ep); /* drop hold for exch in mp */ |
314 | } | 344 | } |
315 | 345 | ||
@@ -322,7 +352,7 @@ static inline void fc_exch_timer_set_locked(struct fc_exch *ep, | |||
322 | if (ep->state & (FC_EX_RST_CLEANUP | FC_EX_DONE)) | 352 | if (ep->state & (FC_EX_RST_CLEANUP | FC_EX_DONE)) |
323 | return; | 353 | return; |
324 | 354 | ||
325 | FC_EXCH_DBG(ep, "Exchange timed out, notifying the upper layer\n"); | 355 | FC_EXCH_DBG(ep, "Exchange timer armed\n"); |
326 | 356 | ||
327 | if (schedule_delayed_work(&ep->timeout_work, | 357 | if (schedule_delayed_work(&ep->timeout_work, |
328 | msecs_to_jiffies(timer_msec))) | 358 | msecs_to_jiffies(timer_msec))) |
@@ -408,6 +438,8 @@ static void fc_exch_timeout(struct work_struct *work) | |||
408 | u32 e_stat; | 438 | u32 e_stat; |
409 | int rc = 1; | 439 | int rc = 1; |
410 | 440 | ||
441 | FC_EXCH_DBG(ep, "Exchange timed out\n"); | ||
442 | |||
411 | spin_lock_bh(&ep->ex_lock); | 443 | spin_lock_bh(&ep->ex_lock); |
412 | if (ep->state & (FC_EX_RST_CLEANUP | FC_EX_DONE)) | 444 | if (ep->state & (FC_EX_RST_CLEANUP | FC_EX_DONE)) |
413 | goto unlock; | 445 | goto unlock; |
@@ -427,7 +459,7 @@ static void fc_exch_timeout(struct work_struct *work) | |||
427 | rc = fc_exch_done_locked(ep); | 459 | rc = fc_exch_done_locked(ep); |
428 | spin_unlock_bh(&ep->ex_lock); | 460 | spin_unlock_bh(&ep->ex_lock); |
429 | if (!rc) | 461 | if (!rc) |
430 | fc_exch_mgr_delete_ep(ep); | 462 | fc_exch_delete(ep); |
431 | if (resp) | 463 | if (resp) |
432 | resp(sp, ERR_PTR(-FC_EX_TIMEOUT), arg); | 464 | resp(sp, ERR_PTR(-FC_EX_TIMEOUT), arg); |
433 | fc_seq_exch_abort(sp, 2 * ep->r_a_tov); | 465 | fc_seq_exch_abort(sp, 2 * ep->r_a_tov); |
@@ -460,65 +492,20 @@ static struct fc_seq *fc_seq_alloc(struct fc_exch *ep, u8 seq_id) | |||
460 | return sp; | 492 | return sp; |
461 | } | 493 | } |
462 | 494 | ||
463 | /* | 495 | /** |
464 | * fc_em_alloc_xid - returns an xid based on request type | 496 | * fc_exch_em_alloc() - allocate an exchange from a specified EM. |
465 | * @lp : ptr to associated lport | 497 | * @lport: ptr to the local port |
466 | * @fp : ptr to the assocated frame | 498 | * @mp: ptr to the exchange manager |
467 | * | 499 | * |
468 | * check the associated fc_fsp_pkt to get scsi command type and | 500 | * Returns pointer to allocated fc_exch with exch lock held. |
469 | * command direction to decide from which range this exch id | ||
470 | * will be allocated from. | ||
471 | * | ||
472 | * Returns : 0 or an valid xid | ||
473 | */ | 501 | */ |
474 | static u16 fc_em_alloc_xid(struct fc_exch_mgr *mp, const struct fc_frame *fp) | 502 | static struct fc_exch *fc_exch_em_alloc(struct fc_lport *lport, |
475 | { | 503 | struct fc_exch_mgr *mp) |
476 | u16 xid, min, max; | ||
477 | u16 *plast; | ||
478 | struct fc_exch *ep = NULL; | ||
479 | |||
480 | if (mp->max_read) { | ||
481 | if (fc_fcp_is_read(fr_fsp(fp))) { | ||
482 | min = mp->min_xid; | ||
483 | max = mp->max_read; | ||
484 | plast = &mp->last_read; | ||
485 | } else { | ||
486 | min = mp->max_read + 1; | ||
487 | max = mp->max_xid; | ||
488 | plast = &mp->last_xid; | ||
489 | } | ||
490 | } else { | ||
491 | min = mp->min_xid; | ||
492 | max = mp->max_xid; | ||
493 | plast = &mp->last_xid; | ||
494 | } | ||
495 | xid = *plast; | ||
496 | do { | ||
497 | xid = (xid == max) ? min : xid + 1; | ||
498 | ep = mp->exches[xid - mp->min_xid]; | ||
499 | } while ((ep != NULL) && (xid != *plast)); | ||
500 | |||
501 | if (unlikely(ep)) | ||
502 | xid = 0; | ||
503 | else | ||
504 | *plast = xid; | ||
505 | |||
506 | return xid; | ||
507 | } | ||
508 | |||
509 | /* | ||
510 | * fc_exch_alloc - allocate an exchange. | ||
511 | * @mp : ptr to the exchange manager | ||
512 | * @xid: input xid | ||
513 | * | ||
514 | * if xid is supplied zero then assign next free exchange ID | ||
515 | * from exchange manager, otherwise use supplied xid. | ||
516 | * Returns with exch lock held. | ||
517 | */ | ||
518 | struct fc_exch *fc_exch_alloc(struct fc_exch_mgr *mp, | ||
519 | struct fc_frame *fp, u16 xid) | ||
520 | { | 504 | { |
521 | struct fc_exch *ep; | 505 | struct fc_exch *ep; |
506 | unsigned int cpu; | ||
507 | u16 index; | ||
508 | struct fc_exch_pool *pool; | ||
522 | 509 | ||
523 | /* allocate memory for exchange */ | 510 | /* allocate memory for exchange */ |
524 | ep = mempool_alloc(mp->ep_pool, GFP_ATOMIC); | 511 | ep = mempool_alloc(mp->ep_pool, GFP_ATOMIC); |
@@ -528,16 +515,17 @@ struct fc_exch *fc_exch_alloc(struct fc_exch_mgr *mp, | |||
528 | } | 515 | } |
529 | memset(ep, 0, sizeof(*ep)); | 516 | memset(ep, 0, sizeof(*ep)); |
530 | 517 | ||
531 | spin_lock_bh(&mp->em_lock); | 518 | cpu = smp_processor_id(); |
532 | /* alloc xid if input xid 0 */ | 519 | pool = per_cpu_ptr(mp->pool, cpu); |
533 | if (!xid) { | 520 | spin_lock_bh(&pool->lock); |
534 | /* alloc a new xid */ | 521 | index = pool->next_index; |
535 | xid = fc_em_alloc_xid(mp, fp); | 522 | /* allocate new exch from pool */ |
536 | if (!xid) { | 523 | while (fc_exch_ptr_get(pool, index)) { |
537 | printk(KERN_WARNING "libfc: Failed to allocate an exhange\n"); | 524 | index = index == mp->pool_max_index ? 0 : index + 1; |
525 | if (index == pool->next_index) | ||
538 | goto err; | 526 | goto err; |
539 | } | ||
540 | } | 527 | } |
528 | pool->next_index = index == mp->pool_max_index ? 0 : index + 1; | ||
541 | 529 | ||
542 | fc_exch_hold(ep); /* hold for exch in mp */ | 530 | fc_exch_hold(ep); /* hold for exch in mp */ |
543 | spin_lock_init(&ep->ex_lock); | 531 | spin_lock_init(&ep->ex_lock); |
@@ -548,18 +536,19 @@ struct fc_exch *fc_exch_alloc(struct fc_exch_mgr *mp, | |||
548 | */ | 536 | */ |
549 | spin_lock_bh(&ep->ex_lock); | 537 | spin_lock_bh(&ep->ex_lock); |
550 | 538 | ||
551 | mp->exches[xid - mp->min_xid] = ep; | 539 | fc_exch_ptr_set(pool, index, ep); |
552 | list_add_tail(&ep->ex_list, &mp->ex_list); | 540 | list_add_tail(&ep->ex_list, &pool->ex_list); |
553 | fc_seq_alloc(ep, ep->seq_id++); | 541 | fc_seq_alloc(ep, ep->seq_id++); |
554 | mp->total_exches++; | 542 | pool->total_exches++; |
555 | spin_unlock_bh(&mp->em_lock); | 543 | spin_unlock_bh(&pool->lock); |
556 | 544 | ||
557 | /* | 545 | /* |
558 | * update exchange | 546 | * update exchange |
559 | */ | 547 | */ |
560 | ep->oxid = ep->xid = xid; | 548 | ep->oxid = ep->xid = (index << fc_cpu_order | cpu) + mp->min_xid; |
561 | ep->em = mp; | 549 | ep->em = mp; |
562 | ep->lp = mp->lp; | 550 | ep->pool = pool; |
551 | ep->lp = lport; | ||
563 | ep->f_ctl = FC_FC_FIRST_SEQ; /* next seq is first seq */ | 552 | ep->f_ctl = FC_FC_FIRST_SEQ; /* next seq is first seq */ |
564 | ep->rxid = FC_XID_UNKNOWN; | 553 | ep->rxid = FC_XID_UNKNOWN; |
565 | ep->class = mp->class; | 554 | ep->class = mp->class; |
@@ -567,11 +556,36 @@ struct fc_exch *fc_exch_alloc(struct fc_exch_mgr *mp, | |||
567 | out: | 556 | out: |
568 | return ep; | 557 | return ep; |
569 | err: | 558 | err: |
570 | spin_unlock_bh(&mp->em_lock); | 559 | spin_unlock_bh(&pool->lock); |
571 | atomic_inc(&mp->stats.no_free_exch_xid); | 560 | atomic_inc(&mp->stats.no_free_exch_xid); |
572 | mempool_free(ep, mp->ep_pool); | 561 | mempool_free(ep, mp->ep_pool); |
573 | return NULL; | 562 | return NULL; |
574 | } | 563 | } |
564 | |||
565 | /** | ||
566 | * fc_exch_alloc() - allocate an exchange. | ||
567 | * @lport: ptr to the local port | ||
568 | * @fp: ptr to the FC frame | ||
569 | * | ||
570 | * This function walks the list of the exchange manager(EM) | ||
571 | * anchors to select a EM for new exchange allocation. The | ||
572 | * EM is selected having either a NULL match function pointer | ||
573 | * or call to match function returning true. | ||
574 | */ | ||
575 | struct fc_exch *fc_exch_alloc(struct fc_lport *lport, struct fc_frame *fp) | ||
576 | { | ||
577 | struct fc_exch_mgr_anchor *ema; | ||
578 | struct fc_exch *ep; | ||
579 | |||
580 | list_for_each_entry(ema, &lport->ema_list, ema_list) { | ||
581 | if (!ema->match || ema->match(fp)) { | ||
582 | ep = fc_exch_em_alloc(lport, ema->mp); | ||
583 | if (ep) | ||
584 | return ep; | ||
585 | } | ||
586 | } | ||
587 | return NULL; | ||
588 | } | ||
575 | EXPORT_SYMBOL(fc_exch_alloc); | 589 | EXPORT_SYMBOL(fc_exch_alloc); |
576 | 590 | ||
577 | /* | 591 | /* |
@@ -579,16 +593,18 @@ EXPORT_SYMBOL(fc_exch_alloc); | |||
579 | */ | 593 | */ |
580 | static struct fc_exch *fc_exch_find(struct fc_exch_mgr *mp, u16 xid) | 594 | static struct fc_exch *fc_exch_find(struct fc_exch_mgr *mp, u16 xid) |
581 | { | 595 | { |
596 | struct fc_exch_pool *pool; | ||
582 | struct fc_exch *ep = NULL; | 597 | struct fc_exch *ep = NULL; |
583 | 598 | ||
584 | if ((xid >= mp->min_xid) && (xid <= mp->max_xid)) { | 599 | if ((xid >= mp->min_xid) && (xid <= mp->max_xid)) { |
585 | spin_lock_bh(&mp->em_lock); | 600 | pool = per_cpu_ptr(mp->pool, xid & fc_cpu_mask); |
586 | ep = mp->exches[xid - mp->min_xid]; | 601 | spin_lock_bh(&pool->lock); |
602 | ep = fc_exch_ptr_get(pool, (xid - mp->min_xid) >> fc_cpu_order); | ||
587 | if (ep) { | 603 | if (ep) { |
588 | fc_exch_hold(ep); | 604 | fc_exch_hold(ep); |
589 | WARN_ON(ep->xid != xid); | 605 | WARN_ON(ep->xid != xid); |
590 | } | 606 | } |
591 | spin_unlock_bh(&mp->em_lock); | 607 | spin_unlock_bh(&pool->lock); |
592 | } | 608 | } |
593 | return ep; | 609 | return ep; |
594 | } | 610 | } |
@@ -602,7 +618,7 @@ void fc_exch_done(struct fc_seq *sp) | |||
602 | rc = fc_exch_done_locked(ep); | 618 | rc = fc_exch_done_locked(ep); |
603 | spin_unlock_bh(&ep->ex_lock); | 619 | spin_unlock_bh(&ep->ex_lock); |
604 | if (!rc) | 620 | if (!rc) |
605 | fc_exch_mgr_delete_ep(ep); | 621 | fc_exch_delete(ep); |
606 | } | 622 | } |
607 | EXPORT_SYMBOL(fc_exch_done); | 623 | EXPORT_SYMBOL(fc_exch_done); |
608 | 624 | ||
@@ -610,12 +626,14 @@ EXPORT_SYMBOL(fc_exch_done); | |||
610 | * Allocate a new exchange as responder. | 626 | * Allocate a new exchange as responder. |
611 | * Sets the responder ID in the frame header. | 627 | * Sets the responder ID in the frame header. |
612 | */ | 628 | */ |
613 | static struct fc_exch *fc_exch_resp(struct fc_exch_mgr *mp, struct fc_frame *fp) | 629 | static struct fc_exch *fc_exch_resp(struct fc_lport *lport, |
630 | struct fc_exch_mgr *mp, | ||
631 | struct fc_frame *fp) | ||
614 | { | 632 | { |
615 | struct fc_exch *ep; | 633 | struct fc_exch *ep; |
616 | struct fc_frame_header *fh; | 634 | struct fc_frame_header *fh; |
617 | 635 | ||
618 | ep = mp->lp->tt.exch_get(mp->lp, fp); | 636 | ep = fc_exch_alloc(lport, fp); |
619 | if (ep) { | 637 | if (ep) { |
620 | ep->class = fc_frame_class(fp); | 638 | ep->class = fc_frame_class(fp); |
621 | 639 | ||
@@ -641,7 +659,7 @@ static struct fc_exch *fc_exch_resp(struct fc_exch_mgr *mp, struct fc_frame *fp) | |||
641 | ep->esb_stat &= ~ESB_ST_SEQ_INIT; | 659 | ep->esb_stat &= ~ESB_ST_SEQ_INIT; |
642 | 660 | ||
643 | fc_exch_hold(ep); /* hold for caller */ | 661 | fc_exch_hold(ep); /* hold for caller */ |
644 | spin_unlock_bh(&ep->ex_lock); /* lock from exch_get */ | 662 | spin_unlock_bh(&ep->ex_lock); /* lock from fc_exch_alloc */ |
645 | } | 663 | } |
646 | return ep; | 664 | return ep; |
647 | } | 665 | } |
@@ -651,7 +669,8 @@ static struct fc_exch *fc_exch_resp(struct fc_exch_mgr *mp, struct fc_frame *fp) | |||
651 | * If fc_pf_rjt_reason is FC_RJT_NONE then this function will have a hold | 669 | * If fc_pf_rjt_reason is FC_RJT_NONE then this function will have a hold |
652 | * on the ep that should be released by the caller. | 670 | * on the ep that should be released by the caller. |
653 | */ | 671 | */ |
654 | static enum fc_pf_rjt_reason fc_seq_lookup_recip(struct fc_exch_mgr *mp, | 672 | static enum fc_pf_rjt_reason fc_seq_lookup_recip(struct fc_lport *lport, |
673 | struct fc_exch_mgr *mp, | ||
655 | struct fc_frame *fp) | 674 | struct fc_frame *fp) |
656 | { | 675 | { |
657 | struct fc_frame_header *fh = fc_frame_header_get(fp); | 676 | struct fc_frame_header *fh = fc_frame_header_get(fp); |
@@ -705,7 +724,7 @@ static enum fc_pf_rjt_reason fc_seq_lookup_recip(struct fc_exch_mgr *mp, | |||
705 | reject = FC_RJT_RX_ID; | 724 | reject = FC_RJT_RX_ID; |
706 | goto rel; | 725 | goto rel; |
707 | } | 726 | } |
708 | ep = fc_exch_resp(mp, fp); | 727 | ep = fc_exch_resp(lport, mp, fp); |
709 | if (!ep) { | 728 | if (!ep) { |
710 | reject = FC_RJT_EXCH_EST; /* XXX */ | 729 | reject = FC_RJT_EXCH_EST; /* XXX */ |
711 | goto out; | 730 | goto out; |
@@ -822,7 +841,6 @@ struct fc_seq *fc_seq_start_next(struct fc_seq *sp) | |||
822 | struct fc_exch *ep = fc_seq_exch(sp); | 841 | struct fc_exch *ep = fc_seq_exch(sp); |
823 | 842 | ||
824 | spin_lock_bh(&ep->ex_lock); | 843 | spin_lock_bh(&ep->ex_lock); |
825 | WARN_ON((ep->esb_stat & ESB_ST_COMPLETE) != 0); | ||
826 | sp = fc_seq_start_next_locked(sp); | 844 | sp = fc_seq_start_next_locked(sp); |
827 | spin_unlock_bh(&ep->ex_lock); | 845 | spin_unlock_bh(&ep->ex_lock); |
828 | 846 | ||
@@ -999,8 +1017,8 @@ static void fc_exch_send_ba_rjt(struct fc_frame *rx_fp, | |||
999 | */ | 1017 | */ |
1000 | memcpy(fh->fh_s_id, rx_fh->fh_d_id, 3); | 1018 | memcpy(fh->fh_s_id, rx_fh->fh_d_id, 3); |
1001 | memcpy(fh->fh_d_id, rx_fh->fh_s_id, 3); | 1019 | memcpy(fh->fh_d_id, rx_fh->fh_s_id, 3); |
1002 | fh->fh_ox_id = rx_fh->fh_rx_id; | 1020 | fh->fh_ox_id = rx_fh->fh_ox_id; |
1003 | fh->fh_rx_id = rx_fh->fh_ox_id; | 1021 | fh->fh_rx_id = rx_fh->fh_rx_id; |
1004 | fh->fh_seq_cnt = rx_fh->fh_seq_cnt; | 1022 | fh->fh_seq_cnt = rx_fh->fh_seq_cnt; |
1005 | fh->fh_r_ctl = FC_RCTL_BA_RJT; | 1023 | fh->fh_r_ctl = FC_RCTL_BA_RJT; |
1006 | fh->fh_type = FC_TYPE_BLS; | 1024 | fh->fh_type = FC_TYPE_BLS; |
@@ -1097,7 +1115,7 @@ static void fc_exch_recv_req(struct fc_lport *lp, struct fc_exch_mgr *mp, | |||
1097 | enum fc_pf_rjt_reason reject; | 1115 | enum fc_pf_rjt_reason reject; |
1098 | 1116 | ||
1099 | fr_seq(fp) = NULL; | 1117 | fr_seq(fp) = NULL; |
1100 | reject = fc_seq_lookup_recip(mp, fp); | 1118 | reject = fc_seq_lookup_recip(lp, mp, fp); |
1101 | if (reject == FC_RJT_NONE) { | 1119 | if (reject == FC_RJT_NONE) { |
1102 | sp = fr_seq(fp); /* sequence will be held */ | 1120 | sp = fr_seq(fp); /* sequence will be held */ |
1103 | ep = fc_seq_exch(sp); | 1121 | ep = fc_seq_exch(sp); |
@@ -1123,7 +1141,7 @@ static void fc_exch_recv_req(struct fc_lport *lp, struct fc_exch_mgr *mp, | |||
1123 | lp->tt.lport_recv(lp, sp, fp); | 1141 | lp->tt.lport_recv(lp, sp, fp); |
1124 | fc_exch_release(ep); /* release from lookup */ | 1142 | fc_exch_release(ep); /* release from lookup */ |
1125 | } else { | 1143 | } else { |
1126 | FC_EM_DBG(mp, "exch/seq lookup failed: reject %x\n", reject); | 1144 | FC_LPORT_DBG(lp, "exch/seq lookup failed: reject %x\n", reject); |
1127 | fc_frame_free(fp); | 1145 | fc_frame_free(fp); |
1128 | } | 1146 | } |
1129 | } | 1147 | } |
@@ -1193,7 +1211,7 @@ static void fc_exch_recv_seq_resp(struct fc_exch_mgr *mp, struct fc_frame *fp) | |||
1193 | WARN_ON(fc_seq_exch(sp) != ep); | 1211 | WARN_ON(fc_seq_exch(sp) != ep); |
1194 | spin_unlock_bh(&ep->ex_lock); | 1212 | spin_unlock_bh(&ep->ex_lock); |
1195 | if (!rc) | 1213 | if (!rc) |
1196 | fc_exch_mgr_delete_ep(ep); | 1214 | fc_exch_delete(ep); |
1197 | } | 1215 | } |
1198 | 1216 | ||
1199 | /* | 1217 | /* |
@@ -1229,13 +1247,12 @@ static void fc_exch_recv_resp(struct fc_exch_mgr *mp, struct fc_frame *fp) | |||
1229 | struct fc_seq *sp; | 1247 | struct fc_seq *sp; |
1230 | 1248 | ||
1231 | sp = fc_seq_lookup_orig(mp, fp); /* doesn't hold sequence */ | 1249 | sp = fc_seq_lookup_orig(mp, fp); /* doesn't hold sequence */ |
1232 | if (!sp) { | 1250 | |
1251 | if (!sp) | ||
1233 | atomic_inc(&mp->stats.xid_not_found); | 1252 | atomic_inc(&mp->stats.xid_not_found); |
1234 | FC_EM_DBG(mp, "seq lookup failed\n"); | 1253 | else |
1235 | } else { | ||
1236 | atomic_inc(&mp->stats.non_bls_resp); | 1254 | atomic_inc(&mp->stats.non_bls_resp); |
1237 | FC_EM_DBG(mp, "non-BLS response to sequence"); | 1255 | |
1238 | } | ||
1239 | fc_frame_free(fp); | 1256 | fc_frame_free(fp); |
1240 | } | 1257 | } |
1241 | 1258 | ||
@@ -1304,7 +1321,7 @@ static void fc_exch_abts_resp(struct fc_exch *ep, struct fc_frame *fp) | |||
1304 | rc = fc_exch_done_locked(ep); | 1321 | rc = fc_exch_done_locked(ep); |
1305 | spin_unlock_bh(&ep->ex_lock); | 1322 | spin_unlock_bh(&ep->ex_lock); |
1306 | if (!rc) | 1323 | if (!rc) |
1307 | fc_exch_mgr_delete_ep(ep); | 1324 | fc_exch_delete(ep); |
1308 | 1325 | ||
1309 | if (resp) | 1326 | if (resp) |
1310 | resp(sp, fp, ex_resp_arg); | 1327 | resp(sp, fp, ex_resp_arg); |
@@ -1447,44 +1464,77 @@ static void fc_exch_reset(struct fc_exch *ep) | |||
1447 | rc = fc_exch_done_locked(ep); | 1464 | rc = fc_exch_done_locked(ep); |
1448 | spin_unlock_bh(&ep->ex_lock); | 1465 | spin_unlock_bh(&ep->ex_lock); |
1449 | if (!rc) | 1466 | if (!rc) |
1450 | fc_exch_mgr_delete_ep(ep); | 1467 | fc_exch_delete(ep); |
1451 | 1468 | ||
1452 | if (resp) | 1469 | if (resp) |
1453 | resp(sp, ERR_PTR(-FC_EX_CLOSED), arg); | 1470 | resp(sp, ERR_PTR(-FC_EX_CLOSED), arg); |
1454 | } | 1471 | } |
1455 | 1472 | ||
1456 | /* | 1473 | /** |
1457 | * Reset an exchange manager, releasing all sequences and exchanges. | 1474 | * fc_exch_pool_reset() - Resets an per cpu exches pool. |
1458 | * If sid is non-zero, reset only exchanges we source from that FID. | 1475 | * @lport: ptr to the local port |
1459 | * If did is non-zero, reset only exchanges destined to that FID. | 1476 | * @pool: ptr to the per cpu exches pool |
1477 | * @sid: source FC ID | ||
1478 | * @did: destination FC ID | ||
1479 | * | ||
1480 | * Resets an per cpu exches pool, releasing its all sequences | ||
1481 | * and exchanges. If sid is non-zero, then reset only exchanges | ||
1482 | * we sourced from that FID. If did is non-zero, reset only | ||
1483 | * exchanges destined to that FID. | ||
1460 | */ | 1484 | */ |
1461 | void fc_exch_mgr_reset(struct fc_lport *lp, u32 sid, u32 did) | 1485 | static void fc_exch_pool_reset(struct fc_lport *lport, |
1486 | struct fc_exch_pool *pool, | ||
1487 | u32 sid, u32 did) | ||
1462 | { | 1488 | { |
1463 | struct fc_exch *ep; | 1489 | struct fc_exch *ep; |
1464 | struct fc_exch *next; | 1490 | struct fc_exch *next; |
1465 | struct fc_exch_mgr *mp = lp->emp; | ||
1466 | 1491 | ||
1467 | spin_lock_bh(&mp->em_lock); | 1492 | spin_lock_bh(&pool->lock); |
1468 | restart: | 1493 | restart: |
1469 | list_for_each_entry_safe(ep, next, &mp->ex_list, ex_list) { | 1494 | list_for_each_entry_safe(ep, next, &pool->ex_list, ex_list) { |
1470 | if ((sid == 0 || sid == ep->sid) && | 1495 | if ((lport == ep->lp) && |
1496 | (sid == 0 || sid == ep->sid) && | ||
1471 | (did == 0 || did == ep->did)) { | 1497 | (did == 0 || did == ep->did)) { |
1472 | fc_exch_hold(ep); | 1498 | fc_exch_hold(ep); |
1473 | spin_unlock_bh(&mp->em_lock); | 1499 | spin_unlock_bh(&pool->lock); |
1474 | 1500 | ||
1475 | fc_exch_reset(ep); | 1501 | fc_exch_reset(ep); |
1476 | 1502 | ||
1477 | fc_exch_release(ep); | 1503 | fc_exch_release(ep); |
1478 | spin_lock_bh(&mp->em_lock); | 1504 | spin_lock_bh(&pool->lock); |
1479 | 1505 | ||
1480 | /* | 1506 | /* |
1481 | * must restart loop incase while lock was down | 1507 | * must restart loop incase while lock |
1482 | * multiple eps were released. | 1508 | * was down multiple eps were released. |
1483 | */ | 1509 | */ |
1484 | goto restart; | 1510 | goto restart; |
1485 | } | 1511 | } |
1486 | } | 1512 | } |
1487 | spin_unlock_bh(&mp->em_lock); | 1513 | spin_unlock_bh(&pool->lock); |
1514 | } | ||
1515 | |||
1516 | /** | ||
1517 | * fc_exch_mgr_reset() - Resets all EMs of a lport | ||
1518 | * @lport: ptr to the local port | ||
1519 | * @sid: source FC ID | ||
1520 | * @did: destination FC ID | ||
1521 | * | ||
1522 | * Reset all EMs of a lport, releasing its all sequences and | ||
1523 | * exchanges. If sid is non-zero, then reset only exchanges | ||
1524 | * we sourced from that FID. If did is non-zero, reset only | ||
1525 | * exchanges destined to that FID. | ||
1526 | */ | ||
1527 | void fc_exch_mgr_reset(struct fc_lport *lport, u32 sid, u32 did) | ||
1528 | { | ||
1529 | struct fc_exch_mgr_anchor *ema; | ||
1530 | unsigned int cpu; | ||
1531 | |||
1532 | list_for_each_entry(ema, &lport->ema_list, ema_list) { | ||
1533 | for_each_possible_cpu(cpu) | ||
1534 | fc_exch_pool_reset(lport, | ||
1535 | per_cpu_ptr(ema->mp->pool, cpu), | ||
1536 | sid, did); | ||
1537 | } | ||
1488 | } | 1538 | } |
1489 | EXPORT_SYMBOL(fc_exch_mgr_reset); | 1539 | EXPORT_SYMBOL(fc_exch_mgr_reset); |
1490 | 1540 | ||
@@ -1730,85 +1780,129 @@ reject: | |||
1730 | fc_frame_free(fp); | 1780 | fc_frame_free(fp); |
1731 | } | 1781 | } |
1732 | 1782 | ||
1783 | struct fc_exch_mgr_anchor *fc_exch_mgr_add(struct fc_lport *lport, | ||
1784 | struct fc_exch_mgr *mp, | ||
1785 | bool (*match)(struct fc_frame *)) | ||
1786 | { | ||
1787 | struct fc_exch_mgr_anchor *ema; | ||
1788 | |||
1789 | ema = kmalloc(sizeof(*ema), GFP_ATOMIC); | ||
1790 | if (!ema) | ||
1791 | return ema; | ||
1792 | |||
1793 | ema->mp = mp; | ||
1794 | ema->match = match; | ||
1795 | /* add EM anchor to EM anchors list */ | ||
1796 | list_add_tail(&ema->ema_list, &lport->ema_list); | ||
1797 | kref_get(&mp->kref); | ||
1798 | return ema; | ||
1799 | } | ||
1800 | EXPORT_SYMBOL(fc_exch_mgr_add); | ||
1801 | |||
1802 | static void fc_exch_mgr_destroy(struct kref *kref) | ||
1803 | { | ||
1804 | struct fc_exch_mgr *mp = container_of(kref, struct fc_exch_mgr, kref); | ||
1805 | |||
1806 | mempool_destroy(mp->ep_pool); | ||
1807 | free_percpu(mp->pool); | ||
1808 | kfree(mp); | ||
1809 | } | ||
1810 | |||
1811 | void fc_exch_mgr_del(struct fc_exch_mgr_anchor *ema) | ||
1812 | { | ||
1813 | /* remove EM anchor from EM anchors list */ | ||
1814 | list_del(&ema->ema_list); | ||
1815 | kref_put(&ema->mp->kref, fc_exch_mgr_destroy); | ||
1816 | kfree(ema); | ||
1817 | } | ||
1818 | EXPORT_SYMBOL(fc_exch_mgr_del); | ||
1819 | |||
1733 | struct fc_exch_mgr *fc_exch_mgr_alloc(struct fc_lport *lp, | 1820 | struct fc_exch_mgr *fc_exch_mgr_alloc(struct fc_lport *lp, |
1734 | enum fc_class class, | 1821 | enum fc_class class, |
1735 | u16 min_xid, u16 max_xid) | 1822 | u16 min_xid, u16 max_xid, |
1823 | bool (*match)(struct fc_frame *)) | ||
1736 | { | 1824 | { |
1737 | struct fc_exch_mgr *mp; | 1825 | struct fc_exch_mgr *mp; |
1738 | size_t len; | 1826 | u16 pool_exch_range; |
1827 | size_t pool_size; | ||
1828 | unsigned int cpu; | ||
1829 | struct fc_exch_pool *pool; | ||
1739 | 1830 | ||
1740 | if (max_xid <= min_xid || min_xid == 0 || max_xid == FC_XID_UNKNOWN) { | 1831 | if (max_xid <= min_xid || max_xid == FC_XID_UNKNOWN || |
1832 | (min_xid & fc_cpu_mask) != 0) { | ||
1741 | FC_LPORT_DBG(lp, "Invalid min_xid 0x:%x and max_xid 0x:%x\n", | 1833 | FC_LPORT_DBG(lp, "Invalid min_xid 0x:%x and max_xid 0x:%x\n", |
1742 | min_xid, max_xid); | 1834 | min_xid, max_xid); |
1743 | return NULL; | 1835 | return NULL; |
1744 | } | 1836 | } |
1745 | 1837 | ||
1746 | /* | 1838 | /* |
1747 | * Memory need for EM | 1839 | * allocate memory for EM |
1748 | */ | 1840 | */ |
1749 | #define xid_ok(i, m1, m2) (((i) >= (m1)) && ((i) <= (m2))) | 1841 | mp = kzalloc(sizeof(struct fc_exch_mgr), GFP_ATOMIC); |
1750 | len = (max_xid - min_xid + 1) * (sizeof(struct fc_exch *)); | ||
1751 | len += sizeof(struct fc_exch_mgr); | ||
1752 | |||
1753 | mp = kzalloc(len, GFP_ATOMIC); | ||
1754 | if (!mp) | 1842 | if (!mp) |
1755 | return NULL; | 1843 | return NULL; |
1756 | 1844 | ||
1757 | mp->class = class; | 1845 | mp->class = class; |
1758 | mp->total_exches = 0; | ||
1759 | mp->exches = (struct fc_exch **)(mp + 1); | ||
1760 | mp->lp = lp; | ||
1761 | /* adjust em exch xid range for offload */ | 1846 | /* adjust em exch xid range for offload */ |
1762 | mp->min_xid = min_xid; | 1847 | mp->min_xid = min_xid; |
1763 | mp->max_xid = max_xid; | 1848 | mp->max_xid = max_xid; |
1764 | mp->last_xid = min_xid - 1; | ||
1765 | mp->max_read = 0; | ||
1766 | mp->last_read = 0; | ||
1767 | if (lp->lro_enabled && xid_ok(lp->lro_xid, min_xid, max_xid)) { | ||
1768 | mp->max_read = lp->lro_xid; | ||
1769 | mp->last_read = min_xid - 1; | ||
1770 | mp->last_xid = mp->max_read; | ||
1771 | } else { | ||
1772 | /* disable lro if no xid control over read */ | ||
1773 | lp->lro_enabled = 0; | ||
1774 | } | ||
1775 | |||
1776 | INIT_LIST_HEAD(&mp->ex_list); | ||
1777 | spin_lock_init(&mp->em_lock); | ||
1778 | 1849 | ||
1779 | mp->ep_pool = mempool_create_slab_pool(2, fc_em_cachep); | 1850 | mp->ep_pool = mempool_create_slab_pool(2, fc_em_cachep); |
1780 | if (!mp->ep_pool) | 1851 | if (!mp->ep_pool) |
1781 | goto free_mp; | 1852 | goto free_mp; |
1782 | 1853 | ||
1854 | /* | ||
1855 | * Setup per cpu exch pool with entire exchange id range equally | ||
1856 | * divided across all cpus. The exch pointers array memory is | ||
1857 | * allocated for exch range per pool. | ||
1858 | */ | ||
1859 | pool_exch_range = (mp->max_xid - mp->min_xid + 1) / (fc_cpu_mask + 1); | ||
1860 | mp->pool_max_index = pool_exch_range - 1; | ||
1861 | |||
1862 | /* | ||
1863 | * Allocate and initialize per cpu exch pool | ||
1864 | */ | ||
1865 | pool_size = sizeof(*pool) + pool_exch_range * sizeof(struct fc_exch *); | ||
1866 | mp->pool = __alloc_percpu(pool_size, __alignof__(struct fc_exch_pool)); | ||
1867 | if (!mp->pool) | ||
1868 | goto free_mempool; | ||
1869 | for_each_possible_cpu(cpu) { | ||
1870 | pool = per_cpu_ptr(mp->pool, cpu); | ||
1871 | spin_lock_init(&pool->lock); | ||
1872 | INIT_LIST_HEAD(&pool->ex_list); | ||
1873 | } | ||
1874 | |||
1875 | kref_init(&mp->kref); | ||
1876 | if (!fc_exch_mgr_add(lp, mp, match)) { | ||
1877 | free_percpu(mp->pool); | ||
1878 | goto free_mempool; | ||
1879 | } | ||
1880 | |||
1881 | /* | ||
1882 | * Above kref_init() sets mp->kref to 1 and then | ||
1883 | * call to fc_exch_mgr_add incremented mp->kref again, | ||
1884 | * so adjust that extra increment. | ||
1885 | */ | ||
1886 | kref_put(&mp->kref, fc_exch_mgr_destroy); | ||
1783 | return mp; | 1887 | return mp; |
1784 | 1888 | ||
1889 | free_mempool: | ||
1890 | mempool_destroy(mp->ep_pool); | ||
1785 | free_mp: | 1891 | free_mp: |
1786 | kfree(mp); | 1892 | kfree(mp); |
1787 | return NULL; | 1893 | return NULL; |
1788 | } | 1894 | } |
1789 | EXPORT_SYMBOL(fc_exch_mgr_alloc); | 1895 | EXPORT_SYMBOL(fc_exch_mgr_alloc); |
1790 | 1896 | ||
1791 | void fc_exch_mgr_free(struct fc_exch_mgr *mp) | 1897 | void fc_exch_mgr_free(struct fc_lport *lport) |
1792 | { | 1898 | { |
1793 | WARN_ON(!mp); | 1899 | struct fc_exch_mgr_anchor *ema, *next; |
1794 | /* | 1900 | |
1795 | * The total exch count must be zero | 1901 | list_for_each_entry_safe(ema, next, &lport->ema_list, ema_list) |
1796 | * before freeing exchange manager. | 1902 | fc_exch_mgr_del(ema); |
1797 | */ | ||
1798 | WARN_ON(mp->total_exches != 0); | ||
1799 | mempool_destroy(mp->ep_pool); | ||
1800 | kfree(mp); | ||
1801 | } | 1903 | } |
1802 | EXPORT_SYMBOL(fc_exch_mgr_free); | 1904 | EXPORT_SYMBOL(fc_exch_mgr_free); |
1803 | 1905 | ||
1804 | struct fc_exch *fc_exch_get(struct fc_lport *lp, struct fc_frame *fp) | ||
1805 | { | ||
1806 | if (!lp || !lp->emp) | ||
1807 | return NULL; | ||
1808 | |||
1809 | return fc_exch_alloc(lp->emp, fp, 0); | ||
1810 | } | ||
1811 | EXPORT_SYMBOL(fc_exch_get); | ||
1812 | 1906 | ||
1813 | struct fc_seq *fc_exch_seq_send(struct fc_lport *lp, | 1907 | struct fc_seq *fc_exch_seq_send(struct fc_lport *lp, |
1814 | struct fc_frame *fp, | 1908 | struct fc_frame *fp, |
@@ -1823,7 +1917,7 @@ struct fc_seq *fc_exch_seq_send(struct fc_lport *lp, | |||
1823 | struct fc_frame_header *fh; | 1917 | struct fc_frame_header *fh; |
1824 | int rc = 1; | 1918 | int rc = 1; |
1825 | 1919 | ||
1826 | ep = lp->tt.exch_get(lp, fp); | 1920 | ep = fc_exch_alloc(lp, fp); |
1827 | if (!ep) { | 1921 | if (!ep) { |
1828 | fc_frame_free(fp); | 1922 | fc_frame_free(fp); |
1829 | return NULL; | 1923 | return NULL; |
@@ -1843,7 +1937,8 @@ struct fc_seq *fc_exch_seq_send(struct fc_lport *lp, | |||
1843 | fc_exch_setup_hdr(ep, fp, ep->f_ctl); | 1937 | fc_exch_setup_hdr(ep, fp, ep->f_ctl); |
1844 | sp->cnt++; | 1938 | sp->cnt++; |
1845 | 1939 | ||
1846 | fc_fcp_ddp_setup(fr_fsp(fp), ep->xid); | 1940 | if (ep->xid <= lp->lro_xid) |
1941 | fc_fcp_ddp_setup(fr_fsp(fp), ep->xid); | ||
1847 | 1942 | ||
1848 | if (unlikely(lp->tt.frame_send(lp, fp))) | 1943 | if (unlikely(lp->tt.frame_send(lp, fp))) |
1849 | goto err; | 1944 | goto err; |
@@ -1860,7 +1955,7 @@ err: | |||
1860 | rc = fc_exch_done_locked(ep); | 1955 | rc = fc_exch_done_locked(ep); |
1861 | spin_unlock_bh(&ep->ex_lock); | 1956 | spin_unlock_bh(&ep->ex_lock); |
1862 | if (!rc) | 1957 | if (!rc) |
1863 | fc_exch_mgr_delete_ep(ep); | 1958 | fc_exch_delete(ep); |
1864 | return NULL; | 1959 | return NULL; |
1865 | } | 1960 | } |
1866 | EXPORT_SYMBOL(fc_exch_seq_send); | 1961 | EXPORT_SYMBOL(fc_exch_seq_send); |
@@ -1868,24 +1963,44 @@ EXPORT_SYMBOL(fc_exch_seq_send); | |||
1868 | /* | 1963 | /* |
1869 | * Receive a frame | 1964 | * Receive a frame |
1870 | */ | 1965 | */ |
1871 | void fc_exch_recv(struct fc_lport *lp, struct fc_exch_mgr *mp, | 1966 | void fc_exch_recv(struct fc_lport *lp, struct fc_frame *fp) |
1872 | struct fc_frame *fp) | ||
1873 | { | 1967 | { |
1874 | struct fc_frame_header *fh = fc_frame_header_get(fp); | 1968 | struct fc_frame_header *fh = fc_frame_header_get(fp); |
1875 | u32 f_ctl; | 1969 | struct fc_exch_mgr_anchor *ema; |
1970 | u32 f_ctl, found = 0; | ||
1971 | u16 oxid; | ||
1876 | 1972 | ||
1877 | /* lport lock ? */ | 1973 | /* lport lock ? */ |
1878 | if (!lp || !mp || (lp->state == LPORT_ST_NONE)) { | 1974 | if (!lp || lp->state == LPORT_ST_DISABLED) { |
1879 | FC_LPORT_DBG(lp, "Receiving frames for an lport that " | 1975 | FC_LPORT_DBG(lp, "Receiving frames for an lport that " |
1880 | "has not been initialized correctly\n"); | 1976 | "has not been initialized correctly\n"); |
1881 | fc_frame_free(fp); | 1977 | fc_frame_free(fp); |
1882 | return; | 1978 | return; |
1883 | } | 1979 | } |
1884 | 1980 | ||
1981 | f_ctl = ntoh24(fh->fh_f_ctl); | ||
1982 | oxid = ntohs(fh->fh_ox_id); | ||
1983 | if (f_ctl & FC_FC_EX_CTX) { | ||
1984 | list_for_each_entry(ema, &lp->ema_list, ema_list) { | ||
1985 | if ((oxid >= ema->mp->min_xid) && | ||
1986 | (oxid <= ema->mp->max_xid)) { | ||
1987 | found = 1; | ||
1988 | break; | ||
1989 | } | ||
1990 | } | ||
1991 | |||
1992 | if (!found) { | ||
1993 | FC_LPORT_DBG(lp, "Received response for out " | ||
1994 | "of range oxid:%hx\n", oxid); | ||
1995 | fc_frame_free(fp); | ||
1996 | return; | ||
1997 | } | ||
1998 | } else | ||
1999 | ema = list_entry(lp->ema_list.prev, typeof(*ema), ema_list); | ||
2000 | |||
1885 | /* | 2001 | /* |
1886 | * If frame is marked invalid, just drop it. | 2002 | * If frame is marked invalid, just drop it. |
1887 | */ | 2003 | */ |
1888 | f_ctl = ntoh24(fh->fh_f_ctl); | ||
1889 | switch (fr_eof(fp)) { | 2004 | switch (fr_eof(fp)) { |
1890 | case FC_EOF_T: | 2005 | case FC_EOF_T: |
1891 | if (f_ctl & FC_FC_END_SEQ) | 2006 | if (f_ctl & FC_FC_END_SEQ) |
@@ -1893,34 +2008,24 @@ void fc_exch_recv(struct fc_lport *lp, struct fc_exch_mgr *mp, | |||
1893 | /* fall through */ | 2008 | /* fall through */ |
1894 | case FC_EOF_N: | 2009 | case FC_EOF_N: |
1895 | if (fh->fh_type == FC_TYPE_BLS) | 2010 | if (fh->fh_type == FC_TYPE_BLS) |
1896 | fc_exch_recv_bls(mp, fp); | 2011 | fc_exch_recv_bls(ema->mp, fp); |
1897 | else if ((f_ctl & (FC_FC_EX_CTX | FC_FC_SEQ_CTX)) == | 2012 | else if ((f_ctl & (FC_FC_EX_CTX | FC_FC_SEQ_CTX)) == |
1898 | FC_FC_EX_CTX) | 2013 | FC_FC_EX_CTX) |
1899 | fc_exch_recv_seq_resp(mp, fp); | 2014 | fc_exch_recv_seq_resp(ema->mp, fp); |
1900 | else if (f_ctl & FC_FC_SEQ_CTX) | 2015 | else if (f_ctl & FC_FC_SEQ_CTX) |
1901 | fc_exch_recv_resp(mp, fp); | 2016 | fc_exch_recv_resp(ema->mp, fp); |
1902 | else | 2017 | else |
1903 | fc_exch_recv_req(lp, mp, fp); | 2018 | fc_exch_recv_req(lp, ema->mp, fp); |
1904 | break; | 2019 | break; |
1905 | default: | 2020 | default: |
1906 | FC_EM_DBG(mp, "dropping invalid frame (eof %x)", fr_eof(fp)); | 2021 | FC_LPORT_DBG(lp, "dropping invalid frame (eof %x)", fr_eof(fp)); |
1907 | fc_frame_free(fp); | 2022 | fc_frame_free(fp); |
1908 | break; | ||
1909 | } | 2023 | } |
1910 | } | 2024 | } |
1911 | EXPORT_SYMBOL(fc_exch_recv); | 2025 | EXPORT_SYMBOL(fc_exch_recv); |
1912 | 2026 | ||
1913 | int fc_exch_init(struct fc_lport *lp) | 2027 | int fc_exch_init(struct fc_lport *lp) |
1914 | { | 2028 | { |
1915 | if (!lp->tt.exch_get) { | ||
1916 | /* | ||
1917 | * exch_put() should be NULL if | ||
1918 | * exch_get() is NULL | ||
1919 | */ | ||
1920 | WARN_ON(lp->tt.exch_put); | ||
1921 | lp->tt.exch_get = fc_exch_get; | ||
1922 | } | ||
1923 | |||
1924 | if (!lp->tt.seq_start_next) | 2029 | if (!lp->tt.seq_start_next) |
1925 | lp->tt.seq_start_next = fc_seq_start_next; | 2030 | lp->tt.seq_start_next = fc_seq_start_next; |
1926 | 2031 | ||
@@ -1942,6 +2047,28 @@ int fc_exch_init(struct fc_lport *lp) | |||
1942 | if (!lp->tt.seq_exch_abort) | 2047 | if (!lp->tt.seq_exch_abort) |
1943 | lp->tt.seq_exch_abort = fc_seq_exch_abort; | 2048 | lp->tt.seq_exch_abort = fc_seq_exch_abort; |
1944 | 2049 | ||
2050 | /* | ||
2051 | * Initialize fc_cpu_mask and fc_cpu_order. The | ||
2052 | * fc_cpu_mask is set for nr_cpu_ids rounded up | ||
2053 | * to order of 2's * power and order is stored | ||
2054 | * in fc_cpu_order as this is later required in | ||
2055 | * mapping between an exch id and exch array index | ||
2056 | * in per cpu exch pool. | ||
2057 | * | ||
2058 | * This round up is required to align fc_cpu_mask | ||
2059 | * to exchange id's lower bits such that all incoming | ||
2060 | * frames of an exchange gets delivered to the same | ||
2061 | * cpu on which exchange originated by simple bitwise | ||
2062 | * AND operation between fc_cpu_mask and exchange id. | ||
2063 | */ | ||
2064 | fc_cpu_mask = 1; | ||
2065 | fc_cpu_order = 0; | ||
2066 | while (fc_cpu_mask < nr_cpu_ids) { | ||
2067 | fc_cpu_mask <<= 1; | ||
2068 | fc_cpu_order++; | ||
2069 | } | ||
2070 | fc_cpu_mask--; | ||
2071 | |||
1945 | return 0; | 2072 | return 0; |
1946 | } | 2073 | } |
1947 | EXPORT_SYMBOL(fc_exch_init); | 2074 | EXPORT_SYMBOL(fc_exch_init); |
diff --git a/drivers/scsi/libfc/fc_fcp.c b/drivers/scsi/libfc/fc_fcp.c index e303e0d12c4b..59a4408b27b5 100644 --- a/drivers/scsi/libfc/fc_fcp.c +++ b/drivers/scsi/libfc/fc_fcp.c | |||
@@ -507,33 +507,6 @@ static int fc_fcp_send_data(struct fc_fcp_pkt *fsp, struct fc_seq *seq, | |||
507 | f_ctl = FC_FC_REL_OFF; | 507 | f_ctl = FC_FC_REL_OFF; |
508 | WARN_ON(!seq); | 508 | WARN_ON(!seq); |
509 | 509 | ||
510 | /* | ||
511 | * If a get_page()/put_page() will fail, don't use sg lists | ||
512 | * in the fc_frame structure. | ||
513 | * | ||
514 | * The put_page() may be long after the I/O has completed | ||
515 | * in the case of FCoE, since the network driver does it | ||
516 | * via free_skb(). See the test in free_pages_check(). | ||
517 | * | ||
518 | * Test this case with 'dd </dev/zero >/dev/st0 bs=64k'. | ||
519 | */ | ||
520 | if (using_sg) { | ||
521 | for (sg = scsi_sglist(sc); sg; sg = sg_next(sg)) { | ||
522 | if (page_count(sg_page(sg)) == 0 || | ||
523 | (sg_page(sg)->flags & (1 << PG_lru | | ||
524 | 1 << PG_private | | ||
525 | 1 << PG_locked | | ||
526 | 1 << PG_active | | ||
527 | 1 << PG_slab | | ||
528 | 1 << PG_swapcache | | ||
529 | 1 << PG_writeback | | ||
530 | 1 << PG_reserved | | ||
531 | 1 << PG_buddy))) { | ||
532 | using_sg = 0; | ||
533 | break; | ||
534 | } | ||
535 | } | ||
536 | } | ||
537 | sg = scsi_sglist(sc); | 510 | sg = scsi_sglist(sc); |
538 | 511 | ||
539 | while (remaining > 0 && sg) { | 512 | while (remaining > 0 && sg) { |
@@ -569,8 +542,6 @@ static int fc_fcp_send_data(struct fc_fcp_pkt *fsp, struct fc_seq *seq, | |||
569 | } | 542 | } |
570 | sg_bytes = min(tlen, sg->length - offset); | 543 | sg_bytes = min(tlen, sg->length - offset); |
571 | if (using_sg) { | 544 | if (using_sg) { |
572 | WARN_ON(skb_shinfo(fp_skb(fp))->nr_frags > | ||
573 | FC_FRAME_SG_LEN); | ||
574 | get_page(sg_page(sg)); | 545 | get_page(sg_page(sg)); |
575 | skb_fill_page_desc(fp_skb(fp), | 546 | skb_fill_page_desc(fp_skb(fp), |
576 | skb_shinfo(fp_skb(fp))->nr_frags, | 547 | skb_shinfo(fp_skb(fp))->nr_frags, |
@@ -1337,7 +1308,7 @@ static void fc_fcp_rec(struct fc_fcp_pkt *fsp) | |||
1337 | fc_fill_fc_hdr(fp, FC_RCTL_ELS_REQ, rport->port_id, | 1308 | fc_fill_fc_hdr(fp, FC_RCTL_ELS_REQ, rport->port_id, |
1338 | fc_host_port_id(rp->local_port->host), FC_TYPE_ELS, | 1309 | fc_host_port_id(rp->local_port->host), FC_TYPE_ELS, |
1339 | FC_FC_FIRST_SEQ | FC_FC_END_SEQ | FC_FC_SEQ_INIT, 0); | 1310 | FC_FC_FIRST_SEQ | FC_FC_END_SEQ | FC_FC_SEQ_INIT, 0); |
1340 | if (lp->tt.elsct_send(lp, rport, fp, ELS_REC, fc_fcp_rec_resp, | 1311 | if (lp->tt.elsct_send(lp, rport->port_id, fp, ELS_REC, fc_fcp_rec_resp, |
1341 | fsp, jiffies_to_msecs(FC_SCSI_REC_TOV))) { | 1312 | fsp, jiffies_to_msecs(FC_SCSI_REC_TOV))) { |
1342 | fc_fcp_pkt_hold(fsp); /* hold while REC outstanding */ | 1313 | fc_fcp_pkt_hold(fsp); /* hold while REC outstanding */ |
1343 | return; | 1314 | return; |
diff --git a/drivers/scsi/libfc/fc_lport.c b/drivers/scsi/libfc/fc_lport.c index 745fa5555d6a..bd2f77197447 100644 --- a/drivers/scsi/libfc/fc_lport.c +++ b/drivers/scsi/libfc/fc_lport.c | |||
@@ -113,7 +113,7 @@ static void fc_lport_enter_ready(struct fc_lport *); | |||
113 | static void fc_lport_enter_logo(struct fc_lport *); | 113 | static void fc_lport_enter_logo(struct fc_lport *); |
114 | 114 | ||
115 | static const char *fc_lport_state_names[] = { | 115 | static const char *fc_lport_state_names[] = { |
116 | [LPORT_ST_NONE] = "none", | 116 | [LPORT_ST_DISABLED] = "disabled", |
117 | [LPORT_ST_FLOGI] = "FLOGI", | 117 | [LPORT_ST_FLOGI] = "FLOGI", |
118 | [LPORT_ST_DNS] = "dNS", | 118 | [LPORT_ST_DNS] = "dNS", |
119 | [LPORT_ST_RPN_ID] = "RPN_ID", | 119 | [LPORT_ST_RPN_ID] = "RPN_ID", |
@@ -133,57 +133,44 @@ static int fc_frame_drop(struct fc_lport *lport, struct fc_frame *fp) | |||
133 | /** | 133 | /** |
134 | * fc_lport_rport_callback() - Event handler for rport events | 134 | * fc_lport_rport_callback() - Event handler for rport events |
135 | * @lport: The lport which is receiving the event | 135 | * @lport: The lport which is receiving the event |
136 | * @rport: The rport which the event has occured on | 136 | * @rdata: private remote port data |
137 | * @event: The event that occured | 137 | * @event: The event that occured |
138 | * | 138 | * |
139 | * Locking Note: The rport lock should not be held when calling | 139 | * Locking Note: The rport lock should not be held when calling |
140 | * this function. | 140 | * this function. |
141 | */ | 141 | */ |
142 | static void fc_lport_rport_callback(struct fc_lport *lport, | 142 | static void fc_lport_rport_callback(struct fc_lport *lport, |
143 | struct fc_rport *rport, | 143 | struct fc_rport_priv *rdata, |
144 | enum fc_rport_event event) | 144 | enum fc_rport_event event) |
145 | { | 145 | { |
146 | FC_LPORT_DBG(lport, "Received a %d event for port (%6x)\n", event, | 146 | FC_LPORT_DBG(lport, "Received a %d event for port (%6x)\n", event, |
147 | rport->port_id); | 147 | rdata->ids.port_id); |
148 | 148 | ||
149 | mutex_lock(&lport->lp_mutex); | ||
149 | switch (event) { | 150 | switch (event) { |
150 | case RPORT_EV_CREATED: | 151 | case RPORT_EV_READY: |
151 | if (rport->port_id == FC_FID_DIR_SERV) { | 152 | if (lport->state == LPORT_ST_DNS) { |
152 | mutex_lock(&lport->lp_mutex); | 153 | lport->dns_rp = rdata; |
153 | if (lport->state == LPORT_ST_DNS) { | 154 | fc_lport_enter_rpn_id(lport); |
154 | lport->dns_rp = rport; | 155 | } else { |
155 | fc_lport_enter_rpn_id(lport); | 156 | FC_LPORT_DBG(lport, "Received an READY event " |
156 | } else { | 157 | "on port (%6x) for the directory " |
157 | FC_LPORT_DBG(lport, "Received an CREATED event " | 158 | "server, but the lport is not " |
158 | "on port (%6x) for the directory " | 159 | "in the DNS state, it's in the " |
159 | "server, but the lport is not " | 160 | "%d state", rdata->ids.port_id, |
160 | "in the DNS state, it's in the " | 161 | lport->state); |
161 | "%d state", rport->port_id, | 162 | lport->tt.rport_logoff(rdata); |
162 | lport->state); | 163 | } |
163 | lport->tt.rport_logoff(rport); | ||
164 | } | ||
165 | mutex_unlock(&lport->lp_mutex); | ||
166 | } else | ||
167 | FC_LPORT_DBG(lport, "Received an event for port (%6x) " | ||
168 | "which is not the directory server\n", | ||
169 | rport->port_id); | ||
170 | break; | 164 | break; |
171 | case RPORT_EV_LOGO: | 165 | case RPORT_EV_LOGO: |
172 | case RPORT_EV_FAILED: | 166 | case RPORT_EV_FAILED: |
173 | case RPORT_EV_STOP: | 167 | case RPORT_EV_STOP: |
174 | if (rport->port_id == FC_FID_DIR_SERV) { | 168 | lport->dns_rp = NULL; |
175 | mutex_lock(&lport->lp_mutex); | ||
176 | lport->dns_rp = NULL; | ||
177 | mutex_unlock(&lport->lp_mutex); | ||
178 | |||
179 | } else | ||
180 | FC_LPORT_DBG(lport, "Received an event for port (%6x) " | ||
181 | "which is not the directory server\n", | ||
182 | rport->port_id); | ||
183 | break; | 169 | break; |
184 | case RPORT_EV_NONE: | 170 | case RPORT_EV_NONE: |
185 | break; | 171 | break; |
186 | } | 172 | } |
173 | mutex_unlock(&lport->lp_mutex); | ||
187 | } | 174 | } |
188 | 175 | ||
189 | /** | 176 | /** |
@@ -211,20 +198,13 @@ static void fc_lport_ptp_setup(struct fc_lport *lport, | |||
211 | u32 remote_fid, u64 remote_wwpn, | 198 | u32 remote_fid, u64 remote_wwpn, |
212 | u64 remote_wwnn) | 199 | u64 remote_wwnn) |
213 | { | 200 | { |
214 | struct fc_disc_port dp; | 201 | mutex_lock(&lport->disc.disc_mutex); |
215 | 202 | if (lport->ptp_rp) | |
216 | dp.lp = lport; | ||
217 | dp.ids.port_id = remote_fid; | ||
218 | dp.ids.port_name = remote_wwpn; | ||
219 | dp.ids.node_name = remote_wwnn; | ||
220 | dp.ids.roles = FC_RPORT_ROLE_UNKNOWN; | ||
221 | |||
222 | if (lport->ptp_rp) { | ||
223 | lport->tt.rport_logoff(lport->ptp_rp); | 203 | lport->tt.rport_logoff(lport->ptp_rp); |
224 | lport->ptp_rp = NULL; | 204 | lport->ptp_rp = lport->tt.rport_create(lport, remote_fid); |
225 | } | 205 | lport->ptp_rp->ids.port_name = remote_wwpn; |
226 | 206 | lport->ptp_rp->ids.node_name = remote_wwnn; | |
227 | lport->ptp_rp = lport->tt.rport_create(&dp); | 207 | mutex_unlock(&lport->disc.disc_mutex); |
228 | 208 | ||
229 | lport->tt.rport_login(lport->ptp_rp); | 209 | lport->tt.rport_login(lport->ptp_rp); |
230 | 210 | ||
@@ -472,56 +452,6 @@ static void fc_lport_recv_rnid_req(struct fc_seq *sp, struct fc_frame *in_fp, | |||
472 | } | 452 | } |
473 | 453 | ||
474 | /** | 454 | /** |
475 | * fc_lport_recv_adisc_req() - Handle received Address Discovery Request | ||
476 | * @lport: Fibre Channel local port recieving the ADISC | ||
477 | * @sp: current sequence in the ADISC exchange | ||
478 | * @fp: ADISC request frame | ||
479 | * | ||
480 | * Locking Note: The lport lock is expected to be held before calling | ||
481 | * this function. | ||
482 | */ | ||
483 | static void fc_lport_recv_adisc_req(struct fc_seq *sp, struct fc_frame *in_fp, | ||
484 | struct fc_lport *lport) | ||
485 | { | ||
486 | struct fc_frame *fp; | ||
487 | struct fc_exch *ep = fc_seq_exch(sp); | ||
488 | struct fc_els_adisc *req, *rp; | ||
489 | struct fc_seq_els_data rjt_data; | ||
490 | size_t len; | ||
491 | u32 f_ctl; | ||
492 | |||
493 | FC_LPORT_DBG(lport, "Received ADISC request while in state %s\n", | ||
494 | fc_lport_state(lport)); | ||
495 | |||
496 | req = fc_frame_payload_get(in_fp, sizeof(*req)); | ||
497 | if (!req) { | ||
498 | rjt_data.fp = NULL; | ||
499 | rjt_data.reason = ELS_RJT_LOGIC; | ||
500 | rjt_data.explan = ELS_EXPL_NONE; | ||
501 | lport->tt.seq_els_rsp_send(sp, ELS_LS_RJT, &rjt_data); | ||
502 | } else { | ||
503 | len = sizeof(*rp); | ||
504 | fp = fc_frame_alloc(lport, len); | ||
505 | if (fp) { | ||
506 | rp = fc_frame_payload_get(fp, len); | ||
507 | memset(rp, 0, len); | ||
508 | rp->adisc_cmd = ELS_LS_ACC; | ||
509 | rp->adisc_wwpn = htonll(lport->wwpn); | ||
510 | rp->adisc_wwnn = htonll(lport->wwnn); | ||
511 | hton24(rp->adisc_port_id, | ||
512 | fc_host_port_id(lport->host)); | ||
513 | sp = lport->tt.seq_start_next(sp); | ||
514 | f_ctl = FC_FC_EX_CTX | FC_FC_LAST_SEQ; | ||
515 | f_ctl |= FC_FC_END_SEQ | FC_FC_SEQ_INIT; | ||
516 | fc_fill_fc_hdr(fp, FC_RCTL_ELS_REP, ep->did, ep->sid, | ||
517 | FC_TYPE_ELS, f_ctl, 0); | ||
518 | lport->tt.seq_send(lport, sp, fp); | ||
519 | } | ||
520 | } | ||
521 | fc_frame_free(in_fp); | ||
522 | } | ||
523 | |||
524 | /** | ||
525 | * fc_lport_recv_logo_req() - Handle received fabric LOGO request | 455 | * fc_lport_recv_logo_req() - Handle received fabric LOGO request |
526 | * @lport: Fibre Channel local port recieving the LOGO | 456 | * @lport: Fibre Channel local port recieving the LOGO |
527 | * @sp: current sequence in the LOGO exchange | 457 | * @sp: current sequence in the LOGO exchange |
@@ -550,7 +480,7 @@ int fc_fabric_login(struct fc_lport *lport) | |||
550 | int rc = -1; | 480 | int rc = -1; |
551 | 481 | ||
552 | mutex_lock(&lport->lp_mutex); | 482 | mutex_lock(&lport->lp_mutex); |
553 | if (lport->state == LPORT_ST_NONE) { | 483 | if (lport->state == LPORT_ST_DISABLED) { |
554 | fc_lport_enter_reset(lport); | 484 | fc_lport_enter_reset(lport); |
555 | rc = 0; | 485 | rc = 0; |
556 | } | 486 | } |
@@ -637,12 +567,13 @@ EXPORT_SYMBOL(fc_fabric_logoff); | |||
637 | int fc_lport_destroy(struct fc_lport *lport) | 567 | int fc_lport_destroy(struct fc_lport *lport) |
638 | { | 568 | { |
639 | mutex_lock(&lport->lp_mutex); | 569 | mutex_lock(&lport->lp_mutex); |
640 | lport->state = LPORT_ST_NONE; | 570 | lport->state = LPORT_ST_DISABLED; |
641 | lport->link_up = 0; | 571 | lport->link_up = 0; |
642 | lport->tt.frame_send = fc_frame_drop; | 572 | lport->tt.frame_send = fc_frame_drop; |
643 | mutex_unlock(&lport->lp_mutex); | 573 | mutex_unlock(&lport->lp_mutex); |
644 | 574 | ||
645 | lport->tt.fcp_abort_io(lport); | 575 | lport->tt.fcp_abort_io(lport); |
576 | lport->tt.disc_stop_final(lport); | ||
646 | lport->tt.exch_mgr_reset(lport, 0, 0); | 577 | lport->tt.exch_mgr_reset(lport, 0, 0); |
647 | return 0; | 578 | return 0; |
648 | } | 579 | } |
@@ -722,7 +653,8 @@ static void fc_lport_enter_ready(struct fc_lport *lport) | |||
722 | 653 | ||
723 | fc_lport_state_enter(lport, LPORT_ST_READY); | 654 | fc_lport_state_enter(lport, LPORT_ST_READY); |
724 | 655 | ||
725 | lport->tt.disc_start(fc_lport_disc_callback, lport); | 656 | if (!lport->ptp_rp) |
657 | lport->tt.disc_start(fc_lport_disc_callback, lport); | ||
726 | } | 658 | } |
727 | 659 | ||
728 | /** | 660 | /** |
@@ -808,8 +740,6 @@ static void fc_lport_recv_flogi_req(struct fc_seq *sp_in, | |||
808 | fc_lport_ptp_setup(lport, remote_fid, remote_wwpn, | 740 | fc_lport_ptp_setup(lport, remote_fid, remote_wwpn, |
809 | get_unaligned_be64(&flp->fl_wwnn)); | 741 | get_unaligned_be64(&flp->fl_wwnn)); |
810 | 742 | ||
811 | lport->tt.disc_start(fc_lport_disc_callback, lport); | ||
812 | |||
813 | out: | 743 | out: |
814 | sp = fr_seq(rx_fp); | 744 | sp = fr_seq(rx_fp); |
815 | fc_frame_free(rx_fp); | 745 | fc_frame_free(rx_fp); |
@@ -832,10 +762,6 @@ static void fc_lport_recv_req(struct fc_lport *lport, struct fc_seq *sp, | |||
832 | { | 762 | { |
833 | struct fc_frame_header *fh = fc_frame_header_get(fp); | 763 | struct fc_frame_header *fh = fc_frame_header_get(fp); |
834 | void (*recv) (struct fc_seq *, struct fc_frame *, struct fc_lport *); | 764 | void (*recv) (struct fc_seq *, struct fc_frame *, struct fc_lport *); |
835 | struct fc_rport *rport; | ||
836 | u32 s_id; | ||
837 | u32 d_id; | ||
838 | struct fc_seq_els_data rjt_data; | ||
839 | 765 | ||
840 | mutex_lock(&lport->lp_mutex); | 766 | mutex_lock(&lport->lp_mutex); |
841 | 767 | ||
@@ -844,11 +770,14 @@ static void fc_lport_recv_req(struct fc_lport *lport, struct fc_seq *sp, | |||
844 | * RSCN here. These don't require a session. | 770 | * RSCN here. These don't require a session. |
845 | * Even if we had a session, it might not be ready. | 771 | * Even if we had a session, it might not be ready. |
846 | */ | 772 | */ |
847 | if (fh->fh_type == FC_TYPE_ELS && fh->fh_r_ctl == FC_RCTL_ELS_REQ) { | 773 | if (!lport->link_up) |
774 | fc_frame_free(fp); | ||
775 | else if (fh->fh_type == FC_TYPE_ELS && | ||
776 | fh->fh_r_ctl == FC_RCTL_ELS_REQ) { | ||
848 | /* | 777 | /* |
849 | * Check opcode. | 778 | * Check opcode. |
850 | */ | 779 | */ |
851 | recv = NULL; | 780 | recv = lport->tt.rport_recv_req; |
852 | switch (fc_frame_payload_op(fp)) { | 781 | switch (fc_frame_payload_op(fp)) { |
853 | case ELS_FLOGI: | 782 | case ELS_FLOGI: |
854 | recv = fc_lport_recv_flogi_req; | 783 | recv = fc_lport_recv_flogi_req; |
@@ -870,34 +799,9 @@ static void fc_lport_recv_req(struct fc_lport *lport, struct fc_seq *sp, | |||
870 | case ELS_RNID: | 799 | case ELS_RNID: |
871 | recv = fc_lport_recv_rnid_req; | 800 | recv = fc_lport_recv_rnid_req; |
872 | break; | 801 | break; |
873 | case ELS_ADISC: | ||
874 | recv = fc_lport_recv_adisc_req; | ||
875 | break; | ||
876 | } | 802 | } |
877 | 803 | ||
878 | if (recv) | 804 | recv(sp, fp, lport); |
879 | recv(sp, fp, lport); | ||
880 | else { | ||
881 | /* | ||
882 | * Find session. | ||
883 | * If this is a new incoming PLOGI, we won't find it. | ||
884 | */ | ||
885 | s_id = ntoh24(fh->fh_s_id); | ||
886 | d_id = ntoh24(fh->fh_d_id); | ||
887 | |||
888 | rport = lport->tt.rport_lookup(lport, s_id); | ||
889 | if (rport) | ||
890 | lport->tt.rport_recv_req(sp, fp, rport); | ||
891 | else { | ||
892 | rjt_data.fp = NULL; | ||
893 | rjt_data.reason = ELS_RJT_UNAB; | ||
894 | rjt_data.explan = ELS_EXPL_NONE; | ||
895 | lport->tt.seq_els_rsp_send(sp, | ||
896 | ELS_LS_RJT, | ||
897 | &rjt_data); | ||
898 | fc_frame_free(fp); | ||
899 | } | ||
900 | } | ||
901 | } else { | 805 | } else { |
902 | FC_LPORT_DBG(lport, "dropping invalid frame (eof %x)\n", | 806 | FC_LPORT_DBG(lport, "dropping invalid frame (eof %x)\n", |
903 | fr_eof(fp)); | 807 | fr_eof(fp)); |
@@ -930,38 +834,61 @@ int fc_lport_reset(struct fc_lport *lport) | |||
930 | EXPORT_SYMBOL(fc_lport_reset); | 834 | EXPORT_SYMBOL(fc_lport_reset); |
931 | 835 | ||
932 | /** | 836 | /** |
933 | * fc_rport_enter_reset() - Reset the local port | 837 | * fc_lport_reset_locked() - Reset the local port |
934 | * @lport: Fibre Channel local port to be reset | 838 | * @lport: Fibre Channel local port to be reset |
935 | * | 839 | * |
936 | * Locking Note: The lport lock is expected to be held before calling | 840 | * Locking Note: The lport lock is expected to be held before calling |
937 | * this routine. | 841 | * this routine. |
938 | */ | 842 | */ |
939 | static void fc_lport_enter_reset(struct fc_lport *lport) | 843 | static void fc_lport_reset_locked(struct fc_lport *lport) |
940 | { | 844 | { |
941 | FC_LPORT_DBG(lport, "Entered RESET state from %s state\n", | ||
942 | fc_lport_state(lport)); | ||
943 | |||
944 | fc_lport_state_enter(lport, LPORT_ST_RESET); | ||
945 | |||
946 | if (lport->dns_rp) | 845 | if (lport->dns_rp) |
947 | lport->tt.rport_logoff(lport->dns_rp); | 846 | lport->tt.rport_logoff(lport->dns_rp); |
948 | 847 | ||
949 | if (lport->ptp_rp) { | 848 | lport->ptp_rp = NULL; |
950 | lport->tt.rport_logoff(lport->ptp_rp); | ||
951 | lport->ptp_rp = NULL; | ||
952 | } | ||
953 | 849 | ||
954 | lport->tt.disc_stop(lport); | 850 | lport->tt.disc_stop(lport); |
955 | 851 | ||
956 | lport->tt.exch_mgr_reset(lport, 0, 0); | 852 | lport->tt.exch_mgr_reset(lport, 0, 0); |
957 | fc_host_fabric_name(lport->host) = 0; | 853 | fc_host_fabric_name(lport->host) = 0; |
958 | fc_host_port_id(lport->host) = 0; | 854 | fc_host_port_id(lport->host) = 0; |
855 | } | ||
959 | 856 | ||
857 | /** | ||
858 | * fc_lport_enter_reset() - Reset the local port | ||
859 | * @lport: Fibre Channel local port to be reset | ||
860 | * | ||
861 | * Locking Note: The lport lock is expected to be held before calling | ||
862 | * this routine. | ||
863 | */ | ||
864 | static void fc_lport_enter_reset(struct fc_lport *lport) | ||
865 | { | ||
866 | FC_LPORT_DBG(lport, "Entered RESET state from %s state\n", | ||
867 | fc_lport_state(lport)); | ||
868 | |||
869 | fc_lport_state_enter(lport, LPORT_ST_RESET); | ||
870 | fc_lport_reset_locked(lport); | ||
960 | if (lport->link_up) | 871 | if (lport->link_up) |
961 | fc_lport_enter_flogi(lport); | 872 | fc_lport_enter_flogi(lport); |
962 | } | 873 | } |
963 | 874 | ||
964 | /** | 875 | /** |
876 | * fc_lport_enter_disabled() - disable the local port | ||
877 | * @lport: Fibre Channel local port to be reset | ||
878 | * | ||
879 | * Locking Note: The lport lock is expected to be held before calling | ||
880 | * this routine. | ||
881 | */ | ||
882 | static void fc_lport_enter_disabled(struct fc_lport *lport) | ||
883 | { | ||
884 | FC_LPORT_DBG(lport, "Entered disabled state from %s state\n", | ||
885 | fc_lport_state(lport)); | ||
886 | |||
887 | fc_lport_state_enter(lport, LPORT_ST_DISABLED); | ||
888 | fc_lport_reset_locked(lport); | ||
889 | } | ||
890 | |||
891 | /** | ||
965 | * fc_lport_error() - Handler for any errors | 892 | * fc_lport_error() - Handler for any errors |
966 | * @lport: The fc_lport object | 893 | * @lport: The fc_lport object |
967 | * @fp: The frame pointer | 894 | * @fp: The frame pointer |
@@ -992,7 +919,7 @@ static void fc_lport_error(struct fc_lport *lport, struct fc_frame *fp) | |||
992 | schedule_delayed_work(&lport->retry_work, delay); | 919 | schedule_delayed_work(&lport->retry_work, delay); |
993 | } else { | 920 | } else { |
994 | switch (lport->state) { | 921 | switch (lport->state) { |
995 | case LPORT_ST_NONE: | 922 | case LPORT_ST_DISABLED: |
996 | case LPORT_ST_READY: | 923 | case LPORT_ST_READY: |
997 | case LPORT_ST_RESET: | 924 | case LPORT_ST_RESET: |
998 | case LPORT_ST_RPN_ID: | 925 | case LPORT_ST_RPN_ID: |
@@ -1026,13 +953,13 @@ static void fc_lport_rft_id_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
1026 | struct fc_frame_header *fh; | 953 | struct fc_frame_header *fh; |
1027 | struct fc_ct_hdr *ct; | 954 | struct fc_ct_hdr *ct; |
1028 | 955 | ||
956 | FC_LPORT_DBG(lport, "Received a RFT_ID %s\n", fc_els_resp_type(fp)); | ||
957 | |||
1029 | if (fp == ERR_PTR(-FC_EX_CLOSED)) | 958 | if (fp == ERR_PTR(-FC_EX_CLOSED)) |
1030 | return; | 959 | return; |
1031 | 960 | ||
1032 | mutex_lock(&lport->lp_mutex); | 961 | mutex_lock(&lport->lp_mutex); |
1033 | 962 | ||
1034 | FC_LPORT_DBG(lport, "Received a RFT_ID response\n"); | ||
1035 | |||
1036 | if (lport->state != LPORT_ST_RFT_ID) { | 963 | if (lport->state != LPORT_ST_RFT_ID) { |
1037 | FC_LPORT_DBG(lport, "Received a RFT_ID response, but in state " | 964 | FC_LPORT_DBG(lport, "Received a RFT_ID response, but in state " |
1038 | "%s\n", fc_lport_state(lport)); | 965 | "%s\n", fc_lport_state(lport)); |
@@ -1080,13 +1007,13 @@ static void fc_lport_rpn_id_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
1080 | struct fc_frame_header *fh; | 1007 | struct fc_frame_header *fh; |
1081 | struct fc_ct_hdr *ct; | 1008 | struct fc_ct_hdr *ct; |
1082 | 1009 | ||
1010 | FC_LPORT_DBG(lport, "Received a RPN_ID %s\n", fc_els_resp_type(fp)); | ||
1011 | |||
1083 | if (fp == ERR_PTR(-FC_EX_CLOSED)) | 1012 | if (fp == ERR_PTR(-FC_EX_CLOSED)) |
1084 | return; | 1013 | return; |
1085 | 1014 | ||
1086 | mutex_lock(&lport->lp_mutex); | 1015 | mutex_lock(&lport->lp_mutex); |
1087 | 1016 | ||
1088 | FC_LPORT_DBG(lport, "Received a RPN_ID response\n"); | ||
1089 | |||
1090 | if (lport->state != LPORT_ST_RPN_ID) { | 1017 | if (lport->state != LPORT_ST_RPN_ID) { |
1091 | FC_LPORT_DBG(lport, "Received a RPN_ID response, but in state " | 1018 | FC_LPORT_DBG(lport, "Received a RPN_ID response, but in state " |
1092 | "%s\n", fc_lport_state(lport)); | 1019 | "%s\n", fc_lport_state(lport)); |
@@ -1132,13 +1059,13 @@ static void fc_lport_scr_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
1132 | struct fc_lport *lport = lp_arg; | 1059 | struct fc_lport *lport = lp_arg; |
1133 | u8 op; | 1060 | u8 op; |
1134 | 1061 | ||
1062 | FC_LPORT_DBG(lport, "Received a SCR %s\n", fc_els_resp_type(fp)); | ||
1063 | |||
1135 | if (fp == ERR_PTR(-FC_EX_CLOSED)) | 1064 | if (fp == ERR_PTR(-FC_EX_CLOSED)) |
1136 | return; | 1065 | return; |
1137 | 1066 | ||
1138 | mutex_lock(&lport->lp_mutex); | 1067 | mutex_lock(&lport->lp_mutex); |
1139 | 1068 | ||
1140 | FC_LPORT_DBG(lport, "Received a SCR response\n"); | ||
1141 | |||
1142 | if (lport->state != LPORT_ST_SCR) { | 1069 | if (lport->state != LPORT_ST_SCR) { |
1143 | FC_LPORT_DBG(lport, "Received a SCR response, but in state " | 1070 | FC_LPORT_DBG(lport, "Received a SCR response, but in state " |
1144 | "%s\n", fc_lport_state(lport)); | 1071 | "%s\n", fc_lport_state(lport)); |
@@ -1186,7 +1113,7 @@ static void fc_lport_enter_scr(struct fc_lport *lport) | |||
1186 | return; | 1113 | return; |
1187 | } | 1114 | } |
1188 | 1115 | ||
1189 | if (!lport->tt.elsct_send(lport, NULL, fp, ELS_SCR, | 1116 | if (!lport->tt.elsct_send(lport, FC_FID_FCTRL, fp, ELS_SCR, |
1190 | fc_lport_scr_resp, lport, lport->e_d_tov)) | 1117 | fc_lport_scr_resp, lport, lport->e_d_tov)) |
1191 | fc_lport_error(lport, fp); | 1118 | fc_lport_error(lport, fp); |
1192 | } | 1119 | } |
@@ -1227,7 +1154,7 @@ static void fc_lport_enter_rft_id(struct fc_lport *lport) | |||
1227 | return; | 1154 | return; |
1228 | } | 1155 | } |
1229 | 1156 | ||
1230 | if (!lport->tt.elsct_send(lport, NULL, fp, FC_NS_RFT_ID, | 1157 | if (!lport->tt.elsct_send(lport, FC_FID_DIR_SERV, fp, FC_NS_RFT_ID, |
1231 | fc_lport_rft_id_resp, | 1158 | fc_lport_rft_id_resp, |
1232 | lport, lport->e_d_tov)) | 1159 | lport, lport->e_d_tov)) |
1233 | fc_lport_error(lport, fp); | 1160 | fc_lport_error(lport, fp); |
@@ -1256,7 +1183,7 @@ static void fc_lport_enter_rpn_id(struct fc_lport *lport) | |||
1256 | return; | 1183 | return; |
1257 | } | 1184 | } |
1258 | 1185 | ||
1259 | if (!lport->tt.elsct_send(lport, NULL, fp, FC_NS_RPN_ID, | 1186 | if (!lport->tt.elsct_send(lport, FC_FID_DIR_SERV, fp, FC_NS_RPN_ID, |
1260 | fc_lport_rpn_id_resp, | 1187 | fc_lport_rpn_id_resp, |
1261 | lport, lport->e_d_tov)) | 1188 | lport, lport->e_d_tov)) |
1262 | fc_lport_error(lport, fp); | 1189 | fc_lport_error(lport, fp); |
@@ -1275,28 +1202,21 @@ static struct fc_rport_operations fc_lport_rport_ops = { | |||
1275 | */ | 1202 | */ |
1276 | static void fc_lport_enter_dns(struct fc_lport *lport) | 1203 | static void fc_lport_enter_dns(struct fc_lport *lport) |
1277 | { | 1204 | { |
1278 | struct fc_rport *rport; | 1205 | struct fc_rport_priv *rdata; |
1279 | struct fc_rport_libfc_priv *rdata; | ||
1280 | struct fc_disc_port dp; | ||
1281 | |||
1282 | dp.ids.port_id = FC_FID_DIR_SERV; | ||
1283 | dp.ids.port_name = -1; | ||
1284 | dp.ids.node_name = -1; | ||
1285 | dp.ids.roles = FC_RPORT_ROLE_UNKNOWN; | ||
1286 | dp.lp = lport; | ||
1287 | 1206 | ||
1288 | FC_LPORT_DBG(lport, "Entered DNS state from %s state\n", | 1207 | FC_LPORT_DBG(lport, "Entered DNS state from %s state\n", |
1289 | fc_lport_state(lport)); | 1208 | fc_lport_state(lport)); |
1290 | 1209 | ||
1291 | fc_lport_state_enter(lport, LPORT_ST_DNS); | 1210 | fc_lport_state_enter(lport, LPORT_ST_DNS); |
1292 | 1211 | ||
1293 | rport = lport->tt.rport_create(&dp); | 1212 | mutex_lock(&lport->disc.disc_mutex); |
1294 | if (!rport) | 1213 | rdata = lport->tt.rport_create(lport, FC_FID_DIR_SERV); |
1214 | mutex_unlock(&lport->disc.disc_mutex); | ||
1215 | if (!rdata) | ||
1295 | goto err; | 1216 | goto err; |
1296 | 1217 | ||
1297 | rdata = rport->dd_data; | ||
1298 | rdata->ops = &fc_lport_rport_ops; | 1218 | rdata->ops = &fc_lport_rport_ops; |
1299 | lport->tt.rport_login(rport); | 1219 | lport->tt.rport_login(rdata); |
1300 | return; | 1220 | return; |
1301 | 1221 | ||
1302 | err: | 1222 | err: |
@@ -1316,7 +1236,7 @@ static void fc_lport_timeout(struct work_struct *work) | |||
1316 | mutex_lock(&lport->lp_mutex); | 1236 | mutex_lock(&lport->lp_mutex); |
1317 | 1237 | ||
1318 | switch (lport->state) { | 1238 | switch (lport->state) { |
1319 | case LPORT_ST_NONE: | 1239 | case LPORT_ST_DISABLED: |
1320 | case LPORT_ST_READY: | 1240 | case LPORT_ST_READY: |
1321 | case LPORT_ST_RESET: | 1241 | case LPORT_ST_RESET: |
1322 | WARN_ON(1); | 1242 | WARN_ON(1); |
@@ -1360,13 +1280,13 @@ static void fc_lport_logo_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
1360 | struct fc_lport *lport = lp_arg; | 1280 | struct fc_lport *lport = lp_arg; |
1361 | u8 op; | 1281 | u8 op; |
1362 | 1282 | ||
1283 | FC_LPORT_DBG(lport, "Received a LOGO %s\n", fc_els_resp_type(fp)); | ||
1284 | |||
1363 | if (fp == ERR_PTR(-FC_EX_CLOSED)) | 1285 | if (fp == ERR_PTR(-FC_EX_CLOSED)) |
1364 | return; | 1286 | return; |
1365 | 1287 | ||
1366 | mutex_lock(&lport->lp_mutex); | 1288 | mutex_lock(&lport->lp_mutex); |
1367 | 1289 | ||
1368 | FC_LPORT_DBG(lport, "Received a LOGO response\n"); | ||
1369 | |||
1370 | if (lport->state != LPORT_ST_LOGO) { | 1290 | if (lport->state != LPORT_ST_LOGO) { |
1371 | FC_LPORT_DBG(lport, "Received a LOGO response, but in state " | 1291 | FC_LPORT_DBG(lport, "Received a LOGO response, but in state " |
1372 | "%s\n", fc_lport_state(lport)); | 1292 | "%s\n", fc_lport_state(lport)); |
@@ -1382,7 +1302,7 @@ static void fc_lport_logo_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
1382 | 1302 | ||
1383 | op = fc_frame_payload_op(fp); | 1303 | op = fc_frame_payload_op(fp); |
1384 | if (op == ELS_LS_ACC) | 1304 | if (op == ELS_LS_ACC) |
1385 | fc_lport_enter_reset(lport); | 1305 | fc_lport_enter_disabled(lport); |
1386 | else | 1306 | else |
1387 | fc_lport_error(lport, fp); | 1307 | fc_lport_error(lport, fp); |
1388 | 1308 | ||
@@ -1415,8 +1335,8 @@ static void fc_lport_enter_logo(struct fc_lport *lport) | |||
1415 | return; | 1335 | return; |
1416 | } | 1336 | } |
1417 | 1337 | ||
1418 | if (!lport->tt.elsct_send(lport, NULL, fp, ELS_LOGO, fc_lport_logo_resp, | 1338 | if (!lport->tt.elsct_send(lport, FC_FID_FLOGI, fp, ELS_LOGO, |
1419 | lport, lport->e_d_tov)) | 1339 | fc_lport_logo_resp, lport, lport->e_d_tov)) |
1420 | fc_lport_error(lport, fp); | 1340 | fc_lport_error(lport, fp); |
1421 | } | 1341 | } |
1422 | 1342 | ||
@@ -1442,13 +1362,13 @@ static void fc_lport_flogi_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
1442 | unsigned int e_d_tov; | 1362 | unsigned int e_d_tov; |
1443 | u16 mfs; | 1363 | u16 mfs; |
1444 | 1364 | ||
1365 | FC_LPORT_DBG(lport, "Received a FLOGI %s\n", fc_els_resp_type(fp)); | ||
1366 | |||
1445 | if (fp == ERR_PTR(-FC_EX_CLOSED)) | 1367 | if (fp == ERR_PTR(-FC_EX_CLOSED)) |
1446 | return; | 1368 | return; |
1447 | 1369 | ||
1448 | mutex_lock(&lport->lp_mutex); | 1370 | mutex_lock(&lport->lp_mutex); |
1449 | 1371 | ||
1450 | FC_LPORT_DBG(lport, "Received a FLOGI response\n"); | ||
1451 | |||
1452 | if (lport->state != LPORT_ST_FLOGI) { | 1372 | if (lport->state != LPORT_ST_FLOGI) { |
1453 | FC_LPORT_DBG(lport, "Received a FLOGI response, but in state " | 1373 | FC_LPORT_DBG(lport, "Received a FLOGI response, but in state " |
1454 | "%s\n", fc_lport_state(lport)); | 1374 | "%s\n", fc_lport_state(lport)); |
@@ -1501,14 +1421,6 @@ static void fc_lport_flogi_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
1501 | fc_lport_enter_dns(lport); | 1421 | fc_lport_enter_dns(lport); |
1502 | } | 1422 | } |
1503 | } | 1423 | } |
1504 | |||
1505 | if (flp) { | ||
1506 | csp_flags = ntohs(flp->fl_csp.sp_features); | ||
1507 | if ((csp_flags & FC_SP_FT_FPORT) == 0) { | ||
1508 | lport->tt.disc_start(fc_lport_disc_callback, | ||
1509 | lport); | ||
1510 | } | ||
1511 | } | ||
1512 | } else { | 1424 | } else { |
1513 | FC_LPORT_DBG(lport, "Bad FLOGI response\n"); | 1425 | FC_LPORT_DBG(lport, "Bad FLOGI response\n"); |
1514 | } | 1426 | } |
@@ -1539,7 +1451,7 @@ void fc_lport_enter_flogi(struct fc_lport *lport) | |||
1539 | if (!fp) | 1451 | if (!fp) |
1540 | return fc_lport_error(lport, fp); | 1452 | return fc_lport_error(lport, fp); |
1541 | 1453 | ||
1542 | if (!lport->tt.elsct_send(lport, NULL, fp, ELS_FLOGI, | 1454 | if (!lport->tt.elsct_send(lport, FC_FID_FLOGI, fp, ELS_FLOGI, |
1543 | fc_lport_flogi_resp, lport, lport->e_d_tov)) | 1455 | fc_lport_flogi_resp, lport, lport->e_d_tov)) |
1544 | fc_lport_error(lport, fp); | 1456 | fc_lport_error(lport, fp); |
1545 | } | 1457 | } |
@@ -1550,7 +1462,7 @@ int fc_lport_config(struct fc_lport *lport) | |||
1550 | INIT_DELAYED_WORK(&lport->retry_work, fc_lport_timeout); | 1462 | INIT_DELAYED_WORK(&lport->retry_work, fc_lport_timeout); |
1551 | mutex_init(&lport->lp_mutex); | 1463 | mutex_init(&lport->lp_mutex); |
1552 | 1464 | ||
1553 | fc_lport_state_enter(lport, LPORT_ST_NONE); | 1465 | fc_lport_state_enter(lport, LPORT_ST_DISABLED); |
1554 | 1466 | ||
1555 | fc_lport_add_fc4_type(lport, FC_TYPE_FCP); | 1467 | fc_lport_add_fc4_type(lport, FC_TYPE_FCP); |
1556 | fc_lport_add_fc4_type(lport, FC_TYPE_CT); | 1468 | fc_lport_add_fc4_type(lport, FC_TYPE_CT); |
@@ -1588,6 +1500,7 @@ int fc_lport_init(struct fc_lport *lport) | |||
1588 | if (lport->link_supported_speeds & FC_PORTSPEED_10GBIT) | 1500 | if (lport->link_supported_speeds & FC_PORTSPEED_10GBIT) |
1589 | fc_host_supported_speeds(lport->host) |= FC_PORTSPEED_10GBIT; | 1501 | fc_host_supported_speeds(lport->host) |= FC_PORTSPEED_10GBIT; |
1590 | 1502 | ||
1503 | INIT_LIST_HEAD(&lport->ema_list); | ||
1591 | return 0; | 1504 | return 0; |
1592 | } | 1505 | } |
1593 | EXPORT_SYMBOL(fc_lport_init); | 1506 | EXPORT_SYMBOL(fc_lport_init); |
diff --git a/drivers/scsi/libfc/fc_rport.c b/drivers/scsi/libfc/fc_rport.c index 7162385f52eb..03ea6748e7ee 100644 --- a/drivers/scsi/libfc/fc_rport.c +++ b/drivers/scsi/libfc/fc_rport.c | |||
@@ -57,94 +57,114 @@ | |||
57 | 57 | ||
58 | struct workqueue_struct *rport_event_queue; | 58 | struct workqueue_struct *rport_event_queue; |
59 | 59 | ||
60 | static void fc_rport_enter_plogi(struct fc_rport *); | 60 | static void fc_rport_enter_plogi(struct fc_rport_priv *); |
61 | static void fc_rport_enter_prli(struct fc_rport *); | 61 | static void fc_rport_enter_prli(struct fc_rport_priv *); |
62 | static void fc_rport_enter_rtv(struct fc_rport *); | 62 | static void fc_rport_enter_rtv(struct fc_rport_priv *); |
63 | static void fc_rport_enter_ready(struct fc_rport *); | 63 | static void fc_rport_enter_ready(struct fc_rport_priv *); |
64 | static void fc_rport_enter_logo(struct fc_rport *); | 64 | static void fc_rport_enter_logo(struct fc_rport_priv *); |
65 | 65 | static void fc_rport_enter_adisc(struct fc_rport_priv *); | |
66 | static void fc_rport_recv_plogi_req(struct fc_rport *, | 66 | |
67 | static void fc_rport_recv_plogi_req(struct fc_lport *, | ||
67 | struct fc_seq *, struct fc_frame *); | 68 | struct fc_seq *, struct fc_frame *); |
68 | static void fc_rport_recv_prli_req(struct fc_rport *, | 69 | static void fc_rport_recv_prli_req(struct fc_rport_priv *, |
69 | struct fc_seq *, struct fc_frame *); | 70 | struct fc_seq *, struct fc_frame *); |
70 | static void fc_rport_recv_prlo_req(struct fc_rport *, | 71 | static void fc_rport_recv_prlo_req(struct fc_rport_priv *, |
71 | struct fc_seq *, struct fc_frame *); | 72 | struct fc_seq *, struct fc_frame *); |
72 | static void fc_rport_recv_logo_req(struct fc_rport *, | 73 | static void fc_rport_recv_logo_req(struct fc_lport *, |
73 | struct fc_seq *, struct fc_frame *); | 74 | struct fc_seq *, struct fc_frame *); |
74 | static void fc_rport_timeout(struct work_struct *); | 75 | static void fc_rport_timeout(struct work_struct *); |
75 | static void fc_rport_error(struct fc_rport *, struct fc_frame *); | 76 | static void fc_rport_error(struct fc_rport_priv *, struct fc_frame *); |
76 | static void fc_rport_error_retry(struct fc_rport *, struct fc_frame *); | 77 | static void fc_rport_error_retry(struct fc_rport_priv *, struct fc_frame *); |
77 | static void fc_rport_work(struct work_struct *); | 78 | static void fc_rport_work(struct work_struct *); |
78 | 79 | ||
79 | static const char *fc_rport_state_names[] = { | 80 | static const char *fc_rport_state_names[] = { |
80 | [RPORT_ST_NONE] = "None", | ||
81 | [RPORT_ST_INIT] = "Init", | 81 | [RPORT_ST_INIT] = "Init", |
82 | [RPORT_ST_PLOGI] = "PLOGI", | 82 | [RPORT_ST_PLOGI] = "PLOGI", |
83 | [RPORT_ST_PRLI] = "PRLI", | 83 | [RPORT_ST_PRLI] = "PRLI", |
84 | [RPORT_ST_RTV] = "RTV", | 84 | [RPORT_ST_RTV] = "RTV", |
85 | [RPORT_ST_READY] = "Ready", | 85 | [RPORT_ST_READY] = "Ready", |
86 | [RPORT_ST_LOGO] = "LOGO", | 86 | [RPORT_ST_LOGO] = "LOGO", |
87 | [RPORT_ST_ADISC] = "ADISC", | ||
88 | [RPORT_ST_DELETE] = "Delete", | ||
87 | }; | 89 | }; |
88 | 90 | ||
89 | static void fc_rport_rogue_destroy(struct device *dev) | 91 | /** |
92 | * fc_rport_lookup() - lookup a remote port by port_id | ||
93 | * @lport: Fibre Channel host port instance | ||
94 | * @port_id: remote port port_id to match | ||
95 | */ | ||
96 | static struct fc_rport_priv *fc_rport_lookup(const struct fc_lport *lport, | ||
97 | u32 port_id) | ||
90 | { | 98 | { |
91 | struct fc_rport *rport = dev_to_rport(dev); | 99 | struct fc_rport_priv *rdata; |
92 | FC_RPORT_DBG(rport, "Destroying rogue rport\n"); | 100 | |
93 | kfree(rport); | 101 | list_for_each_entry(rdata, &lport->disc.rports, peers) |
102 | if (rdata->ids.port_id == port_id && | ||
103 | rdata->rp_state != RPORT_ST_DELETE) | ||
104 | return rdata; | ||
105 | return NULL; | ||
94 | } | 106 | } |
95 | 107 | ||
96 | struct fc_rport *fc_rport_rogue_create(struct fc_disc_port *dp) | 108 | /** |
109 | * fc_rport_create() - Create a new remote port | ||
110 | * @lport: The local port that the new remote port is for | ||
111 | * @port_id: The port ID for the new remote port | ||
112 | * | ||
113 | * Locking note: must be called with the disc_mutex held. | ||
114 | */ | ||
115 | static struct fc_rport_priv *fc_rport_create(struct fc_lport *lport, | ||
116 | u32 port_id) | ||
97 | { | 117 | { |
98 | struct fc_rport *rport; | 118 | struct fc_rport_priv *rdata; |
99 | struct fc_rport_libfc_priv *rdata; | ||
100 | rport = kzalloc(sizeof(*rport) + sizeof(*rdata), GFP_KERNEL); | ||
101 | 119 | ||
102 | if (!rport) | 120 | rdata = lport->tt.rport_lookup(lport, port_id); |
103 | return NULL; | 121 | if (rdata) |
122 | return rdata; | ||
104 | 123 | ||
105 | rdata = RPORT_TO_PRIV(rport); | 124 | rdata = kzalloc(sizeof(*rdata), GFP_KERNEL); |
125 | if (!rdata) | ||
126 | return NULL; | ||
106 | 127 | ||
107 | rport->dd_data = rdata; | 128 | rdata->ids.node_name = -1; |
108 | rport->port_id = dp->ids.port_id; | 129 | rdata->ids.port_name = -1; |
109 | rport->port_name = dp->ids.port_name; | 130 | rdata->ids.port_id = port_id; |
110 | rport->node_name = dp->ids.node_name; | 131 | rdata->ids.roles = FC_RPORT_ROLE_UNKNOWN; |
111 | rport->roles = dp->ids.roles; | ||
112 | rport->maxframe_size = FC_MIN_MAX_PAYLOAD; | ||
113 | /* | ||
114 | * Note: all this libfc rogue rport code will be removed for | ||
115 | * upstream so it fine that this is really ugly and hacky right now. | ||
116 | */ | ||
117 | device_initialize(&rport->dev); | ||
118 | rport->dev.release = fc_rport_rogue_destroy; | ||
119 | 132 | ||
133 | kref_init(&rdata->kref); | ||
120 | mutex_init(&rdata->rp_mutex); | 134 | mutex_init(&rdata->rp_mutex); |
121 | rdata->local_port = dp->lp; | 135 | rdata->local_port = lport; |
122 | rdata->trans_state = FC_PORTSTATE_ROGUE; | ||
123 | rdata->rp_state = RPORT_ST_INIT; | 136 | rdata->rp_state = RPORT_ST_INIT; |
124 | rdata->event = RPORT_EV_NONE; | 137 | rdata->event = RPORT_EV_NONE; |
125 | rdata->flags = FC_RP_FLAGS_REC_SUPPORTED; | 138 | rdata->flags = FC_RP_FLAGS_REC_SUPPORTED; |
126 | rdata->ops = NULL; | 139 | rdata->e_d_tov = lport->e_d_tov; |
127 | rdata->e_d_tov = dp->lp->e_d_tov; | 140 | rdata->r_a_tov = lport->r_a_tov; |
128 | rdata->r_a_tov = dp->lp->r_a_tov; | 141 | rdata->maxframe_size = FC_MIN_MAX_PAYLOAD; |
129 | INIT_DELAYED_WORK(&rdata->retry_work, fc_rport_timeout); | 142 | INIT_DELAYED_WORK(&rdata->retry_work, fc_rport_timeout); |
130 | INIT_WORK(&rdata->event_work, fc_rport_work); | 143 | INIT_WORK(&rdata->event_work, fc_rport_work); |
131 | /* | 144 | if (port_id != FC_FID_DIR_SERV) |
132 | * For good measure, but not necessary as we should only | 145 | list_add(&rdata->peers, &lport->disc.rports); |
133 | * add REAL rport to the lport list. | 146 | return rdata; |
134 | */ | 147 | } |
135 | INIT_LIST_HEAD(&rdata->peers); | 148 | |
149 | /** | ||
150 | * fc_rport_destroy() - free a remote port after last reference is released. | ||
151 | * @kref: pointer to kref inside struct fc_rport_priv | ||
152 | */ | ||
153 | static void fc_rport_destroy(struct kref *kref) | ||
154 | { | ||
155 | struct fc_rport_priv *rdata; | ||
136 | 156 | ||
137 | return rport; | 157 | rdata = container_of(kref, struct fc_rport_priv, kref); |
158 | kfree(rdata); | ||
138 | } | 159 | } |
139 | 160 | ||
140 | /** | 161 | /** |
141 | * fc_rport_state() - return a string for the state the rport is in | 162 | * fc_rport_state() - return a string for the state the rport is in |
142 | * @rport: The rport whose state we want to get a string for | 163 | * @rdata: remote port private data |
143 | */ | 164 | */ |
144 | static const char *fc_rport_state(struct fc_rport *rport) | 165 | static const char *fc_rport_state(struct fc_rport_priv *rdata) |
145 | { | 166 | { |
146 | const char *cp; | 167 | const char *cp; |
147 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
148 | 168 | ||
149 | cp = fc_rport_state_names[rdata->rp_state]; | 169 | cp = fc_rport_state_names[rdata->rp_state]; |
150 | if (!cp) | 170 | if (!cp) |
@@ -191,15 +211,14 @@ static unsigned int fc_plogi_get_maxframe(struct fc_els_flogi *flp, | |||
191 | 211 | ||
192 | /** | 212 | /** |
193 | * fc_rport_state_enter() - Change the rport's state | 213 | * fc_rport_state_enter() - Change the rport's state |
194 | * @rport: The rport whose state should change | 214 | * @rdata: The rport whose state should change |
195 | * @new: The new state of the rport | 215 | * @new: The new state of the rport |
196 | * | 216 | * |
197 | * Locking Note: Called with the rport lock held | 217 | * Locking Note: Called with the rport lock held |
198 | */ | 218 | */ |
199 | static void fc_rport_state_enter(struct fc_rport *rport, | 219 | static void fc_rport_state_enter(struct fc_rport_priv *rdata, |
200 | enum fc_rport_state new) | 220 | enum fc_rport_state new) |
201 | { | 221 | { |
202 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
203 | if (rdata->rp_state != new) | 222 | if (rdata->rp_state != new) |
204 | rdata->retries = 0; | 223 | rdata->retries = 0; |
205 | rdata->rp_state = new; | 224 | rdata->rp_state = new; |
@@ -208,147 +227,187 @@ static void fc_rport_state_enter(struct fc_rport *rport, | |||
208 | static void fc_rport_work(struct work_struct *work) | 227 | static void fc_rport_work(struct work_struct *work) |
209 | { | 228 | { |
210 | u32 port_id; | 229 | u32 port_id; |
211 | struct fc_rport_libfc_priv *rdata = | 230 | struct fc_rport_priv *rdata = |
212 | container_of(work, struct fc_rport_libfc_priv, event_work); | 231 | container_of(work, struct fc_rport_priv, event_work); |
232 | struct fc_rport_libfc_priv *rp; | ||
213 | enum fc_rport_event event; | 233 | enum fc_rport_event event; |
214 | enum fc_rport_trans_state trans_state; | ||
215 | struct fc_lport *lport = rdata->local_port; | 234 | struct fc_lport *lport = rdata->local_port; |
216 | struct fc_rport_operations *rport_ops; | 235 | struct fc_rport_operations *rport_ops; |
217 | struct fc_rport *rport = PRIV_TO_RPORT(rdata); | 236 | struct fc_rport_identifiers ids; |
237 | struct fc_rport *rport; | ||
218 | 238 | ||
219 | mutex_lock(&rdata->rp_mutex); | 239 | mutex_lock(&rdata->rp_mutex); |
220 | event = rdata->event; | 240 | event = rdata->event; |
221 | rport_ops = rdata->ops; | 241 | rport_ops = rdata->ops; |
242 | rport = rdata->rport; | ||
222 | 243 | ||
223 | if (event == RPORT_EV_CREATED) { | 244 | FC_RPORT_DBG(rdata, "work event %u\n", event); |
224 | struct fc_rport *new_rport; | ||
225 | struct fc_rport_libfc_priv *new_rdata; | ||
226 | struct fc_rport_identifiers ids; | ||
227 | 245 | ||
228 | ids.port_id = rport->port_id; | 246 | switch (event) { |
229 | ids.roles = rport->roles; | 247 | case RPORT_EV_READY: |
230 | ids.port_name = rport->port_name; | 248 | ids = rdata->ids; |
231 | ids.node_name = rport->node_name; | 249 | rdata->event = RPORT_EV_NONE; |
250 | kref_get(&rdata->kref); | ||
251 | mutex_unlock(&rdata->rp_mutex); | ||
232 | 252 | ||
253 | if (!rport) | ||
254 | rport = fc_remote_port_add(lport->host, 0, &ids); | ||
255 | if (!rport) { | ||
256 | FC_RPORT_DBG(rdata, "Failed to add the rport\n"); | ||
257 | lport->tt.rport_logoff(rdata); | ||
258 | kref_put(&rdata->kref, lport->tt.rport_destroy); | ||
259 | return; | ||
260 | } | ||
261 | mutex_lock(&rdata->rp_mutex); | ||
262 | if (rdata->rport) | ||
263 | FC_RPORT_DBG(rdata, "rport already allocated\n"); | ||
264 | rdata->rport = rport; | ||
265 | rport->maxframe_size = rdata->maxframe_size; | ||
266 | rport->supported_classes = rdata->supported_classes; | ||
267 | |||
268 | rp = rport->dd_data; | ||
269 | rp->local_port = lport; | ||
270 | rp->rp_state = rdata->rp_state; | ||
271 | rp->flags = rdata->flags; | ||
272 | rp->e_d_tov = rdata->e_d_tov; | ||
273 | rp->r_a_tov = rdata->r_a_tov; | ||
233 | mutex_unlock(&rdata->rp_mutex); | 274 | mutex_unlock(&rdata->rp_mutex); |
234 | 275 | ||
235 | new_rport = fc_remote_port_add(lport->host, 0, &ids); | 276 | if (rport_ops && rport_ops->event_callback) { |
236 | if (new_rport) { | 277 | FC_RPORT_DBG(rdata, "callback ev %d\n", event); |
237 | /* | 278 | rport_ops->event_callback(lport, rdata, event); |
238 | * Switch from the rogue rport to the rport | ||
239 | * returned by the FC class. | ||
240 | */ | ||
241 | new_rport->maxframe_size = rport->maxframe_size; | ||
242 | |||
243 | new_rdata = new_rport->dd_data; | ||
244 | new_rdata->e_d_tov = rdata->e_d_tov; | ||
245 | new_rdata->r_a_tov = rdata->r_a_tov; | ||
246 | new_rdata->ops = rdata->ops; | ||
247 | new_rdata->local_port = rdata->local_port; | ||
248 | new_rdata->flags = FC_RP_FLAGS_REC_SUPPORTED; | ||
249 | new_rdata->trans_state = FC_PORTSTATE_REAL; | ||
250 | mutex_init(&new_rdata->rp_mutex); | ||
251 | INIT_DELAYED_WORK(&new_rdata->retry_work, | ||
252 | fc_rport_timeout); | ||
253 | INIT_LIST_HEAD(&new_rdata->peers); | ||
254 | INIT_WORK(&new_rdata->event_work, fc_rport_work); | ||
255 | |||
256 | fc_rport_state_enter(new_rport, RPORT_ST_READY); | ||
257 | } else { | ||
258 | printk(KERN_WARNING "libfc: Failed to allocate " | ||
259 | " memory for rport (%6x)\n", ids.port_id); | ||
260 | event = RPORT_EV_FAILED; | ||
261 | } | 279 | } |
262 | if (rport->port_id != FC_FID_DIR_SERV) | 280 | kref_put(&rdata->kref, lport->tt.rport_destroy); |
263 | if (rport_ops->event_callback) | 281 | break; |
264 | rport_ops->event_callback(lport, rport, | 282 | |
265 | RPORT_EV_FAILED); | 283 | case RPORT_EV_FAILED: |
266 | put_device(&rport->dev); | 284 | case RPORT_EV_LOGO: |
267 | rport = new_rport; | 285 | case RPORT_EV_STOP: |
268 | rdata = new_rport->dd_data; | 286 | port_id = rdata->ids.port_id; |
269 | if (rport_ops->event_callback) | ||
270 | rport_ops->event_callback(lport, rport, event); | ||
271 | } else if ((event == RPORT_EV_FAILED) || | ||
272 | (event == RPORT_EV_LOGO) || | ||
273 | (event == RPORT_EV_STOP)) { | ||
274 | trans_state = rdata->trans_state; | ||
275 | mutex_unlock(&rdata->rp_mutex); | 287 | mutex_unlock(&rdata->rp_mutex); |
276 | if (rport_ops->event_callback) | 288 | |
277 | rport_ops->event_callback(lport, rport, event); | 289 | if (port_id != FC_FID_DIR_SERV) { |
278 | if (trans_state == FC_PORTSTATE_ROGUE) | 290 | mutex_lock(&lport->disc.disc_mutex); |
279 | put_device(&rport->dev); | 291 | list_del(&rdata->peers); |
280 | else { | 292 | mutex_unlock(&lport->disc.disc_mutex); |
281 | port_id = rport->port_id; | 293 | } |
294 | |||
295 | if (rport_ops && rport_ops->event_callback) { | ||
296 | FC_RPORT_DBG(rdata, "callback ev %d\n", event); | ||
297 | rport_ops->event_callback(lport, rdata, event); | ||
298 | } | ||
299 | cancel_delayed_work_sync(&rdata->retry_work); | ||
300 | |||
301 | /* | ||
302 | * Reset any outstanding exchanges before freeing rport. | ||
303 | */ | ||
304 | lport->tt.exch_mgr_reset(lport, 0, port_id); | ||
305 | lport->tt.exch_mgr_reset(lport, port_id, 0); | ||
306 | |||
307 | if (rport) { | ||
308 | rp = rport->dd_data; | ||
309 | rp->rp_state = RPORT_ST_DELETE; | ||
310 | mutex_lock(&rdata->rp_mutex); | ||
311 | rdata->rport = NULL; | ||
312 | mutex_unlock(&rdata->rp_mutex); | ||
282 | fc_remote_port_delete(rport); | 313 | fc_remote_port_delete(rport); |
283 | lport->tt.exch_mgr_reset(lport, 0, port_id); | ||
284 | lport->tt.exch_mgr_reset(lport, port_id, 0); | ||
285 | } | 314 | } |
286 | } else | 315 | kref_put(&rdata->kref, lport->tt.rport_destroy); |
316 | break; | ||
317 | |||
318 | default: | ||
287 | mutex_unlock(&rdata->rp_mutex); | 319 | mutex_unlock(&rdata->rp_mutex); |
320 | break; | ||
321 | } | ||
288 | } | 322 | } |
289 | 323 | ||
290 | /** | 324 | /** |
291 | * fc_rport_login() - Start the remote port login state machine | 325 | * fc_rport_login() - Start the remote port login state machine |
292 | * @rport: Fibre Channel remote port | 326 | * @rdata: private remote port |
293 | * | 327 | * |
294 | * Locking Note: Called without the rport lock held. This | 328 | * Locking Note: Called without the rport lock held. This |
295 | * function will hold the rport lock, call an _enter_* | 329 | * function will hold the rport lock, call an _enter_* |
296 | * function and then unlock the rport. | 330 | * function and then unlock the rport. |
331 | * | ||
332 | * This indicates the intent to be logged into the remote port. | ||
333 | * If it appears we are already logged in, ADISC is used to verify | ||
334 | * the setup. | ||
297 | */ | 335 | */ |
298 | int fc_rport_login(struct fc_rport *rport) | 336 | int fc_rport_login(struct fc_rport_priv *rdata) |
299 | { | 337 | { |
300 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
301 | |||
302 | mutex_lock(&rdata->rp_mutex); | 338 | mutex_lock(&rdata->rp_mutex); |
303 | 339 | ||
304 | FC_RPORT_DBG(rport, "Login to port\n"); | 340 | switch (rdata->rp_state) { |
305 | 341 | case RPORT_ST_READY: | |
306 | fc_rport_enter_plogi(rport); | 342 | FC_RPORT_DBG(rdata, "ADISC port\n"); |
307 | 343 | fc_rport_enter_adisc(rdata); | |
344 | break; | ||
345 | default: | ||
346 | FC_RPORT_DBG(rdata, "Login to port\n"); | ||
347 | fc_rport_enter_plogi(rdata); | ||
348 | break; | ||
349 | } | ||
308 | mutex_unlock(&rdata->rp_mutex); | 350 | mutex_unlock(&rdata->rp_mutex); |
309 | 351 | ||
310 | return 0; | 352 | return 0; |
311 | } | 353 | } |
312 | 354 | ||
313 | /** | 355 | /** |
356 | * fc_rport_enter_delete() - schedule a remote port to be deleted. | ||
357 | * @rdata: private remote port | ||
358 | * @event: event to report as the reason for deletion | ||
359 | * | ||
360 | * Locking Note: Called with the rport lock held. | ||
361 | * | ||
362 | * Allow state change into DELETE only once. | ||
363 | * | ||
364 | * Call queue_work only if there's no event already pending. | ||
365 | * Set the new event so that the old pending event will not occur. | ||
366 | * Since we have the mutex, even if fc_rport_work() is already started, | ||
367 | * it'll see the new event. | ||
368 | */ | ||
369 | static void fc_rport_enter_delete(struct fc_rport_priv *rdata, | ||
370 | enum fc_rport_event event) | ||
371 | { | ||
372 | if (rdata->rp_state == RPORT_ST_DELETE) | ||
373 | return; | ||
374 | |||
375 | FC_RPORT_DBG(rdata, "Delete port\n"); | ||
376 | |||
377 | fc_rport_state_enter(rdata, RPORT_ST_DELETE); | ||
378 | |||
379 | if (rdata->event == RPORT_EV_NONE) | ||
380 | queue_work(rport_event_queue, &rdata->event_work); | ||
381 | rdata->event = event; | ||
382 | } | ||
383 | |||
384 | /** | ||
314 | * fc_rport_logoff() - Logoff and remove an rport | 385 | * fc_rport_logoff() - Logoff and remove an rport |
315 | * @rport: Fibre Channel remote port to be removed | 386 | * @rdata: private remote port |
316 | * | 387 | * |
317 | * Locking Note: Called without the rport lock held. This | 388 | * Locking Note: Called without the rport lock held. This |
318 | * function will hold the rport lock, call an _enter_* | 389 | * function will hold the rport lock, call an _enter_* |
319 | * function and then unlock the rport. | 390 | * function and then unlock the rport. |
320 | */ | 391 | */ |
321 | int fc_rport_logoff(struct fc_rport *rport) | 392 | int fc_rport_logoff(struct fc_rport_priv *rdata) |
322 | { | 393 | { |
323 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
324 | |||
325 | mutex_lock(&rdata->rp_mutex); | 394 | mutex_lock(&rdata->rp_mutex); |
326 | 395 | ||
327 | FC_RPORT_DBG(rport, "Remove port\n"); | 396 | FC_RPORT_DBG(rdata, "Remove port\n"); |
328 | 397 | ||
329 | if (rdata->rp_state == RPORT_ST_NONE) { | 398 | if (rdata->rp_state == RPORT_ST_DELETE) { |
330 | FC_RPORT_DBG(rport, "Port in NONE state, not removing\n"); | 399 | FC_RPORT_DBG(rdata, "Port in Delete state, not removing\n"); |
331 | mutex_unlock(&rdata->rp_mutex); | 400 | mutex_unlock(&rdata->rp_mutex); |
332 | goto out; | 401 | goto out; |
333 | } | 402 | } |
334 | 403 | ||
335 | fc_rport_enter_logo(rport); | 404 | fc_rport_enter_logo(rdata); |
336 | 405 | ||
337 | /* | 406 | /* |
338 | * Change the state to NONE so that we discard | 407 | * Change the state to Delete so that we discard |
339 | * the response. | 408 | * the response. |
340 | */ | 409 | */ |
341 | fc_rport_state_enter(rport, RPORT_ST_NONE); | 410 | fc_rport_enter_delete(rdata, RPORT_EV_STOP); |
342 | |||
343 | mutex_unlock(&rdata->rp_mutex); | ||
344 | |||
345 | cancel_delayed_work_sync(&rdata->retry_work); | ||
346 | |||
347 | mutex_lock(&rdata->rp_mutex); | ||
348 | |||
349 | rdata->event = RPORT_EV_STOP; | ||
350 | queue_work(rport_event_queue, &rdata->event_work); | ||
351 | |||
352 | mutex_unlock(&rdata->rp_mutex); | 411 | mutex_unlock(&rdata->rp_mutex); |
353 | 412 | ||
354 | out: | 413 | out: |
@@ -357,26 +416,25 @@ out: | |||
357 | 416 | ||
358 | /** | 417 | /** |
359 | * fc_rport_enter_ready() - The rport is ready | 418 | * fc_rport_enter_ready() - The rport is ready |
360 | * @rport: Fibre Channel remote port that is ready | 419 | * @rdata: private remote port |
361 | * | 420 | * |
362 | * Locking Note: The rport lock is expected to be held before calling | 421 | * Locking Note: The rport lock is expected to be held before calling |
363 | * this routine. | 422 | * this routine. |
364 | */ | 423 | */ |
365 | static void fc_rport_enter_ready(struct fc_rport *rport) | 424 | static void fc_rport_enter_ready(struct fc_rport_priv *rdata) |
366 | { | 425 | { |
367 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | 426 | fc_rport_state_enter(rdata, RPORT_ST_READY); |
368 | |||
369 | fc_rport_state_enter(rport, RPORT_ST_READY); | ||
370 | 427 | ||
371 | FC_RPORT_DBG(rport, "Port is Ready\n"); | 428 | FC_RPORT_DBG(rdata, "Port is Ready\n"); |
372 | 429 | ||
373 | rdata->event = RPORT_EV_CREATED; | 430 | if (rdata->event == RPORT_EV_NONE) |
374 | queue_work(rport_event_queue, &rdata->event_work); | 431 | queue_work(rport_event_queue, &rdata->event_work); |
432 | rdata->event = RPORT_EV_READY; | ||
375 | } | 433 | } |
376 | 434 | ||
377 | /** | 435 | /** |
378 | * fc_rport_timeout() - Handler for the retry_work timer. | 436 | * fc_rport_timeout() - Handler for the retry_work timer. |
379 | * @work: The work struct of the fc_rport_libfc_priv | 437 | * @work: The work struct of the fc_rport_priv |
380 | * | 438 | * |
381 | * Locking Note: Called without the rport lock held. This | 439 | * Locking Note: Called without the rport lock held. This |
382 | * function will hold the rport lock, call an _enter_* | 440 | * function will hold the rport lock, call an _enter_* |
@@ -384,63 +442,63 @@ static void fc_rport_enter_ready(struct fc_rport *rport) | |||
384 | */ | 442 | */ |
385 | static void fc_rport_timeout(struct work_struct *work) | 443 | static void fc_rport_timeout(struct work_struct *work) |
386 | { | 444 | { |
387 | struct fc_rport_libfc_priv *rdata = | 445 | struct fc_rport_priv *rdata = |
388 | container_of(work, struct fc_rport_libfc_priv, retry_work.work); | 446 | container_of(work, struct fc_rport_priv, retry_work.work); |
389 | struct fc_rport *rport = PRIV_TO_RPORT(rdata); | ||
390 | 447 | ||
391 | mutex_lock(&rdata->rp_mutex); | 448 | mutex_lock(&rdata->rp_mutex); |
392 | 449 | ||
393 | switch (rdata->rp_state) { | 450 | switch (rdata->rp_state) { |
394 | case RPORT_ST_PLOGI: | 451 | case RPORT_ST_PLOGI: |
395 | fc_rport_enter_plogi(rport); | 452 | fc_rport_enter_plogi(rdata); |
396 | break; | 453 | break; |
397 | case RPORT_ST_PRLI: | 454 | case RPORT_ST_PRLI: |
398 | fc_rport_enter_prli(rport); | 455 | fc_rport_enter_prli(rdata); |
399 | break; | 456 | break; |
400 | case RPORT_ST_RTV: | 457 | case RPORT_ST_RTV: |
401 | fc_rport_enter_rtv(rport); | 458 | fc_rport_enter_rtv(rdata); |
402 | break; | 459 | break; |
403 | case RPORT_ST_LOGO: | 460 | case RPORT_ST_LOGO: |
404 | fc_rport_enter_logo(rport); | 461 | fc_rport_enter_logo(rdata); |
462 | break; | ||
463 | case RPORT_ST_ADISC: | ||
464 | fc_rport_enter_adisc(rdata); | ||
405 | break; | 465 | break; |
406 | case RPORT_ST_READY: | 466 | case RPORT_ST_READY: |
407 | case RPORT_ST_INIT: | 467 | case RPORT_ST_INIT: |
408 | case RPORT_ST_NONE: | 468 | case RPORT_ST_DELETE: |
409 | break; | 469 | break; |
410 | } | 470 | } |
411 | 471 | ||
412 | mutex_unlock(&rdata->rp_mutex); | 472 | mutex_unlock(&rdata->rp_mutex); |
413 | put_device(&rport->dev); | ||
414 | } | 473 | } |
415 | 474 | ||
416 | /** | 475 | /** |
417 | * fc_rport_error() - Error handler, called once retries have been exhausted | 476 | * fc_rport_error() - Error handler, called once retries have been exhausted |
418 | * @rport: The fc_rport object | 477 | * @rdata: private remote port |
419 | * @fp: The frame pointer | 478 | * @fp: The frame pointer |
420 | * | 479 | * |
421 | * Locking Note: The rport lock is expected to be held before | 480 | * Locking Note: The rport lock is expected to be held before |
422 | * calling this routine | 481 | * calling this routine |
423 | */ | 482 | */ |
424 | static void fc_rport_error(struct fc_rport *rport, struct fc_frame *fp) | 483 | static void fc_rport_error(struct fc_rport_priv *rdata, struct fc_frame *fp) |
425 | { | 484 | { |
426 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | 485 | FC_RPORT_DBG(rdata, "Error %ld in state %s, retries %d\n", |
427 | 486 | IS_ERR(fp) ? -PTR_ERR(fp) : 0, | |
428 | FC_RPORT_DBG(rport, "Error %ld in state %s, retries %d\n", | 487 | fc_rport_state(rdata), rdata->retries); |
429 | PTR_ERR(fp), fc_rport_state(rport), rdata->retries); | ||
430 | 488 | ||
431 | switch (rdata->rp_state) { | 489 | switch (rdata->rp_state) { |
432 | case RPORT_ST_PLOGI: | 490 | case RPORT_ST_PLOGI: |
433 | case RPORT_ST_PRLI: | ||
434 | case RPORT_ST_LOGO: | 491 | case RPORT_ST_LOGO: |
435 | rdata->event = RPORT_EV_FAILED; | 492 | fc_rport_enter_delete(rdata, RPORT_EV_FAILED); |
436 | fc_rport_state_enter(rport, RPORT_ST_NONE); | ||
437 | queue_work(rport_event_queue, | ||
438 | &rdata->event_work); | ||
439 | break; | 493 | break; |
440 | case RPORT_ST_RTV: | 494 | case RPORT_ST_RTV: |
441 | fc_rport_enter_ready(rport); | 495 | fc_rport_enter_ready(rdata); |
442 | break; | 496 | break; |
443 | case RPORT_ST_NONE: | 497 | case RPORT_ST_PRLI: |
498 | case RPORT_ST_ADISC: | ||
499 | fc_rport_enter_logo(rdata); | ||
500 | break; | ||
501 | case RPORT_ST_DELETE: | ||
444 | case RPORT_ST_READY: | 502 | case RPORT_ST_READY: |
445 | case RPORT_ST_INIT: | 503 | case RPORT_ST_INIT: |
446 | break; | 504 | break; |
@@ -449,7 +507,7 @@ static void fc_rport_error(struct fc_rport *rport, struct fc_frame *fp) | |||
449 | 507 | ||
450 | /** | 508 | /** |
451 | * fc_rport_error_retry() - Error handler when retries are desired | 509 | * fc_rport_error_retry() - Error handler when retries are desired |
452 | * @rport: The fc_rport object | 510 | * @rdata: private remote port data |
453 | * @fp: The frame pointer | 511 | * @fp: The frame pointer |
454 | * | 512 | * |
455 | * If the error was an exchange timeout retry immediately, | 513 | * If the error was an exchange timeout retry immediately, |
@@ -458,45 +516,43 @@ static void fc_rport_error(struct fc_rport *rport, struct fc_frame *fp) | |||
458 | * Locking Note: The rport lock is expected to be held before | 516 | * Locking Note: The rport lock is expected to be held before |
459 | * calling this routine | 517 | * calling this routine |
460 | */ | 518 | */ |
461 | static void fc_rport_error_retry(struct fc_rport *rport, struct fc_frame *fp) | 519 | static void fc_rport_error_retry(struct fc_rport_priv *rdata, |
520 | struct fc_frame *fp) | ||
462 | { | 521 | { |
463 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
464 | unsigned long delay = FC_DEF_E_D_TOV; | 522 | unsigned long delay = FC_DEF_E_D_TOV; |
465 | 523 | ||
466 | /* make sure this isn't an FC_EX_CLOSED error, never retry those */ | 524 | /* make sure this isn't an FC_EX_CLOSED error, never retry those */ |
467 | if (PTR_ERR(fp) == -FC_EX_CLOSED) | 525 | if (PTR_ERR(fp) == -FC_EX_CLOSED) |
468 | return fc_rport_error(rport, fp); | 526 | return fc_rport_error(rdata, fp); |
469 | 527 | ||
470 | if (rdata->retries < rdata->local_port->max_rport_retry_count) { | 528 | if (rdata->retries < rdata->local_port->max_rport_retry_count) { |
471 | FC_RPORT_DBG(rport, "Error %ld in state %s, retrying\n", | 529 | FC_RPORT_DBG(rdata, "Error %ld in state %s, retrying\n", |
472 | PTR_ERR(fp), fc_rport_state(rport)); | 530 | PTR_ERR(fp), fc_rport_state(rdata)); |
473 | rdata->retries++; | 531 | rdata->retries++; |
474 | /* no additional delay on exchange timeouts */ | 532 | /* no additional delay on exchange timeouts */ |
475 | if (PTR_ERR(fp) == -FC_EX_TIMEOUT) | 533 | if (PTR_ERR(fp) == -FC_EX_TIMEOUT) |
476 | delay = 0; | 534 | delay = 0; |
477 | get_device(&rport->dev); | ||
478 | schedule_delayed_work(&rdata->retry_work, delay); | 535 | schedule_delayed_work(&rdata->retry_work, delay); |
479 | return; | 536 | return; |
480 | } | 537 | } |
481 | 538 | ||
482 | return fc_rport_error(rport, fp); | 539 | return fc_rport_error(rdata, fp); |
483 | } | 540 | } |
484 | 541 | ||
485 | /** | 542 | /** |
486 | * fc_rport_plogi_recv_resp() - Handle incoming ELS PLOGI response | 543 | * fc_rport_plogi_recv_resp() - Handle incoming ELS PLOGI response |
487 | * @sp: current sequence in the PLOGI exchange | 544 | * @sp: current sequence in the PLOGI exchange |
488 | * @fp: response frame | 545 | * @fp: response frame |
489 | * @rp_arg: Fibre Channel remote port | 546 | * @rdata_arg: private remote port data |
490 | * | 547 | * |
491 | * Locking Note: This function will be called without the rport lock | 548 | * Locking Note: This function will be called without the rport lock |
492 | * held, but it will lock, call an _enter_* function or fc_rport_error | 549 | * held, but it will lock, call an _enter_* function or fc_rport_error |
493 | * and then unlock the rport. | 550 | * and then unlock the rport. |
494 | */ | 551 | */ |
495 | static void fc_rport_plogi_resp(struct fc_seq *sp, struct fc_frame *fp, | 552 | static void fc_rport_plogi_resp(struct fc_seq *sp, struct fc_frame *fp, |
496 | void *rp_arg) | 553 | void *rdata_arg) |
497 | { | 554 | { |
498 | struct fc_rport *rport = rp_arg; | 555 | struct fc_rport_priv *rdata = rdata_arg; |
499 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
500 | struct fc_lport *lport = rdata->local_port; | 556 | struct fc_lport *lport = rdata->local_port; |
501 | struct fc_els_flogi *plp = NULL; | 557 | struct fc_els_flogi *plp = NULL; |
502 | unsigned int tov; | 558 | unsigned int tov; |
@@ -506,26 +562,26 @@ static void fc_rport_plogi_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
506 | 562 | ||
507 | mutex_lock(&rdata->rp_mutex); | 563 | mutex_lock(&rdata->rp_mutex); |
508 | 564 | ||
509 | FC_RPORT_DBG(rport, "Received a PLOGI response\n"); | 565 | FC_RPORT_DBG(rdata, "Received a PLOGI %s\n", fc_els_resp_type(fp)); |
510 | 566 | ||
511 | if (rdata->rp_state != RPORT_ST_PLOGI) { | 567 | if (rdata->rp_state != RPORT_ST_PLOGI) { |
512 | FC_RPORT_DBG(rport, "Received a PLOGI response, but in state " | 568 | FC_RPORT_DBG(rdata, "Received a PLOGI response, but in state " |
513 | "%s\n", fc_rport_state(rport)); | 569 | "%s\n", fc_rport_state(rdata)); |
514 | if (IS_ERR(fp)) | 570 | if (IS_ERR(fp)) |
515 | goto err; | 571 | goto err; |
516 | goto out; | 572 | goto out; |
517 | } | 573 | } |
518 | 574 | ||
519 | if (IS_ERR(fp)) { | 575 | if (IS_ERR(fp)) { |
520 | fc_rport_error_retry(rport, fp); | 576 | fc_rport_error_retry(rdata, fp); |
521 | goto err; | 577 | goto err; |
522 | } | 578 | } |
523 | 579 | ||
524 | op = fc_frame_payload_op(fp); | 580 | op = fc_frame_payload_op(fp); |
525 | if (op == ELS_LS_ACC && | 581 | if (op == ELS_LS_ACC && |
526 | (plp = fc_frame_payload_get(fp, sizeof(*plp))) != NULL) { | 582 | (plp = fc_frame_payload_get(fp, sizeof(*plp))) != NULL) { |
527 | rport->port_name = get_unaligned_be64(&plp->fl_wwpn); | 583 | rdata->ids.port_name = get_unaligned_be64(&plp->fl_wwpn); |
528 | rport->node_name = get_unaligned_be64(&plp->fl_wwnn); | 584 | rdata->ids.node_name = get_unaligned_be64(&plp->fl_wwnn); |
529 | 585 | ||
530 | tov = ntohl(plp->fl_csp.sp_e_d_tov); | 586 | tov = ntohl(plp->fl_csp.sp_e_d_tov); |
531 | if (ntohs(plp->fl_csp.sp_features) & FC_SP_FT_EDTR) | 587 | if (ntohs(plp->fl_csp.sp_features) & FC_SP_FT_EDTR) |
@@ -537,75 +593,64 @@ static void fc_rport_plogi_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
537 | if (cssp_seq < csp_seq) | 593 | if (cssp_seq < csp_seq) |
538 | csp_seq = cssp_seq; | 594 | csp_seq = cssp_seq; |
539 | rdata->max_seq = csp_seq; | 595 | rdata->max_seq = csp_seq; |
540 | rport->maxframe_size = | 596 | rdata->maxframe_size = fc_plogi_get_maxframe(plp, lport->mfs); |
541 | fc_plogi_get_maxframe(plp, lport->mfs); | 597 | fc_rport_enter_prli(rdata); |
542 | |||
543 | /* | ||
544 | * If the rport is one of the well known addresses | ||
545 | * we skip PRLI and RTV and go straight to READY. | ||
546 | */ | ||
547 | if (rport->port_id >= FC_FID_DOM_MGR) | ||
548 | fc_rport_enter_ready(rport); | ||
549 | else | ||
550 | fc_rport_enter_prli(rport); | ||
551 | } else | 598 | } else |
552 | fc_rport_error_retry(rport, fp); | 599 | fc_rport_error_retry(rdata, fp); |
553 | 600 | ||
554 | out: | 601 | out: |
555 | fc_frame_free(fp); | 602 | fc_frame_free(fp); |
556 | err: | 603 | err: |
557 | mutex_unlock(&rdata->rp_mutex); | 604 | mutex_unlock(&rdata->rp_mutex); |
558 | put_device(&rport->dev); | 605 | kref_put(&rdata->kref, rdata->local_port->tt.rport_destroy); |
559 | } | 606 | } |
560 | 607 | ||
561 | /** | 608 | /** |
562 | * fc_rport_enter_plogi() - Send Port Login (PLOGI) request to peer | 609 | * fc_rport_enter_plogi() - Send Port Login (PLOGI) request to peer |
563 | * @rport: Fibre Channel remote port to send PLOGI to | 610 | * @rdata: private remote port data |
564 | * | 611 | * |
565 | * Locking Note: The rport lock is expected to be held before calling | 612 | * Locking Note: The rport lock is expected to be held before calling |
566 | * this routine. | 613 | * this routine. |
567 | */ | 614 | */ |
568 | static void fc_rport_enter_plogi(struct fc_rport *rport) | 615 | static void fc_rport_enter_plogi(struct fc_rport_priv *rdata) |
569 | { | 616 | { |
570 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
571 | struct fc_lport *lport = rdata->local_port; | 617 | struct fc_lport *lport = rdata->local_port; |
572 | struct fc_frame *fp; | 618 | struct fc_frame *fp; |
573 | 619 | ||
574 | FC_RPORT_DBG(rport, "Port entered PLOGI state from %s state\n", | 620 | FC_RPORT_DBG(rdata, "Port entered PLOGI state from %s state\n", |
575 | fc_rport_state(rport)); | 621 | fc_rport_state(rdata)); |
576 | 622 | ||
577 | fc_rport_state_enter(rport, RPORT_ST_PLOGI); | 623 | fc_rport_state_enter(rdata, RPORT_ST_PLOGI); |
578 | 624 | ||
579 | rport->maxframe_size = FC_MIN_MAX_PAYLOAD; | 625 | rdata->maxframe_size = FC_MIN_MAX_PAYLOAD; |
580 | fp = fc_frame_alloc(lport, sizeof(struct fc_els_flogi)); | 626 | fp = fc_frame_alloc(lport, sizeof(struct fc_els_flogi)); |
581 | if (!fp) { | 627 | if (!fp) { |
582 | fc_rport_error_retry(rport, fp); | 628 | fc_rport_error_retry(rdata, fp); |
583 | return; | 629 | return; |
584 | } | 630 | } |
585 | rdata->e_d_tov = lport->e_d_tov; | 631 | rdata->e_d_tov = lport->e_d_tov; |
586 | 632 | ||
587 | if (!lport->tt.elsct_send(lport, rport, fp, ELS_PLOGI, | 633 | if (!lport->tt.elsct_send(lport, rdata->ids.port_id, fp, ELS_PLOGI, |
588 | fc_rport_plogi_resp, rport, lport->e_d_tov)) | 634 | fc_rport_plogi_resp, rdata, lport->e_d_tov)) |
589 | fc_rport_error_retry(rport, fp); | 635 | fc_rport_error_retry(rdata, fp); |
590 | else | 636 | else |
591 | get_device(&rport->dev); | 637 | kref_get(&rdata->kref); |
592 | } | 638 | } |
593 | 639 | ||
594 | /** | 640 | /** |
595 | * fc_rport_prli_resp() - Process Login (PRLI) response handler | 641 | * fc_rport_prli_resp() - Process Login (PRLI) response handler |
596 | * @sp: current sequence in the PRLI exchange | 642 | * @sp: current sequence in the PRLI exchange |
597 | * @fp: response frame | 643 | * @fp: response frame |
598 | * @rp_arg: Fibre Channel remote port | 644 | * @rdata_arg: private remote port data |
599 | * | 645 | * |
600 | * Locking Note: This function will be called without the rport lock | 646 | * Locking Note: This function will be called without the rport lock |
601 | * held, but it will lock, call an _enter_* function or fc_rport_error | 647 | * held, but it will lock, call an _enter_* function or fc_rport_error |
602 | * and then unlock the rport. | 648 | * and then unlock the rport. |
603 | */ | 649 | */ |
604 | static void fc_rport_prli_resp(struct fc_seq *sp, struct fc_frame *fp, | 650 | static void fc_rport_prli_resp(struct fc_seq *sp, struct fc_frame *fp, |
605 | void *rp_arg) | 651 | void *rdata_arg) |
606 | { | 652 | { |
607 | struct fc_rport *rport = rp_arg; | 653 | struct fc_rport_priv *rdata = rdata_arg; |
608 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
609 | struct { | 654 | struct { |
610 | struct fc_els_prli prli; | 655 | struct fc_els_prli prli; |
611 | struct fc_els_spp spp; | 656 | struct fc_els_spp spp; |
@@ -616,21 +661,24 @@ static void fc_rport_prli_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
616 | 661 | ||
617 | mutex_lock(&rdata->rp_mutex); | 662 | mutex_lock(&rdata->rp_mutex); |
618 | 663 | ||
619 | FC_RPORT_DBG(rport, "Received a PRLI response\n"); | 664 | FC_RPORT_DBG(rdata, "Received a PRLI %s\n", fc_els_resp_type(fp)); |
620 | 665 | ||
621 | if (rdata->rp_state != RPORT_ST_PRLI) { | 666 | if (rdata->rp_state != RPORT_ST_PRLI) { |
622 | FC_RPORT_DBG(rport, "Received a PRLI response, but in state " | 667 | FC_RPORT_DBG(rdata, "Received a PRLI response, but in state " |
623 | "%s\n", fc_rport_state(rport)); | 668 | "%s\n", fc_rport_state(rdata)); |
624 | if (IS_ERR(fp)) | 669 | if (IS_ERR(fp)) |
625 | goto err; | 670 | goto err; |
626 | goto out; | 671 | goto out; |
627 | } | 672 | } |
628 | 673 | ||
629 | if (IS_ERR(fp)) { | 674 | if (IS_ERR(fp)) { |
630 | fc_rport_error_retry(rport, fp); | 675 | fc_rport_error_retry(rdata, fp); |
631 | goto err; | 676 | goto err; |
632 | } | 677 | } |
633 | 678 | ||
679 | /* reinitialize remote port roles */ | ||
680 | rdata->ids.roles = FC_RPORT_ROLE_UNKNOWN; | ||
681 | |||
634 | op = fc_frame_payload_op(fp); | 682 | op = fc_frame_payload_op(fp); |
635 | if (op == ELS_LS_ACC) { | 683 | if (op == ELS_LS_ACC) { |
636 | pp = fc_frame_payload_get(fp, sizeof(*pp)); | 684 | pp = fc_frame_payload_get(fp, sizeof(*pp)); |
@@ -640,90 +688,82 @@ static void fc_rport_prli_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
640 | rdata->flags |= FC_RP_FLAGS_RETRY; | 688 | rdata->flags |= FC_RP_FLAGS_RETRY; |
641 | } | 689 | } |
642 | 690 | ||
643 | rport->supported_classes = FC_COS_CLASS3; | 691 | rdata->supported_classes = FC_COS_CLASS3; |
644 | if (fcp_parm & FCP_SPPF_INIT_FCN) | 692 | if (fcp_parm & FCP_SPPF_INIT_FCN) |
645 | roles |= FC_RPORT_ROLE_FCP_INITIATOR; | 693 | roles |= FC_RPORT_ROLE_FCP_INITIATOR; |
646 | if (fcp_parm & FCP_SPPF_TARG_FCN) | 694 | if (fcp_parm & FCP_SPPF_TARG_FCN) |
647 | roles |= FC_RPORT_ROLE_FCP_TARGET; | 695 | roles |= FC_RPORT_ROLE_FCP_TARGET; |
648 | 696 | ||
649 | rport->roles = roles; | 697 | rdata->ids.roles = roles; |
650 | fc_rport_enter_rtv(rport); | 698 | fc_rport_enter_rtv(rdata); |
651 | 699 | ||
652 | } else { | 700 | } else { |
653 | FC_RPORT_DBG(rport, "Bad ELS response for PRLI command\n"); | 701 | FC_RPORT_DBG(rdata, "Bad ELS response for PRLI command\n"); |
654 | rdata->event = RPORT_EV_FAILED; | 702 | fc_rport_enter_delete(rdata, RPORT_EV_FAILED); |
655 | fc_rport_state_enter(rport, RPORT_ST_NONE); | ||
656 | queue_work(rport_event_queue, &rdata->event_work); | ||
657 | } | 703 | } |
658 | 704 | ||
659 | out: | 705 | out: |
660 | fc_frame_free(fp); | 706 | fc_frame_free(fp); |
661 | err: | 707 | err: |
662 | mutex_unlock(&rdata->rp_mutex); | 708 | mutex_unlock(&rdata->rp_mutex); |
663 | put_device(&rport->dev); | 709 | kref_put(&rdata->kref, rdata->local_port->tt.rport_destroy); |
664 | } | 710 | } |
665 | 711 | ||
666 | /** | 712 | /** |
667 | * fc_rport_logo_resp() - Logout (LOGO) response handler | 713 | * fc_rport_logo_resp() - Logout (LOGO) response handler |
668 | * @sp: current sequence in the LOGO exchange | 714 | * @sp: current sequence in the LOGO exchange |
669 | * @fp: response frame | 715 | * @fp: response frame |
670 | * @rp_arg: Fibre Channel remote port | 716 | * @rdata_arg: private remote port data |
671 | * | 717 | * |
672 | * Locking Note: This function will be called without the rport lock | 718 | * Locking Note: This function will be called without the rport lock |
673 | * held, but it will lock, call an _enter_* function or fc_rport_error | 719 | * held, but it will lock, call an _enter_* function or fc_rport_error |
674 | * and then unlock the rport. | 720 | * and then unlock the rport. |
675 | */ | 721 | */ |
676 | static void fc_rport_logo_resp(struct fc_seq *sp, struct fc_frame *fp, | 722 | static void fc_rport_logo_resp(struct fc_seq *sp, struct fc_frame *fp, |
677 | void *rp_arg) | 723 | void *rdata_arg) |
678 | { | 724 | { |
679 | struct fc_rport *rport = rp_arg; | 725 | struct fc_rport_priv *rdata = rdata_arg; |
680 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
681 | u8 op; | 726 | u8 op; |
682 | 727 | ||
683 | mutex_lock(&rdata->rp_mutex); | 728 | mutex_lock(&rdata->rp_mutex); |
684 | 729 | ||
685 | FC_RPORT_DBG(rport, "Received a LOGO response\n"); | 730 | FC_RPORT_DBG(rdata, "Received a LOGO %s\n", fc_els_resp_type(fp)); |
686 | 731 | ||
687 | if (rdata->rp_state != RPORT_ST_LOGO) { | 732 | if (rdata->rp_state != RPORT_ST_LOGO) { |
688 | FC_RPORT_DBG(rport, "Received a LOGO response, but in state " | 733 | FC_RPORT_DBG(rdata, "Received a LOGO response, but in state " |
689 | "%s\n", fc_rport_state(rport)); | 734 | "%s\n", fc_rport_state(rdata)); |
690 | if (IS_ERR(fp)) | 735 | if (IS_ERR(fp)) |
691 | goto err; | 736 | goto err; |
692 | goto out; | 737 | goto out; |
693 | } | 738 | } |
694 | 739 | ||
695 | if (IS_ERR(fp)) { | 740 | if (IS_ERR(fp)) { |
696 | fc_rport_error_retry(rport, fp); | 741 | fc_rport_error_retry(rdata, fp); |
697 | goto err; | 742 | goto err; |
698 | } | 743 | } |
699 | 744 | ||
700 | op = fc_frame_payload_op(fp); | 745 | op = fc_frame_payload_op(fp); |
701 | if (op == ELS_LS_ACC) { | 746 | if (op != ELS_LS_ACC) |
702 | fc_rport_enter_rtv(rport); | 747 | FC_RPORT_DBG(rdata, "Bad ELS response op %x for LOGO command\n", |
703 | } else { | 748 | op); |
704 | FC_RPORT_DBG(rport, "Bad ELS response for LOGO command\n"); | 749 | fc_rport_enter_delete(rdata, RPORT_EV_LOGO); |
705 | rdata->event = RPORT_EV_LOGO; | ||
706 | fc_rport_state_enter(rport, RPORT_ST_NONE); | ||
707 | queue_work(rport_event_queue, &rdata->event_work); | ||
708 | } | ||
709 | 750 | ||
710 | out: | 751 | out: |
711 | fc_frame_free(fp); | 752 | fc_frame_free(fp); |
712 | err: | 753 | err: |
713 | mutex_unlock(&rdata->rp_mutex); | 754 | mutex_unlock(&rdata->rp_mutex); |
714 | put_device(&rport->dev); | 755 | kref_put(&rdata->kref, rdata->local_port->tt.rport_destroy); |
715 | } | 756 | } |
716 | 757 | ||
717 | /** | 758 | /** |
718 | * fc_rport_enter_prli() - Send Process Login (PRLI) request to peer | 759 | * fc_rport_enter_prli() - Send Process Login (PRLI) request to peer |
719 | * @rport: Fibre Channel remote port to send PRLI to | 760 | * @rdata: private remote port data |
720 | * | 761 | * |
721 | * Locking Note: The rport lock is expected to be held before calling | 762 | * Locking Note: The rport lock is expected to be held before calling |
722 | * this routine. | 763 | * this routine. |
723 | */ | 764 | */ |
724 | static void fc_rport_enter_prli(struct fc_rport *rport) | 765 | static void fc_rport_enter_prli(struct fc_rport_priv *rdata) |
725 | { | 766 | { |
726 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
727 | struct fc_lport *lport = rdata->local_port; | 767 | struct fc_lport *lport = rdata->local_port; |
728 | struct { | 768 | struct { |
729 | struct fc_els_prli prli; | 769 | struct fc_els_prli prli; |
@@ -731,29 +771,38 @@ static void fc_rport_enter_prli(struct fc_rport *rport) | |||
731 | } *pp; | 771 | } *pp; |
732 | struct fc_frame *fp; | 772 | struct fc_frame *fp; |
733 | 773 | ||
734 | FC_RPORT_DBG(rport, "Port entered PRLI state from %s state\n", | 774 | /* |
735 | fc_rport_state(rport)); | 775 | * If the rport is one of the well known addresses |
776 | * we skip PRLI and RTV and go straight to READY. | ||
777 | */ | ||
778 | if (rdata->ids.port_id >= FC_FID_DOM_MGR) { | ||
779 | fc_rport_enter_ready(rdata); | ||
780 | return; | ||
781 | } | ||
782 | |||
783 | FC_RPORT_DBG(rdata, "Port entered PRLI state from %s state\n", | ||
784 | fc_rport_state(rdata)); | ||
736 | 785 | ||
737 | fc_rport_state_enter(rport, RPORT_ST_PRLI); | 786 | fc_rport_state_enter(rdata, RPORT_ST_PRLI); |
738 | 787 | ||
739 | fp = fc_frame_alloc(lport, sizeof(*pp)); | 788 | fp = fc_frame_alloc(lport, sizeof(*pp)); |
740 | if (!fp) { | 789 | if (!fp) { |
741 | fc_rport_error_retry(rport, fp); | 790 | fc_rport_error_retry(rdata, fp); |
742 | return; | 791 | return; |
743 | } | 792 | } |
744 | 793 | ||
745 | if (!lport->tt.elsct_send(lport, rport, fp, ELS_PRLI, | 794 | if (!lport->tt.elsct_send(lport, rdata->ids.port_id, fp, ELS_PRLI, |
746 | fc_rport_prli_resp, rport, lport->e_d_tov)) | 795 | fc_rport_prli_resp, rdata, lport->e_d_tov)) |
747 | fc_rport_error_retry(rport, fp); | 796 | fc_rport_error_retry(rdata, fp); |
748 | else | 797 | else |
749 | get_device(&rport->dev); | 798 | kref_get(&rdata->kref); |
750 | } | 799 | } |
751 | 800 | ||
752 | /** | 801 | /** |
753 | * fc_rport_els_rtv_resp() - Request Timeout Value response handler | 802 | * fc_rport_els_rtv_resp() - Request Timeout Value response handler |
754 | * @sp: current sequence in the RTV exchange | 803 | * @sp: current sequence in the RTV exchange |
755 | * @fp: response frame | 804 | * @fp: response frame |
756 | * @rp_arg: Fibre Channel remote port | 805 | * @rdata_arg: private remote port data |
757 | * | 806 | * |
758 | * Many targets don't seem to support this. | 807 | * Many targets don't seem to support this. |
759 | * | 808 | * |
@@ -762,26 +811,25 @@ static void fc_rport_enter_prli(struct fc_rport *rport) | |||
762 | * and then unlock the rport. | 811 | * and then unlock the rport. |
763 | */ | 812 | */ |
764 | static void fc_rport_rtv_resp(struct fc_seq *sp, struct fc_frame *fp, | 813 | static void fc_rport_rtv_resp(struct fc_seq *sp, struct fc_frame *fp, |
765 | void *rp_arg) | 814 | void *rdata_arg) |
766 | { | 815 | { |
767 | struct fc_rport *rport = rp_arg; | 816 | struct fc_rport_priv *rdata = rdata_arg; |
768 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
769 | u8 op; | 817 | u8 op; |
770 | 818 | ||
771 | mutex_lock(&rdata->rp_mutex); | 819 | mutex_lock(&rdata->rp_mutex); |
772 | 820 | ||
773 | FC_RPORT_DBG(rport, "Received a RTV response\n"); | 821 | FC_RPORT_DBG(rdata, "Received a RTV %s\n", fc_els_resp_type(fp)); |
774 | 822 | ||
775 | if (rdata->rp_state != RPORT_ST_RTV) { | 823 | if (rdata->rp_state != RPORT_ST_RTV) { |
776 | FC_RPORT_DBG(rport, "Received a RTV response, but in state " | 824 | FC_RPORT_DBG(rdata, "Received a RTV response, but in state " |
777 | "%s\n", fc_rport_state(rport)); | 825 | "%s\n", fc_rport_state(rdata)); |
778 | if (IS_ERR(fp)) | 826 | if (IS_ERR(fp)) |
779 | goto err; | 827 | goto err; |
780 | goto out; | 828 | goto out; |
781 | } | 829 | } |
782 | 830 | ||
783 | if (IS_ERR(fp)) { | 831 | if (IS_ERR(fp)) { |
784 | fc_rport_error(rport, fp); | 832 | fc_rport_error(rdata, fp); |
785 | goto err; | 833 | goto err; |
786 | } | 834 | } |
787 | 835 | ||
@@ -807,184 +855,376 @@ static void fc_rport_rtv_resp(struct fc_seq *sp, struct fc_frame *fp, | |||
807 | } | 855 | } |
808 | } | 856 | } |
809 | 857 | ||
810 | fc_rport_enter_ready(rport); | 858 | fc_rport_enter_ready(rdata); |
811 | 859 | ||
812 | out: | 860 | out: |
813 | fc_frame_free(fp); | 861 | fc_frame_free(fp); |
814 | err: | 862 | err: |
815 | mutex_unlock(&rdata->rp_mutex); | 863 | mutex_unlock(&rdata->rp_mutex); |
816 | put_device(&rport->dev); | 864 | kref_put(&rdata->kref, rdata->local_port->tt.rport_destroy); |
817 | } | 865 | } |
818 | 866 | ||
819 | /** | 867 | /** |
820 | * fc_rport_enter_rtv() - Send Request Timeout Value (RTV) request to peer | 868 | * fc_rport_enter_rtv() - Send Request Timeout Value (RTV) request to peer |
821 | * @rport: Fibre Channel remote port to send RTV to | 869 | * @rdata: private remote port data |
822 | * | 870 | * |
823 | * Locking Note: The rport lock is expected to be held before calling | 871 | * Locking Note: The rport lock is expected to be held before calling |
824 | * this routine. | 872 | * this routine. |
825 | */ | 873 | */ |
826 | static void fc_rport_enter_rtv(struct fc_rport *rport) | 874 | static void fc_rport_enter_rtv(struct fc_rport_priv *rdata) |
827 | { | 875 | { |
828 | struct fc_frame *fp; | 876 | struct fc_frame *fp; |
829 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
830 | struct fc_lport *lport = rdata->local_port; | 877 | struct fc_lport *lport = rdata->local_port; |
831 | 878 | ||
832 | FC_RPORT_DBG(rport, "Port entered RTV state from %s state\n", | 879 | FC_RPORT_DBG(rdata, "Port entered RTV state from %s state\n", |
833 | fc_rport_state(rport)); | 880 | fc_rport_state(rdata)); |
834 | 881 | ||
835 | fc_rport_state_enter(rport, RPORT_ST_RTV); | 882 | fc_rport_state_enter(rdata, RPORT_ST_RTV); |
836 | 883 | ||
837 | fp = fc_frame_alloc(lport, sizeof(struct fc_els_rtv)); | 884 | fp = fc_frame_alloc(lport, sizeof(struct fc_els_rtv)); |
838 | if (!fp) { | 885 | if (!fp) { |
839 | fc_rport_error_retry(rport, fp); | 886 | fc_rport_error_retry(rdata, fp); |
840 | return; | 887 | return; |
841 | } | 888 | } |
842 | 889 | ||
843 | if (!lport->tt.elsct_send(lport, rport, fp, ELS_RTV, | 890 | if (!lport->tt.elsct_send(lport, rdata->ids.port_id, fp, ELS_RTV, |
844 | fc_rport_rtv_resp, rport, lport->e_d_tov)) | 891 | fc_rport_rtv_resp, rdata, lport->e_d_tov)) |
845 | fc_rport_error_retry(rport, fp); | 892 | fc_rport_error_retry(rdata, fp); |
846 | else | 893 | else |
847 | get_device(&rport->dev); | 894 | kref_get(&rdata->kref); |
848 | } | 895 | } |
849 | 896 | ||
850 | /** | 897 | /** |
851 | * fc_rport_enter_logo() - Send Logout (LOGO) request to peer | 898 | * fc_rport_enter_logo() - Send Logout (LOGO) request to peer |
852 | * @rport: Fibre Channel remote port to send LOGO to | 899 | * @rdata: private remote port data |
853 | * | 900 | * |
854 | * Locking Note: The rport lock is expected to be held before calling | 901 | * Locking Note: The rport lock is expected to be held before calling |
855 | * this routine. | 902 | * this routine. |
856 | */ | 903 | */ |
857 | static void fc_rport_enter_logo(struct fc_rport *rport) | 904 | static void fc_rport_enter_logo(struct fc_rport_priv *rdata) |
858 | { | 905 | { |
859 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
860 | struct fc_lport *lport = rdata->local_port; | 906 | struct fc_lport *lport = rdata->local_port; |
861 | struct fc_frame *fp; | 907 | struct fc_frame *fp; |
862 | 908 | ||
863 | FC_RPORT_DBG(rport, "Port entered LOGO state from %s state\n", | 909 | FC_RPORT_DBG(rdata, "Port entered LOGO state from %s state\n", |
864 | fc_rport_state(rport)); | 910 | fc_rport_state(rdata)); |
865 | 911 | ||
866 | fc_rport_state_enter(rport, RPORT_ST_LOGO); | 912 | fc_rport_state_enter(rdata, RPORT_ST_LOGO); |
867 | 913 | ||
868 | fp = fc_frame_alloc(lport, sizeof(struct fc_els_logo)); | 914 | fp = fc_frame_alloc(lport, sizeof(struct fc_els_logo)); |
869 | if (!fp) { | 915 | if (!fp) { |
870 | fc_rport_error_retry(rport, fp); | 916 | fc_rport_error_retry(rdata, fp); |
871 | return; | 917 | return; |
872 | } | 918 | } |
873 | 919 | ||
874 | if (!lport->tt.elsct_send(lport, rport, fp, ELS_LOGO, | 920 | if (!lport->tt.elsct_send(lport, rdata->ids.port_id, fp, ELS_LOGO, |
875 | fc_rport_logo_resp, rport, lport->e_d_tov)) | 921 | fc_rport_logo_resp, rdata, lport->e_d_tov)) |
876 | fc_rport_error_retry(rport, fp); | 922 | fc_rport_error_retry(rdata, fp); |
877 | else | 923 | else |
878 | get_device(&rport->dev); | 924 | kref_get(&rdata->kref); |
879 | } | 925 | } |
880 | 926 | ||
881 | |||
882 | /** | 927 | /** |
883 | * fc_rport_recv_req() - Receive a request from a rport | 928 | * fc_rport_els_adisc_resp() - Address Discovery response handler |
884 | * @sp: current sequence in the PLOGI exchange | 929 | * @sp: current sequence in the ADISC exchange |
885 | * @fp: response frame | 930 | * @fp: response frame |
886 | * @rp_arg: Fibre Channel remote port | 931 | * @rdata_arg: remote port private. |
887 | * | 932 | * |
888 | * Locking Note: Called without the rport lock held. This | 933 | * Locking Note: This function will be called without the rport lock |
889 | * function will hold the rport lock, call an _enter_* | 934 | * held, but it will lock, call an _enter_* function or fc_rport_error |
890 | * function and then unlock the rport. | 935 | * and then unlock the rport. |
891 | */ | 936 | */ |
892 | void fc_rport_recv_req(struct fc_seq *sp, struct fc_frame *fp, | 937 | static void fc_rport_adisc_resp(struct fc_seq *sp, struct fc_frame *fp, |
893 | struct fc_rport *rport) | 938 | void *rdata_arg) |
939 | { | ||
940 | struct fc_rport_priv *rdata = rdata_arg; | ||
941 | struct fc_els_adisc *adisc; | ||
942 | u8 op; | ||
943 | |||
944 | mutex_lock(&rdata->rp_mutex); | ||
945 | |||
946 | FC_RPORT_DBG(rdata, "Received a ADISC response\n"); | ||
947 | |||
948 | if (rdata->rp_state != RPORT_ST_ADISC) { | ||
949 | FC_RPORT_DBG(rdata, "Received a ADISC resp but in state %s\n", | ||
950 | fc_rport_state(rdata)); | ||
951 | if (IS_ERR(fp)) | ||
952 | goto err; | ||
953 | goto out; | ||
954 | } | ||
955 | |||
956 | if (IS_ERR(fp)) { | ||
957 | fc_rport_error(rdata, fp); | ||
958 | goto err; | ||
959 | } | ||
960 | |||
961 | /* | ||
962 | * If address verification failed. Consider us logged out of the rport. | ||
963 | * Since the rport is still in discovery, we want to be | ||
964 | * logged in, so go to PLOGI state. Otherwise, go back to READY. | ||
965 | */ | ||
966 | op = fc_frame_payload_op(fp); | ||
967 | adisc = fc_frame_payload_get(fp, sizeof(*adisc)); | ||
968 | if (op != ELS_LS_ACC || !adisc || | ||
969 | ntoh24(adisc->adisc_port_id) != rdata->ids.port_id || | ||
970 | get_unaligned_be64(&adisc->adisc_wwpn) != rdata->ids.port_name || | ||
971 | get_unaligned_be64(&adisc->adisc_wwnn) != rdata->ids.node_name) { | ||
972 | FC_RPORT_DBG(rdata, "ADISC error or mismatch\n"); | ||
973 | fc_rport_enter_plogi(rdata); | ||
974 | } else { | ||
975 | FC_RPORT_DBG(rdata, "ADISC OK\n"); | ||
976 | fc_rport_enter_ready(rdata); | ||
977 | } | ||
978 | out: | ||
979 | fc_frame_free(fp); | ||
980 | err: | ||
981 | mutex_unlock(&rdata->rp_mutex); | ||
982 | kref_put(&rdata->kref, rdata->local_port->tt.rport_destroy); | ||
983 | } | ||
984 | |||
985 | /** | ||
986 | * fc_rport_enter_adisc() - Send Address Discover (ADISC) request to peer | ||
987 | * @rdata: remote port private data | ||
988 | * | ||
989 | * Locking Note: The rport lock is expected to be held before calling | ||
990 | * this routine. | ||
991 | */ | ||
992 | static void fc_rport_enter_adisc(struct fc_rport_priv *rdata) | ||
894 | { | 993 | { |
895 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
896 | struct fc_lport *lport = rdata->local_port; | 994 | struct fc_lport *lport = rdata->local_port; |
995 | struct fc_frame *fp; | ||
996 | |||
997 | FC_RPORT_DBG(rdata, "sending ADISC from %s state\n", | ||
998 | fc_rport_state(rdata)); | ||
897 | 999 | ||
1000 | fc_rport_state_enter(rdata, RPORT_ST_ADISC); | ||
1001 | |||
1002 | fp = fc_frame_alloc(lport, sizeof(struct fc_els_adisc)); | ||
1003 | if (!fp) { | ||
1004 | fc_rport_error_retry(rdata, fp); | ||
1005 | return; | ||
1006 | } | ||
1007 | if (!lport->tt.elsct_send(lport, rdata->ids.port_id, fp, ELS_ADISC, | ||
1008 | fc_rport_adisc_resp, rdata, lport->e_d_tov)) | ||
1009 | fc_rport_error_retry(rdata, fp); | ||
1010 | else | ||
1011 | kref_get(&rdata->kref); | ||
1012 | } | ||
1013 | |||
1014 | /** | ||
1015 | * fc_rport_recv_adisc_req() - Handle incoming Address Discovery (ADISC) Request | ||
1016 | * @rdata: remote port private | ||
1017 | * @sp: current sequence in the ADISC exchange | ||
1018 | * @in_fp: ADISC request frame | ||
1019 | * | ||
1020 | * Locking Note: Called with the lport and rport locks held. | ||
1021 | */ | ||
1022 | static void fc_rport_recv_adisc_req(struct fc_rport_priv *rdata, | ||
1023 | struct fc_seq *sp, struct fc_frame *in_fp) | ||
1024 | { | ||
1025 | struct fc_lport *lport = rdata->local_port; | ||
1026 | struct fc_frame *fp; | ||
1027 | struct fc_exch *ep = fc_seq_exch(sp); | ||
1028 | struct fc_els_adisc *adisc; | ||
1029 | struct fc_seq_els_data rjt_data; | ||
1030 | u32 f_ctl; | ||
1031 | |||
1032 | FC_RPORT_DBG(rdata, "Received ADISC request\n"); | ||
1033 | |||
1034 | adisc = fc_frame_payload_get(in_fp, sizeof(*adisc)); | ||
1035 | if (!adisc) { | ||
1036 | rjt_data.fp = NULL; | ||
1037 | rjt_data.reason = ELS_RJT_PROT; | ||
1038 | rjt_data.explan = ELS_EXPL_INV_LEN; | ||
1039 | lport->tt.seq_els_rsp_send(sp, ELS_LS_RJT, &rjt_data); | ||
1040 | goto drop; | ||
1041 | } | ||
1042 | |||
1043 | fp = fc_frame_alloc(lport, sizeof(*adisc)); | ||
1044 | if (!fp) | ||
1045 | goto drop; | ||
1046 | fc_adisc_fill(lport, fp); | ||
1047 | adisc = fc_frame_payload_get(fp, sizeof(*adisc)); | ||
1048 | adisc->adisc_cmd = ELS_LS_ACC; | ||
1049 | sp = lport->tt.seq_start_next(sp); | ||
1050 | f_ctl = FC_FC_EX_CTX | FC_FC_LAST_SEQ | FC_FC_END_SEQ | FC_FC_SEQ_INIT; | ||
1051 | fc_fill_fc_hdr(fp, FC_RCTL_ELS_REP, ep->did, ep->sid, | ||
1052 | FC_TYPE_ELS, f_ctl, 0); | ||
1053 | lport->tt.seq_send(lport, sp, fp); | ||
1054 | drop: | ||
1055 | fc_frame_free(in_fp); | ||
1056 | } | ||
1057 | |||
1058 | /** | ||
1059 | * fc_rport_recv_els_req() - handle a validated ELS request. | ||
1060 | * @lport: Fibre Channel local port | ||
1061 | * @sp: current sequence in the PLOGI exchange | ||
1062 | * @fp: response frame | ||
1063 | * | ||
1064 | * Handle incoming ELS requests that require port login. | ||
1065 | * The ELS opcode has already been validated by the caller. | ||
1066 | * | ||
1067 | * Locking Note: Called with the lport lock held. | ||
1068 | */ | ||
1069 | static void fc_rport_recv_els_req(struct fc_lport *lport, | ||
1070 | struct fc_seq *sp, struct fc_frame *fp) | ||
1071 | { | ||
1072 | struct fc_rport_priv *rdata; | ||
898 | struct fc_frame_header *fh; | 1073 | struct fc_frame_header *fh; |
899 | struct fc_seq_els_data els_data; | 1074 | struct fc_seq_els_data els_data; |
900 | u8 op; | ||
901 | |||
902 | mutex_lock(&rdata->rp_mutex); | ||
903 | 1075 | ||
904 | els_data.fp = NULL; | 1076 | els_data.fp = NULL; |
905 | els_data.explan = ELS_EXPL_NONE; | 1077 | els_data.reason = ELS_RJT_UNAB; |
906 | els_data.reason = ELS_RJT_NONE; | 1078 | els_data.explan = ELS_EXPL_PLOGI_REQD; |
907 | 1079 | ||
908 | fh = fc_frame_header_get(fp); | 1080 | fh = fc_frame_header_get(fp); |
909 | 1081 | ||
910 | if (fh->fh_r_ctl == FC_RCTL_ELS_REQ && fh->fh_type == FC_TYPE_ELS) { | 1082 | mutex_lock(&lport->disc.disc_mutex); |
911 | op = fc_frame_payload_op(fp); | 1083 | rdata = lport->tt.rport_lookup(lport, ntoh24(fh->fh_s_id)); |
912 | switch (op) { | 1084 | if (!rdata) { |
913 | case ELS_PLOGI: | 1085 | mutex_unlock(&lport->disc.disc_mutex); |
914 | fc_rport_recv_plogi_req(rport, sp, fp); | 1086 | goto reject; |
915 | break; | 1087 | } |
916 | case ELS_PRLI: | 1088 | mutex_lock(&rdata->rp_mutex); |
917 | fc_rport_recv_prli_req(rport, sp, fp); | 1089 | mutex_unlock(&lport->disc.disc_mutex); |
918 | break; | 1090 | |
919 | case ELS_PRLO: | 1091 | switch (rdata->rp_state) { |
920 | fc_rport_recv_prlo_req(rport, sp, fp); | 1092 | case RPORT_ST_PRLI: |
921 | break; | 1093 | case RPORT_ST_RTV: |
922 | case ELS_LOGO: | 1094 | case RPORT_ST_READY: |
923 | fc_rport_recv_logo_req(rport, sp, fp); | 1095 | case RPORT_ST_ADISC: |
924 | break; | 1096 | break; |
925 | case ELS_RRQ: | 1097 | default: |
926 | els_data.fp = fp; | 1098 | mutex_unlock(&rdata->rp_mutex); |
927 | lport->tt.seq_els_rsp_send(sp, ELS_RRQ, &els_data); | 1099 | goto reject; |
928 | break; | 1100 | } |
929 | case ELS_REC: | 1101 | |
930 | els_data.fp = fp; | 1102 | switch (fc_frame_payload_op(fp)) { |
931 | lport->tt.seq_els_rsp_send(sp, ELS_REC, &els_data); | 1103 | case ELS_PRLI: |
932 | break; | 1104 | fc_rport_recv_prli_req(rdata, sp, fp); |
933 | default: | 1105 | break; |
934 | els_data.reason = ELS_RJT_UNSUP; | 1106 | case ELS_PRLO: |
935 | lport->tt.seq_els_rsp_send(sp, ELS_LS_RJT, &els_data); | 1107 | fc_rport_recv_prlo_req(rdata, sp, fp); |
936 | break; | 1108 | break; |
937 | } | 1109 | case ELS_ADISC: |
1110 | fc_rport_recv_adisc_req(rdata, sp, fp); | ||
1111 | break; | ||
1112 | case ELS_RRQ: | ||
1113 | els_data.fp = fp; | ||
1114 | lport->tt.seq_els_rsp_send(sp, ELS_RRQ, &els_data); | ||
1115 | break; | ||
1116 | case ELS_REC: | ||
1117 | els_data.fp = fp; | ||
1118 | lport->tt.seq_els_rsp_send(sp, ELS_REC, &els_data); | ||
1119 | break; | ||
1120 | default: | ||
1121 | fc_frame_free(fp); /* can't happen */ | ||
1122 | break; | ||
938 | } | 1123 | } |
939 | 1124 | ||
940 | mutex_unlock(&rdata->rp_mutex); | 1125 | mutex_unlock(&rdata->rp_mutex); |
1126 | return; | ||
1127 | |||
1128 | reject: | ||
1129 | lport->tt.seq_els_rsp_send(sp, ELS_LS_RJT, &els_data); | ||
1130 | fc_frame_free(fp); | ||
1131 | } | ||
1132 | |||
1133 | /** | ||
1134 | * fc_rport_recv_req() - Handle a received ELS request from a rport | ||
1135 | * @sp: current sequence in the PLOGI exchange | ||
1136 | * @fp: response frame | ||
1137 | * @lport: Fibre Channel local port | ||
1138 | * | ||
1139 | * Locking Note: Called with the lport lock held. | ||
1140 | */ | ||
1141 | void fc_rport_recv_req(struct fc_seq *sp, struct fc_frame *fp, | ||
1142 | struct fc_lport *lport) | ||
1143 | { | ||
1144 | struct fc_seq_els_data els_data; | ||
1145 | |||
1146 | /* | ||
1147 | * Handle PLOGI and LOGO requests separately, since they | ||
1148 | * don't require prior login. | ||
1149 | * Check for unsupported opcodes first and reject them. | ||
1150 | * For some ops, it would be incorrect to reject with "PLOGI required". | ||
1151 | */ | ||
1152 | switch (fc_frame_payload_op(fp)) { | ||
1153 | case ELS_PLOGI: | ||
1154 | fc_rport_recv_plogi_req(lport, sp, fp); | ||
1155 | break; | ||
1156 | case ELS_LOGO: | ||
1157 | fc_rport_recv_logo_req(lport, sp, fp); | ||
1158 | break; | ||
1159 | case ELS_PRLI: | ||
1160 | case ELS_PRLO: | ||
1161 | case ELS_ADISC: | ||
1162 | case ELS_RRQ: | ||
1163 | case ELS_REC: | ||
1164 | fc_rport_recv_els_req(lport, sp, fp); | ||
1165 | break; | ||
1166 | default: | ||
1167 | fc_frame_free(fp); | ||
1168 | els_data.fp = NULL; | ||
1169 | els_data.reason = ELS_RJT_UNSUP; | ||
1170 | els_data.explan = ELS_EXPL_NONE; | ||
1171 | lport->tt.seq_els_rsp_send(sp, ELS_LS_RJT, &els_data); | ||
1172 | break; | ||
1173 | } | ||
941 | } | 1174 | } |
942 | 1175 | ||
943 | /** | 1176 | /** |
944 | * fc_rport_recv_plogi_req() - Handle incoming Port Login (PLOGI) request | 1177 | * fc_rport_recv_plogi_req() - Handle incoming Port Login (PLOGI) request |
945 | * @rport: Fibre Channel remote port that initiated PLOGI | 1178 | * @lport: local port |
946 | * @sp: current sequence in the PLOGI exchange | 1179 | * @sp: current sequence in the PLOGI exchange |
947 | * @fp: PLOGI request frame | 1180 | * @fp: PLOGI request frame |
948 | * | 1181 | * |
949 | * Locking Note: The rport lock is exected to be held before calling | 1182 | * Locking Note: The rport lock is held before calling this function. |
950 | * this function. | ||
951 | */ | 1183 | */ |
952 | static void fc_rport_recv_plogi_req(struct fc_rport *rport, | 1184 | static void fc_rport_recv_plogi_req(struct fc_lport *lport, |
953 | struct fc_seq *sp, struct fc_frame *rx_fp) | 1185 | struct fc_seq *sp, struct fc_frame *rx_fp) |
954 | { | 1186 | { |
955 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | 1187 | struct fc_disc *disc; |
956 | struct fc_lport *lport = rdata->local_port; | 1188 | struct fc_rport_priv *rdata; |
957 | struct fc_frame *fp = rx_fp; | 1189 | struct fc_frame *fp = rx_fp; |
958 | struct fc_exch *ep; | 1190 | struct fc_exch *ep; |
959 | struct fc_frame_header *fh; | 1191 | struct fc_frame_header *fh; |
960 | struct fc_els_flogi *pl; | 1192 | struct fc_els_flogi *pl; |
961 | struct fc_seq_els_data rjt_data; | 1193 | struct fc_seq_els_data rjt_data; |
962 | u32 sid; | 1194 | u32 sid, f_ctl; |
963 | u64 wwpn; | ||
964 | u64 wwnn; | ||
965 | enum fc_els_rjt_reason reject = 0; | ||
966 | u32 f_ctl; | ||
967 | rjt_data.fp = NULL; | ||
968 | 1195 | ||
1196 | rjt_data.fp = NULL; | ||
969 | fh = fc_frame_header_get(fp); | 1197 | fh = fc_frame_header_get(fp); |
1198 | sid = ntoh24(fh->fh_s_id); | ||
970 | 1199 | ||
971 | FC_RPORT_DBG(rport, "Received PLOGI request while in state %s\n", | 1200 | FC_RPORT_ID_DBG(lport, sid, "Received PLOGI request\n"); |
972 | fc_rport_state(rport)); | ||
973 | 1201 | ||
974 | sid = ntoh24(fh->fh_s_id); | ||
975 | pl = fc_frame_payload_get(fp, sizeof(*pl)); | 1202 | pl = fc_frame_payload_get(fp, sizeof(*pl)); |
976 | if (!pl) { | 1203 | if (!pl) { |
977 | FC_RPORT_DBG(rport, "Received PLOGI too short\n"); | 1204 | FC_RPORT_ID_DBG(lport, sid, "Received PLOGI too short\n"); |
978 | WARN_ON(1); | 1205 | rjt_data.reason = ELS_RJT_PROT; |
979 | /* XXX TBD: send reject? */ | 1206 | rjt_data.explan = ELS_EXPL_INV_LEN; |
980 | fc_frame_free(fp); | 1207 | goto reject; |
981 | return; | 1208 | } |
1209 | |||
1210 | disc = &lport->disc; | ||
1211 | mutex_lock(&disc->disc_mutex); | ||
1212 | rdata = lport->tt.rport_create(lport, sid); | ||
1213 | if (!rdata) { | ||
1214 | mutex_unlock(&disc->disc_mutex); | ||
1215 | rjt_data.reason = ELS_RJT_UNAB; | ||
1216 | rjt_data.explan = ELS_EXPL_INSUF_RES; | ||
1217 | goto reject; | ||
982 | } | 1218 | } |
983 | wwpn = get_unaligned_be64(&pl->fl_wwpn); | 1219 | |
984 | wwnn = get_unaligned_be64(&pl->fl_wwnn); | 1220 | mutex_lock(&rdata->rp_mutex); |
1221 | mutex_unlock(&disc->disc_mutex); | ||
1222 | |||
1223 | rdata->ids.port_name = get_unaligned_be64(&pl->fl_wwpn); | ||
1224 | rdata->ids.node_name = get_unaligned_be64(&pl->fl_wwnn); | ||
985 | 1225 | ||
986 | /* | 1226 | /* |
987 | * If the session was just created, possibly due to the incoming PLOGI, | 1227 | * If the rport was just created, possibly due to the incoming PLOGI, |
988 | * set the state appropriately and accept the PLOGI. | 1228 | * set the state appropriately and accept the PLOGI. |
989 | * | 1229 | * |
990 | * If we had also sent a PLOGI, and if the received PLOGI is from a | 1230 | * If we had also sent a PLOGI, and if the received PLOGI is from a |
@@ -996,86 +1236,76 @@ static void fc_rport_recv_plogi_req(struct fc_rport *rport, | |||
996 | */ | 1236 | */ |
997 | switch (rdata->rp_state) { | 1237 | switch (rdata->rp_state) { |
998 | case RPORT_ST_INIT: | 1238 | case RPORT_ST_INIT: |
999 | FC_RPORT_DBG(rport, "Received PLOGI, wwpn %llx state INIT " | 1239 | FC_RPORT_DBG(rdata, "Received PLOGI in INIT state\n"); |
1000 | "- reject\n", (unsigned long long)wwpn); | ||
1001 | reject = ELS_RJT_UNSUP; | ||
1002 | break; | 1240 | break; |
1003 | case RPORT_ST_PLOGI: | 1241 | case RPORT_ST_PLOGI: |
1004 | FC_RPORT_DBG(rport, "Received PLOGI in PLOGI state %d\n", | 1242 | FC_RPORT_DBG(rdata, "Received PLOGI in PLOGI state\n"); |
1005 | rdata->rp_state); | 1243 | if (rdata->ids.port_name < lport->wwpn) { |
1006 | if (wwpn < lport->wwpn) | 1244 | mutex_unlock(&rdata->rp_mutex); |
1007 | reject = ELS_RJT_INPROG; | 1245 | rjt_data.reason = ELS_RJT_INPROG; |
1246 | rjt_data.explan = ELS_EXPL_NONE; | ||
1247 | goto reject; | ||
1248 | } | ||
1008 | break; | 1249 | break; |
1009 | case RPORT_ST_PRLI: | 1250 | case RPORT_ST_PRLI: |
1010 | case RPORT_ST_READY: | 1251 | case RPORT_ST_READY: |
1011 | FC_RPORT_DBG(rport, "Received PLOGI in logged-in state %d " | 1252 | case RPORT_ST_ADISC: |
1253 | FC_RPORT_DBG(rdata, "Received PLOGI in logged-in state %d " | ||
1012 | "- ignored for now\n", rdata->rp_state); | 1254 | "- ignored for now\n", rdata->rp_state); |
1013 | /* XXX TBD - should reset */ | 1255 | /* XXX TBD - should reset */ |
1014 | break; | 1256 | break; |
1015 | case RPORT_ST_NONE: | 1257 | case RPORT_ST_DELETE: |
1016 | default: | 1258 | default: |
1017 | FC_RPORT_DBG(rport, "Received PLOGI in unexpected " | 1259 | FC_RPORT_DBG(rdata, "Received PLOGI in unexpected state %d\n", |
1018 | "state %d\n", rdata->rp_state); | 1260 | rdata->rp_state); |
1019 | fc_frame_free(fp); | 1261 | fc_frame_free(rx_fp); |
1020 | return; | 1262 | goto out; |
1021 | break; | ||
1022 | } | 1263 | } |
1023 | 1264 | ||
1024 | if (reject) { | 1265 | /* |
1025 | rjt_data.reason = reject; | 1266 | * Get session payload size from incoming PLOGI. |
1026 | rjt_data.explan = ELS_EXPL_NONE; | 1267 | */ |
1027 | lport->tt.seq_els_rsp_send(sp, ELS_LS_RJT, &rjt_data); | 1268 | rdata->maxframe_size = fc_plogi_get_maxframe(pl, lport->mfs); |
1028 | fc_frame_free(fp); | 1269 | fc_frame_free(rx_fp); |
1029 | } else { | 1270 | |
1030 | fp = fc_frame_alloc(lport, sizeof(*pl)); | 1271 | /* |
1031 | if (fp == NULL) { | 1272 | * Send LS_ACC. If this fails, the originator should retry. |
1032 | fp = rx_fp; | 1273 | */ |
1033 | rjt_data.reason = ELS_RJT_UNAB; | 1274 | sp = lport->tt.seq_start_next(sp); |
1034 | rjt_data.explan = ELS_EXPL_NONE; | 1275 | if (!sp) |
1035 | lport->tt.seq_els_rsp_send(sp, ELS_LS_RJT, &rjt_data); | 1276 | goto out; |
1036 | fc_frame_free(fp); | 1277 | fp = fc_frame_alloc(lport, sizeof(*pl)); |
1037 | } else { | 1278 | if (!fp) |
1038 | sp = lport->tt.seq_start_next(sp); | 1279 | goto out; |
1039 | WARN_ON(!sp); | 1280 | |
1040 | fc_rport_set_name(rport, wwpn, wwnn); | 1281 | fc_plogi_fill(lport, fp, ELS_LS_ACC); |
1041 | 1282 | f_ctl = FC_FC_EX_CTX | FC_FC_LAST_SEQ | FC_FC_END_SEQ | FC_FC_SEQ_INIT; | |
1042 | /* | 1283 | ep = fc_seq_exch(sp); |
1043 | * Get session payload size from incoming PLOGI. | 1284 | fc_fill_fc_hdr(fp, FC_RCTL_ELS_REP, ep->did, ep->sid, |
1044 | */ | 1285 | FC_TYPE_ELS, f_ctl, 0); |
1045 | rport->maxframe_size = | 1286 | lport->tt.seq_send(lport, sp, fp); |
1046 | fc_plogi_get_maxframe(pl, lport->mfs); | 1287 | fc_rport_enter_prli(rdata); |
1047 | fc_frame_free(rx_fp); | 1288 | out: |
1048 | fc_plogi_fill(lport, fp, ELS_LS_ACC); | 1289 | mutex_unlock(&rdata->rp_mutex); |
1049 | 1290 | return; | |
1050 | /* | 1291 | |
1051 | * Send LS_ACC. If this fails, | 1292 | reject: |
1052 | * the originator should retry. | 1293 | lport->tt.seq_els_rsp_send(sp, ELS_LS_RJT, &rjt_data); |
1053 | */ | 1294 | fc_frame_free(fp); |
1054 | f_ctl = FC_FC_EX_CTX | FC_FC_LAST_SEQ; | ||
1055 | f_ctl |= FC_FC_END_SEQ | FC_FC_SEQ_INIT; | ||
1056 | ep = fc_seq_exch(sp); | ||
1057 | fc_fill_fc_hdr(fp, FC_RCTL_ELS_REP, ep->did, ep->sid, | ||
1058 | FC_TYPE_ELS, f_ctl, 0); | ||
1059 | lport->tt.seq_send(lport, sp, fp); | ||
1060 | if (rdata->rp_state == RPORT_ST_PLOGI) | ||
1061 | fc_rport_enter_prli(rport); | ||
1062 | } | ||
1063 | } | ||
1064 | } | 1295 | } |
1065 | 1296 | ||
1066 | /** | 1297 | /** |
1067 | * fc_rport_recv_prli_req() - Handle incoming Process Login (PRLI) request | 1298 | * fc_rport_recv_prli_req() - Handle incoming Process Login (PRLI) request |
1068 | * @rport: Fibre Channel remote port that initiated PRLI | 1299 | * @rdata: private remote port data |
1069 | * @sp: current sequence in the PRLI exchange | 1300 | * @sp: current sequence in the PRLI exchange |
1070 | * @fp: PRLI request frame | 1301 | * @fp: PRLI request frame |
1071 | * | 1302 | * |
1072 | * Locking Note: The rport lock is exected to be held before calling | 1303 | * Locking Note: The rport lock is exected to be held before calling |
1073 | * this function. | 1304 | * this function. |
1074 | */ | 1305 | */ |
1075 | static void fc_rport_recv_prli_req(struct fc_rport *rport, | 1306 | static void fc_rport_recv_prli_req(struct fc_rport_priv *rdata, |
1076 | struct fc_seq *sp, struct fc_frame *rx_fp) | 1307 | struct fc_seq *sp, struct fc_frame *rx_fp) |
1077 | { | 1308 | { |
1078 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
1079 | struct fc_lport *lport = rdata->local_port; | 1309 | struct fc_lport *lport = rdata->local_port; |
1080 | struct fc_exch *ep; | 1310 | struct fc_exch *ep; |
1081 | struct fc_frame *fp; | 1311 | struct fc_frame *fp; |
@@ -1099,12 +1329,14 @@ static void fc_rport_recv_prli_req(struct fc_rport *rport, | |||
1099 | 1329 | ||
1100 | fh = fc_frame_header_get(rx_fp); | 1330 | fh = fc_frame_header_get(rx_fp); |
1101 | 1331 | ||
1102 | FC_RPORT_DBG(rport, "Received PRLI request while in state %s\n", | 1332 | FC_RPORT_DBG(rdata, "Received PRLI request while in state %s\n", |
1103 | fc_rport_state(rport)); | 1333 | fc_rport_state(rdata)); |
1104 | 1334 | ||
1105 | switch (rdata->rp_state) { | 1335 | switch (rdata->rp_state) { |
1106 | case RPORT_ST_PRLI: | 1336 | case RPORT_ST_PRLI: |
1337 | case RPORT_ST_RTV: | ||
1107 | case RPORT_ST_READY: | 1338 | case RPORT_ST_READY: |
1339 | case RPORT_ST_ADISC: | ||
1108 | reason = ELS_RJT_NONE; | 1340 | reason = ELS_RJT_NONE; |
1109 | break; | 1341 | break; |
1110 | default: | 1342 | default: |
@@ -1149,6 +1381,9 @@ static void fc_rport_recv_prli_req(struct fc_rport *rport, | |||
1149 | pp->prli.prli_len = htons(len); | 1381 | pp->prli.prli_len = htons(len); |
1150 | len -= sizeof(struct fc_els_prli); | 1382 | len -= sizeof(struct fc_els_prli); |
1151 | 1383 | ||
1384 | /* reinitialize remote port roles */ | ||
1385 | rdata->ids.roles = FC_RPORT_ROLE_UNKNOWN; | ||
1386 | |||
1152 | /* | 1387 | /* |
1153 | * Go through all the service parameter pages and build | 1388 | * Go through all the service parameter pages and build |
1154 | * response. If plen indicates longer SPP than standard, | 1389 | * response. If plen indicates longer SPP than standard, |
@@ -1169,12 +1404,12 @@ static void fc_rport_recv_prli_req(struct fc_rport *rport, | |||
1169 | fcp_parm = ntohl(rspp->spp_params); | 1404 | fcp_parm = ntohl(rspp->spp_params); |
1170 | if (fcp_parm * FCP_SPPF_RETRY) | 1405 | if (fcp_parm * FCP_SPPF_RETRY) |
1171 | rdata->flags |= FC_RP_FLAGS_RETRY; | 1406 | rdata->flags |= FC_RP_FLAGS_RETRY; |
1172 | rport->supported_classes = FC_COS_CLASS3; | 1407 | rdata->supported_classes = FC_COS_CLASS3; |
1173 | if (fcp_parm & FCP_SPPF_INIT_FCN) | 1408 | if (fcp_parm & FCP_SPPF_INIT_FCN) |
1174 | roles |= FC_RPORT_ROLE_FCP_INITIATOR; | 1409 | roles |= FC_RPORT_ROLE_FCP_INITIATOR; |
1175 | if (fcp_parm & FCP_SPPF_TARG_FCN) | 1410 | if (fcp_parm & FCP_SPPF_TARG_FCN) |
1176 | roles |= FC_RPORT_ROLE_FCP_TARGET; | 1411 | roles |= FC_RPORT_ROLE_FCP_TARGET; |
1177 | rport->roles = roles; | 1412 | rdata->ids.roles = roles; |
1178 | 1413 | ||
1179 | spp->spp_params = | 1414 | spp->spp_params = |
1180 | htonl(lport->service_params); | 1415 | htonl(lport->service_params); |
@@ -1204,9 +1439,10 @@ static void fc_rport_recv_prli_req(struct fc_rport *rport, | |||
1204 | */ | 1439 | */ |
1205 | switch (rdata->rp_state) { | 1440 | switch (rdata->rp_state) { |
1206 | case RPORT_ST_PRLI: | 1441 | case RPORT_ST_PRLI: |
1207 | fc_rport_enter_ready(rport); | 1442 | fc_rport_enter_ready(rdata); |
1208 | break; | 1443 | break; |
1209 | case RPORT_ST_READY: | 1444 | case RPORT_ST_READY: |
1445 | case RPORT_ST_ADISC: | ||
1210 | break; | 1446 | break; |
1211 | default: | 1447 | default: |
1212 | break; | 1448 | break; |
@@ -1217,17 +1453,17 @@ static void fc_rport_recv_prli_req(struct fc_rport *rport, | |||
1217 | 1453 | ||
1218 | /** | 1454 | /** |
1219 | * fc_rport_recv_prlo_req() - Handle incoming Process Logout (PRLO) request | 1455 | * fc_rport_recv_prlo_req() - Handle incoming Process Logout (PRLO) request |
1220 | * @rport: Fibre Channel remote port that initiated PRLO | 1456 | * @rdata: private remote port data |
1221 | * @sp: current sequence in the PRLO exchange | 1457 | * @sp: current sequence in the PRLO exchange |
1222 | * @fp: PRLO request frame | 1458 | * @fp: PRLO request frame |
1223 | * | 1459 | * |
1224 | * Locking Note: The rport lock is exected to be held before calling | 1460 | * Locking Note: The rport lock is exected to be held before calling |
1225 | * this function. | 1461 | * this function. |
1226 | */ | 1462 | */ |
1227 | static void fc_rport_recv_prlo_req(struct fc_rport *rport, struct fc_seq *sp, | 1463 | static void fc_rport_recv_prlo_req(struct fc_rport_priv *rdata, |
1464 | struct fc_seq *sp, | ||
1228 | struct fc_frame *fp) | 1465 | struct fc_frame *fp) |
1229 | { | 1466 | { |
1230 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | ||
1231 | struct fc_lport *lport = rdata->local_port; | 1467 | struct fc_lport *lport = rdata->local_port; |
1232 | 1468 | ||
1233 | struct fc_frame_header *fh; | 1469 | struct fc_frame_header *fh; |
@@ -1235,13 +1471,8 @@ static void fc_rport_recv_prlo_req(struct fc_rport *rport, struct fc_seq *sp, | |||
1235 | 1471 | ||
1236 | fh = fc_frame_header_get(fp); | 1472 | fh = fc_frame_header_get(fp); |
1237 | 1473 | ||
1238 | FC_RPORT_DBG(rport, "Received PRLO request while in state %s\n", | 1474 | FC_RPORT_DBG(rdata, "Received PRLO request while in state %s\n", |
1239 | fc_rport_state(rport)); | 1475 | fc_rport_state(rdata)); |
1240 | |||
1241 | if (rdata->rp_state == RPORT_ST_NONE) { | ||
1242 | fc_frame_free(fp); | ||
1243 | return; | ||
1244 | } | ||
1245 | 1476 | ||
1246 | rjt_data.fp = NULL; | 1477 | rjt_data.fp = NULL; |
1247 | rjt_data.reason = ELS_RJT_UNAB; | 1478 | rjt_data.reason = ELS_RJT_UNAB; |
@@ -1252,35 +1483,46 @@ static void fc_rport_recv_prlo_req(struct fc_rport *rport, struct fc_seq *sp, | |||
1252 | 1483 | ||
1253 | /** | 1484 | /** |
1254 | * fc_rport_recv_logo_req() - Handle incoming Logout (LOGO) request | 1485 | * fc_rport_recv_logo_req() - Handle incoming Logout (LOGO) request |
1255 | * @rport: Fibre Channel remote port that initiated LOGO | 1486 | * @lport: local port. |
1256 | * @sp: current sequence in the LOGO exchange | 1487 | * @sp: current sequence in the LOGO exchange |
1257 | * @fp: LOGO request frame | 1488 | * @fp: LOGO request frame |
1258 | * | 1489 | * |
1259 | * Locking Note: The rport lock is exected to be held before calling | 1490 | * Locking Note: The rport lock is exected to be held before calling |
1260 | * this function. | 1491 | * this function. |
1261 | */ | 1492 | */ |
1262 | static void fc_rport_recv_logo_req(struct fc_rport *rport, struct fc_seq *sp, | 1493 | static void fc_rport_recv_logo_req(struct fc_lport *lport, |
1494 | struct fc_seq *sp, | ||
1263 | struct fc_frame *fp) | 1495 | struct fc_frame *fp) |
1264 | { | 1496 | { |
1265 | struct fc_frame_header *fh; | 1497 | struct fc_frame_header *fh; |
1266 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | 1498 | struct fc_rport_priv *rdata; |
1267 | struct fc_lport *lport = rdata->local_port; | 1499 | u32 sid; |
1268 | |||
1269 | fh = fc_frame_header_get(fp); | ||
1270 | 1500 | ||
1271 | FC_RPORT_DBG(rport, "Received LOGO request while in state %s\n", | 1501 | lport->tt.seq_els_rsp_send(sp, ELS_LS_ACC, NULL); |
1272 | fc_rport_state(rport)); | ||
1273 | 1502 | ||
1274 | if (rdata->rp_state == RPORT_ST_NONE) { | 1503 | fh = fc_frame_header_get(fp); |
1275 | fc_frame_free(fp); | 1504 | sid = ntoh24(fh->fh_s_id); |
1276 | return; | ||
1277 | } | ||
1278 | 1505 | ||
1279 | rdata->event = RPORT_EV_LOGO; | 1506 | mutex_lock(&lport->disc.disc_mutex); |
1280 | fc_rport_state_enter(rport, RPORT_ST_NONE); | 1507 | rdata = lport->tt.rport_lookup(lport, sid); |
1281 | queue_work(rport_event_queue, &rdata->event_work); | 1508 | if (rdata) { |
1509 | mutex_lock(&rdata->rp_mutex); | ||
1510 | FC_RPORT_DBG(rdata, "Received LOGO request while in state %s\n", | ||
1511 | fc_rport_state(rdata)); | ||
1282 | 1512 | ||
1283 | lport->tt.seq_els_rsp_send(sp, ELS_LS_ACC, NULL); | 1513 | /* |
1514 | * If the remote port was created due to discovery, | ||
1515 | * log back in. It may have seen a stale RSCN about us. | ||
1516 | */ | ||
1517 | if (rdata->rp_state != RPORT_ST_DELETE && rdata->disc_id) | ||
1518 | fc_rport_enter_plogi(rdata); | ||
1519 | else | ||
1520 | fc_rport_enter_delete(rdata, RPORT_EV_LOGO); | ||
1521 | mutex_unlock(&rdata->rp_mutex); | ||
1522 | } else | ||
1523 | FC_RPORT_ID_DBG(lport, sid, | ||
1524 | "Received LOGO from non-logged-in port\n"); | ||
1525 | mutex_unlock(&lport->disc.disc_mutex); | ||
1284 | fc_frame_free(fp); | 1526 | fc_frame_free(fp); |
1285 | } | 1527 | } |
1286 | 1528 | ||
@@ -1291,8 +1533,11 @@ static void fc_rport_flush_queue(void) | |||
1291 | 1533 | ||
1292 | int fc_rport_init(struct fc_lport *lport) | 1534 | int fc_rport_init(struct fc_lport *lport) |
1293 | { | 1535 | { |
1536 | if (!lport->tt.rport_lookup) | ||
1537 | lport->tt.rport_lookup = fc_rport_lookup; | ||
1538 | |||
1294 | if (!lport->tt.rport_create) | 1539 | if (!lport->tt.rport_create) |
1295 | lport->tt.rport_create = fc_rport_rogue_create; | 1540 | lport->tt.rport_create = fc_rport_create; |
1296 | 1541 | ||
1297 | if (!lport->tt.rport_login) | 1542 | if (!lport->tt.rport_login) |
1298 | lport->tt.rport_login = fc_rport_login; | 1543 | lport->tt.rport_login = fc_rport_login; |
@@ -1306,6 +1551,9 @@ int fc_rport_init(struct fc_lport *lport) | |||
1306 | if (!lport->tt.rport_flush_queue) | 1551 | if (!lport->tt.rport_flush_queue) |
1307 | lport->tt.rport_flush_queue = fc_rport_flush_queue; | 1552 | lport->tt.rport_flush_queue = fc_rport_flush_queue; |
1308 | 1553 | ||
1554 | if (!lport->tt.rport_destroy) | ||
1555 | lport->tt.rport_destroy = fc_rport_destroy; | ||
1556 | |||
1309 | return 0; | 1557 | return 0; |
1310 | } | 1558 | } |
1311 | EXPORT_SYMBOL(fc_rport_init); | 1559 | EXPORT_SYMBOL(fc_rport_init); |
@@ -1327,8 +1575,8 @@ EXPORT_SYMBOL(fc_destroy_rport); | |||
1327 | 1575 | ||
1328 | void fc_rport_terminate_io(struct fc_rport *rport) | 1576 | void fc_rport_terminate_io(struct fc_rport *rport) |
1329 | { | 1577 | { |
1330 | struct fc_rport_libfc_priv *rdata = rport->dd_data; | 1578 | struct fc_rport_libfc_priv *rp = rport->dd_data; |
1331 | struct fc_lport *lport = rdata->local_port; | 1579 | struct fc_lport *lport = rp->local_port; |
1332 | 1580 | ||
1333 | lport->tt.exch_mgr_reset(lport, 0, rport->port_id); | 1581 | lport->tt.exch_mgr_reset(lport, 0, rport->port_id); |
1334 | lport->tt.exch_mgr_reset(lport, rport->port_id, 0); | 1582 | lport->tt.exch_mgr_reset(lport, rport->port_id, 0); |
diff --git a/drivers/scsi/libiscsi.c b/drivers/scsi/libiscsi.c index a751f6230c22..8dc73c489a17 100644 --- a/drivers/scsi/libiscsi.c +++ b/drivers/scsi/libiscsi.c | |||
@@ -109,12 +109,9 @@ inline void iscsi_conn_queue_work(struct iscsi_conn *conn) | |||
109 | } | 109 | } |
110 | EXPORT_SYMBOL_GPL(iscsi_conn_queue_work); | 110 | EXPORT_SYMBOL_GPL(iscsi_conn_queue_work); |
111 | 111 | ||
112 | void | 112 | static void __iscsi_update_cmdsn(struct iscsi_session *session, |
113 | iscsi_update_cmdsn(struct iscsi_session *session, struct iscsi_nopin *hdr) | 113 | uint32_t exp_cmdsn, uint32_t max_cmdsn) |
114 | { | 114 | { |
115 | uint32_t max_cmdsn = be32_to_cpu(hdr->max_cmdsn); | ||
116 | uint32_t exp_cmdsn = be32_to_cpu(hdr->exp_cmdsn); | ||
117 | |||
118 | /* | 115 | /* |
119 | * standard specifies this check for when to update expected and | 116 | * standard specifies this check for when to update expected and |
120 | * max sequence numbers | 117 | * max sequence numbers |
@@ -138,6 +135,12 @@ iscsi_update_cmdsn(struct iscsi_session *session, struct iscsi_nopin *hdr) | |||
138 | iscsi_conn_queue_work(session->leadconn); | 135 | iscsi_conn_queue_work(session->leadconn); |
139 | } | 136 | } |
140 | } | 137 | } |
138 | |||
139 | void iscsi_update_cmdsn(struct iscsi_session *session, struct iscsi_nopin *hdr) | ||
140 | { | ||
141 | __iscsi_update_cmdsn(session, be32_to_cpu(hdr->exp_cmdsn), | ||
142 | be32_to_cpu(hdr->max_cmdsn)); | ||
143 | } | ||
141 | EXPORT_SYMBOL_GPL(iscsi_update_cmdsn); | 144 | EXPORT_SYMBOL_GPL(iscsi_update_cmdsn); |
142 | 145 | ||
143 | /** | 146 | /** |
@@ -301,8 +304,6 @@ static int iscsi_prep_scsi_cmd_pdu(struct iscsi_task *task) | |||
301 | hdr->flags = ISCSI_ATTR_SIMPLE; | 304 | hdr->flags = ISCSI_ATTR_SIMPLE; |
302 | int_to_scsilun(sc->device->lun, (struct scsi_lun *)hdr->lun); | 305 | int_to_scsilun(sc->device->lun, (struct scsi_lun *)hdr->lun); |
303 | memcpy(task->lun, hdr->lun, sizeof(task->lun)); | 306 | memcpy(task->lun, hdr->lun, sizeof(task->lun)); |
304 | hdr->cmdsn = task->cmdsn = cpu_to_be32(session->cmdsn); | ||
305 | session->cmdsn++; | ||
306 | hdr->exp_statsn = cpu_to_be32(conn->exp_statsn); | 307 | hdr->exp_statsn = cpu_to_be32(conn->exp_statsn); |
307 | cmd_len = sc->cmd_len; | 308 | cmd_len = sc->cmd_len; |
308 | if (cmd_len < ISCSI_CDB_SIZE) | 309 | if (cmd_len < ISCSI_CDB_SIZE) |
@@ -388,6 +389,8 @@ static int iscsi_prep_scsi_cmd_pdu(struct iscsi_task *task) | |||
388 | return -EIO; | 389 | return -EIO; |
389 | 390 | ||
390 | task->state = ISCSI_TASK_RUNNING; | 391 | task->state = ISCSI_TASK_RUNNING; |
392 | hdr->cmdsn = task->cmdsn = cpu_to_be32(session->cmdsn); | ||
393 | session->cmdsn++; | ||
391 | 394 | ||
392 | conn->scsicmd_pdus_cnt++; | 395 | conn->scsicmd_pdus_cnt++; |
393 | ISCSI_DBG_SESSION(session, "iscsi prep [%s cid %d sc %p cdb 0x%x " | 396 | ISCSI_DBG_SESSION(session, "iscsi prep [%s cid %d sc %p cdb 0x%x " |
@@ -499,6 +502,31 @@ static void iscsi_complete_task(struct iscsi_task *task, int state) | |||
499 | __iscsi_put_task(task); | 502 | __iscsi_put_task(task); |
500 | } | 503 | } |
501 | 504 | ||
505 | /** | ||
506 | * iscsi_complete_scsi_task - finish scsi task normally | ||
507 | * @task: iscsi task for scsi cmd | ||
508 | * @exp_cmdsn: expected cmd sn in cpu format | ||
509 | * @max_cmdsn: max cmd sn in cpu format | ||
510 | * | ||
511 | * This is used when drivers do not need or cannot perform | ||
512 | * lower level pdu processing. | ||
513 | * | ||
514 | * Called with session lock | ||
515 | */ | ||
516 | void iscsi_complete_scsi_task(struct iscsi_task *task, | ||
517 | uint32_t exp_cmdsn, uint32_t max_cmdsn) | ||
518 | { | ||
519 | struct iscsi_conn *conn = task->conn; | ||
520 | |||
521 | ISCSI_DBG_SESSION(conn->session, "[itt 0x%x]\n", task->itt); | ||
522 | |||
523 | conn->last_recv = jiffies; | ||
524 | __iscsi_update_cmdsn(conn->session, exp_cmdsn, max_cmdsn); | ||
525 | iscsi_complete_task(task, ISCSI_TASK_COMPLETED); | ||
526 | } | ||
527 | EXPORT_SYMBOL_GPL(iscsi_complete_scsi_task); | ||
528 | |||
529 | |||
502 | /* | 530 | /* |
503 | * session lock must be held and if not called for a task that is | 531 | * session lock must be held and if not called for a task that is |
504 | * still pending or from the xmit thread, then xmit thread must | 532 | * still pending or from the xmit thread, then xmit thread must |
@@ -857,27 +885,102 @@ static void iscsi_send_nopout(struct iscsi_conn *conn, struct iscsi_nopin *rhdr) | |||
857 | } | 885 | } |
858 | } | 886 | } |
859 | 887 | ||
888 | static int iscsi_nop_out_rsp(struct iscsi_task *task, | ||
889 | struct iscsi_nopin *nop, char *data, int datalen) | ||
890 | { | ||
891 | struct iscsi_conn *conn = task->conn; | ||
892 | int rc = 0; | ||
893 | |||
894 | if (conn->ping_task != task) { | ||
895 | /* | ||
896 | * If this is not in response to one of our | ||
897 | * nops then it must be from userspace. | ||
898 | */ | ||
899 | if (iscsi_recv_pdu(conn->cls_conn, (struct iscsi_hdr *)nop, | ||
900 | data, datalen)) | ||
901 | rc = ISCSI_ERR_CONN_FAILED; | ||
902 | } else | ||
903 | mod_timer(&conn->transport_timer, jiffies + conn->recv_timeout); | ||
904 | iscsi_complete_task(task, ISCSI_TASK_COMPLETED); | ||
905 | return rc; | ||
906 | } | ||
907 | |||
860 | static int iscsi_handle_reject(struct iscsi_conn *conn, struct iscsi_hdr *hdr, | 908 | static int iscsi_handle_reject(struct iscsi_conn *conn, struct iscsi_hdr *hdr, |
861 | char *data, int datalen) | 909 | char *data, int datalen) |
862 | { | 910 | { |
863 | struct iscsi_reject *reject = (struct iscsi_reject *)hdr; | 911 | struct iscsi_reject *reject = (struct iscsi_reject *)hdr; |
864 | struct iscsi_hdr rejected_pdu; | 912 | struct iscsi_hdr rejected_pdu; |
913 | int opcode, rc = 0; | ||
865 | 914 | ||
866 | conn->exp_statsn = be32_to_cpu(reject->statsn) + 1; | 915 | conn->exp_statsn = be32_to_cpu(reject->statsn) + 1; |
867 | 916 | ||
868 | if (reject->reason == ISCSI_REASON_DATA_DIGEST_ERROR) { | 917 | if (ntoh24(reject->dlength) > datalen || |
869 | if (ntoh24(reject->dlength) > datalen) | 918 | ntoh24(reject->dlength) < sizeof(struct iscsi_hdr)) { |
870 | return ISCSI_ERR_PROTO; | 919 | iscsi_conn_printk(KERN_ERR, conn, "Cannot handle rejected " |
920 | "pdu. Invalid data length (pdu dlength " | ||
921 | "%u, datalen %d\n", ntoh24(reject->dlength), | ||
922 | datalen); | ||
923 | return ISCSI_ERR_PROTO; | ||
924 | } | ||
925 | memcpy(&rejected_pdu, data, sizeof(struct iscsi_hdr)); | ||
926 | opcode = rejected_pdu.opcode & ISCSI_OPCODE_MASK; | ||
927 | |||
928 | switch (reject->reason) { | ||
929 | case ISCSI_REASON_DATA_DIGEST_ERROR: | ||
930 | iscsi_conn_printk(KERN_ERR, conn, | ||
931 | "pdu (op 0x%x itt 0x%x) rejected " | ||
932 | "due to DataDigest error.\n", | ||
933 | rejected_pdu.itt, opcode); | ||
934 | break; | ||
935 | case ISCSI_REASON_IMM_CMD_REJECT: | ||
936 | iscsi_conn_printk(KERN_ERR, conn, | ||
937 | "pdu (op 0x%x itt 0x%x) rejected. Too many " | ||
938 | "immediate commands.\n", | ||
939 | rejected_pdu.itt, opcode); | ||
940 | /* | ||
941 | * We only send one TMF at a time so if the target could not | ||
942 | * handle it, then it should get fixed (RFC mandates that | ||
943 | * a target can handle one immediate TMF per conn). | ||
944 | * | ||
945 | * For nops-outs, we could have sent more than one if | ||
946 | * the target is sending us lots of nop-ins | ||
947 | */ | ||
948 | if (opcode != ISCSI_OP_NOOP_OUT) | ||
949 | return 0; | ||
871 | 950 | ||
872 | if (ntoh24(reject->dlength) >= sizeof(struct iscsi_hdr)) { | 951 | if (rejected_pdu.itt == cpu_to_be32(ISCSI_RESERVED_TAG)) |
873 | memcpy(&rejected_pdu, data, sizeof(struct iscsi_hdr)); | 952 | /* |
874 | iscsi_conn_printk(KERN_ERR, conn, | 953 | * nop-out in response to target's nop-out rejected. |
875 | "pdu (op 0x%x) rejected " | 954 | * Just resend. |
876 | "due to DataDigest error.\n", | 955 | */ |
877 | rejected_pdu.opcode); | 956 | iscsi_send_nopout(conn, |
957 | (struct iscsi_nopin*)&rejected_pdu); | ||
958 | else { | ||
959 | struct iscsi_task *task; | ||
960 | /* | ||
961 | * Our nop as ping got dropped. We know the target | ||
962 | * and transport are ok so just clean up | ||
963 | */ | ||
964 | task = iscsi_itt_to_task(conn, rejected_pdu.itt); | ||
965 | if (!task) { | ||
966 | iscsi_conn_printk(KERN_ERR, conn, | ||
967 | "Invalid pdu reject. Could " | ||
968 | "not lookup rejected task.\n"); | ||
969 | rc = ISCSI_ERR_BAD_ITT; | ||
970 | } else | ||
971 | rc = iscsi_nop_out_rsp(task, | ||
972 | (struct iscsi_nopin*)&rejected_pdu, | ||
973 | NULL, 0); | ||
878 | } | 974 | } |
975 | break; | ||
976 | default: | ||
977 | iscsi_conn_printk(KERN_ERR, conn, | ||
978 | "pdu (op 0x%x itt 0x%x) rejected. Reason " | ||
979 | "code 0x%x\n", rejected_pdu.itt, | ||
980 | rejected_pdu.opcode, reject->reason); | ||
981 | break; | ||
879 | } | 982 | } |
880 | return 0; | 983 | return rc; |
881 | } | 984 | } |
882 | 985 | ||
883 | /** | 986 | /** |
@@ -1038,15 +1141,8 @@ int __iscsi_complete_pdu(struct iscsi_conn *conn, struct iscsi_hdr *hdr, | |||
1038 | } | 1141 | } |
1039 | conn->exp_statsn = be32_to_cpu(hdr->statsn) + 1; | 1142 | conn->exp_statsn = be32_to_cpu(hdr->statsn) + 1; |
1040 | 1143 | ||
1041 | if (conn->ping_task != task) | 1144 | rc = iscsi_nop_out_rsp(task, (struct iscsi_nopin*)hdr, |
1042 | /* | 1145 | data, datalen); |
1043 | * If this is not in response to one of our | ||
1044 | * nops then it must be from userspace. | ||
1045 | */ | ||
1046 | goto recv_pdu; | ||
1047 | |||
1048 | mod_timer(&conn->transport_timer, jiffies + conn->recv_timeout); | ||
1049 | iscsi_complete_task(task, ISCSI_TASK_COMPLETED); | ||
1050 | break; | 1146 | break; |
1051 | default: | 1147 | default: |
1052 | rc = ISCSI_ERR_BAD_OPCODE; | 1148 | rc = ISCSI_ERR_BAD_OPCODE; |
@@ -1212,6 +1308,9 @@ static int iscsi_xmit_task(struct iscsi_conn *conn) | |||
1212 | struct iscsi_task *task = conn->task; | 1308 | struct iscsi_task *task = conn->task; |
1213 | int rc; | 1309 | int rc; |
1214 | 1310 | ||
1311 | if (test_bit(ISCSI_SUSPEND_BIT, &conn->suspend_tx)) | ||
1312 | return -ENODATA; | ||
1313 | |||
1215 | __iscsi_get_task(task); | 1314 | __iscsi_get_task(task); |
1216 | spin_unlock_bh(&conn->session->lock); | 1315 | spin_unlock_bh(&conn->session->lock); |
1217 | rc = conn->session->tt->xmit_task(task); | 1316 | rc = conn->session->tt->xmit_task(task); |
@@ -1261,7 +1360,7 @@ static int iscsi_data_xmit(struct iscsi_conn *conn) | |||
1261 | int rc = 0; | 1360 | int rc = 0; |
1262 | 1361 | ||
1263 | spin_lock_bh(&conn->session->lock); | 1362 | spin_lock_bh(&conn->session->lock); |
1264 | if (unlikely(conn->suspend_tx)) { | 1363 | if (test_bit(ISCSI_SUSPEND_BIT, &conn->suspend_tx)) { |
1265 | ISCSI_DBG_SESSION(conn->session, "Tx suspended!\n"); | 1364 | ISCSI_DBG_SESSION(conn->session, "Tx suspended!\n"); |
1266 | spin_unlock_bh(&conn->session->lock); | 1365 | spin_unlock_bh(&conn->session->lock); |
1267 | return -ENODATA; | 1366 | return -ENODATA; |
@@ -1270,7 +1369,7 @@ static int iscsi_data_xmit(struct iscsi_conn *conn) | |||
1270 | if (conn->task) { | 1369 | if (conn->task) { |
1271 | rc = iscsi_xmit_task(conn); | 1370 | rc = iscsi_xmit_task(conn); |
1272 | if (rc) | 1371 | if (rc) |
1273 | goto again; | 1372 | goto done; |
1274 | } | 1373 | } |
1275 | 1374 | ||
1276 | /* | 1375 | /* |
@@ -1290,7 +1389,7 @@ check_mgmt: | |||
1290 | } | 1389 | } |
1291 | rc = iscsi_xmit_task(conn); | 1390 | rc = iscsi_xmit_task(conn); |
1292 | if (rc) | 1391 | if (rc) |
1293 | goto again; | 1392 | goto done; |
1294 | } | 1393 | } |
1295 | 1394 | ||
1296 | /* process pending command queue */ | 1395 | /* process pending command queue */ |
@@ -1311,14 +1410,14 @@ check_mgmt: | |||
1311 | list_add_tail(&conn->task->running, | 1410 | list_add_tail(&conn->task->running, |
1312 | &conn->cmdqueue); | 1411 | &conn->cmdqueue); |
1313 | conn->task = NULL; | 1412 | conn->task = NULL; |
1314 | goto again; | 1413 | goto done; |
1315 | } else | 1414 | } else |
1316 | fail_scsi_task(conn->task, DID_ABORT); | 1415 | fail_scsi_task(conn->task, DID_ABORT); |
1317 | continue; | 1416 | continue; |
1318 | } | 1417 | } |
1319 | rc = iscsi_xmit_task(conn); | 1418 | rc = iscsi_xmit_task(conn); |
1320 | if (rc) | 1419 | if (rc) |
1321 | goto again; | 1420 | goto done; |
1322 | /* | 1421 | /* |
1323 | * we could continuously get new task requests so | 1422 | * we could continuously get new task requests so |
1324 | * we need to check the mgmt queue for nops that need to | 1423 | * we need to check the mgmt queue for nops that need to |
@@ -1344,16 +1443,14 @@ check_mgmt: | |||
1344 | conn->task->state = ISCSI_TASK_RUNNING; | 1443 | conn->task->state = ISCSI_TASK_RUNNING; |
1345 | rc = iscsi_xmit_task(conn); | 1444 | rc = iscsi_xmit_task(conn); |
1346 | if (rc) | 1445 | if (rc) |
1347 | goto again; | 1446 | goto done; |
1348 | if (!list_empty(&conn->mgmtqueue)) | 1447 | if (!list_empty(&conn->mgmtqueue)) |
1349 | goto check_mgmt; | 1448 | goto check_mgmt; |
1350 | } | 1449 | } |
1351 | spin_unlock_bh(&conn->session->lock); | 1450 | spin_unlock_bh(&conn->session->lock); |
1352 | return -ENODATA; | 1451 | return -ENODATA; |
1353 | 1452 | ||
1354 | again: | 1453 | done: |
1355 | if (unlikely(conn->suspend_tx)) | ||
1356 | rc = -ENODATA; | ||
1357 | spin_unlock_bh(&conn->session->lock); | 1454 | spin_unlock_bh(&conn->session->lock); |
1358 | return rc; | 1455 | return rc; |
1359 | } | 1456 | } |
@@ -1474,6 +1571,12 @@ int iscsi_queuecommand(struct scsi_cmnd *sc, void (*done)(struct scsi_cmnd *)) | |||
1474 | goto fault; | 1571 | goto fault; |
1475 | } | 1572 | } |
1476 | 1573 | ||
1574 | if (test_bit(ISCSI_SUSPEND_BIT, &conn->suspend_tx)) { | ||
1575 | reason = FAILURE_SESSION_IN_RECOVERY; | ||
1576 | sc->result = DID_REQUEUE; | ||
1577 | goto fault; | ||
1578 | } | ||
1579 | |||
1477 | if (iscsi_check_cmdsn_window_closed(conn)) { | 1580 | if (iscsi_check_cmdsn_window_closed(conn)) { |
1478 | reason = FAILURE_WINDOW_CLOSED; | 1581 | reason = FAILURE_WINDOW_CLOSED; |
1479 | goto reject; | 1582 | goto reject; |
@@ -1497,6 +1600,7 @@ int iscsi_queuecommand(struct scsi_cmnd *sc, void (*done)(struct scsi_cmnd *)) | |||
1497 | } | 1600 | } |
1498 | } | 1601 | } |
1499 | if (session->tt->xmit_task(task)) { | 1602 | if (session->tt->xmit_task(task)) { |
1603 | session->cmdsn--; | ||
1500 | reason = FAILURE_SESSION_NOT_READY; | 1604 | reason = FAILURE_SESSION_NOT_READY; |
1501 | goto prepd_reject; | 1605 | goto prepd_reject; |
1502 | } | 1606 | } |
@@ -1712,6 +1816,33 @@ static void fail_scsi_tasks(struct iscsi_conn *conn, unsigned lun, | |||
1712 | } | 1816 | } |
1713 | } | 1817 | } |
1714 | 1818 | ||
1819 | /** | ||
1820 | * iscsi_suspend_queue - suspend iscsi_queuecommand | ||
1821 | * @conn: iscsi conn to stop queueing IO on | ||
1822 | * | ||
1823 | * This grabs the session lock to make sure no one is in | ||
1824 | * xmit_task/queuecommand, and then sets suspend to prevent | ||
1825 | * new commands from being queued. This only needs to be called | ||
1826 | * by offload drivers that need to sync a path like ep disconnect | ||
1827 | * with the iscsi_queuecommand/xmit_task. To start IO again libiscsi | ||
1828 | * will call iscsi_start_tx and iscsi_unblock_session when in FFP. | ||
1829 | */ | ||
1830 | void iscsi_suspend_queue(struct iscsi_conn *conn) | ||
1831 | { | ||
1832 | spin_lock_bh(&conn->session->lock); | ||
1833 | set_bit(ISCSI_SUSPEND_BIT, &conn->suspend_tx); | ||
1834 | spin_unlock_bh(&conn->session->lock); | ||
1835 | } | ||
1836 | EXPORT_SYMBOL_GPL(iscsi_suspend_queue); | ||
1837 | |||
1838 | /** | ||
1839 | * iscsi_suspend_tx - suspend iscsi_data_xmit | ||
1840 | * @conn: iscsi conn tp stop processing IO on. | ||
1841 | * | ||
1842 | * This function sets the suspend bit to prevent iscsi_data_xmit | ||
1843 | * from sending new IO, and if work is queued on the xmit thread | ||
1844 | * it will wait for it to be completed. | ||
1845 | */ | ||
1715 | void iscsi_suspend_tx(struct iscsi_conn *conn) | 1846 | void iscsi_suspend_tx(struct iscsi_conn *conn) |
1716 | { | 1847 | { |
1717 | struct Scsi_Host *shost = conn->session->host; | 1848 | struct Scsi_Host *shost = conn->session->host; |
diff --git a/drivers/scsi/libsrp.c b/drivers/scsi/libsrp.c index 2742ae8a3d09..9ad38e81e343 100644 --- a/drivers/scsi/libsrp.c +++ b/drivers/scsi/libsrp.c | |||
@@ -124,6 +124,7 @@ static void srp_ring_free(struct device *dev, struct srp_buf **ring, size_t max, | |||
124 | dma_free_coherent(dev, size, ring[i]->buf, ring[i]->dma); | 124 | dma_free_coherent(dev, size, ring[i]->buf, ring[i]->dma); |
125 | kfree(ring[i]); | 125 | kfree(ring[i]); |
126 | } | 126 | } |
127 | kfree(ring); | ||
127 | } | 128 | } |
128 | 129 | ||
129 | int srp_target_alloc(struct srp_target *target, struct device *dev, | 130 | int srp_target_alloc(struct srp_target *target, struct device *dev, |
diff --git a/drivers/scsi/lpfc/Makefile b/drivers/scsi/lpfc/Makefile index 1c286707dd5f..ad05d6edb8f6 100644 --- a/drivers/scsi/lpfc/Makefile +++ b/drivers/scsi/lpfc/Makefile | |||
@@ -28,4 +28,4 @@ obj-$(CONFIG_SCSI_LPFC) := lpfc.o | |||
28 | 28 | ||
29 | lpfc-objs := lpfc_mem.o lpfc_sli.o lpfc_ct.o lpfc_els.o lpfc_hbadisc.o \ | 29 | lpfc-objs := lpfc_mem.o lpfc_sli.o lpfc_ct.o lpfc_els.o lpfc_hbadisc.o \ |
30 | lpfc_init.o lpfc_mbox.o lpfc_nportdisc.o lpfc_scsi.o lpfc_attr.o \ | 30 | lpfc_init.o lpfc_mbox.o lpfc_nportdisc.o lpfc_scsi.o lpfc_attr.o \ |
31 | lpfc_vport.o lpfc_debugfs.o | 31 | lpfc_vport.o lpfc_debugfs.o lpfc_bsg.o |
diff --git a/drivers/scsi/lpfc/lpfc.h b/drivers/scsi/lpfc/lpfc.h index 1877d9811831..aa10f7951634 100644 --- a/drivers/scsi/lpfc/lpfc.h +++ b/drivers/scsi/lpfc/lpfc.h | |||
@@ -312,6 +312,7 @@ struct lpfc_vport { | |||
312 | #define FC_BYPASSED_MODE 0x20000 /* NPort is in bypassed mode */ | 312 | #define FC_BYPASSED_MODE 0x20000 /* NPort is in bypassed mode */ |
313 | #define FC_VPORT_NEEDS_REG_VPI 0x80000 /* Needs to have its vpi registered */ | 313 | #define FC_VPORT_NEEDS_REG_VPI 0x80000 /* Needs to have its vpi registered */ |
314 | #define FC_RSCN_DEFERRED 0x100000 /* A deferred RSCN being processed */ | 314 | #define FC_RSCN_DEFERRED 0x100000 /* A deferred RSCN being processed */ |
315 | #define FC_VPORT_NEEDS_INIT_VPI 0x200000 /* Need to INIT_VPI before FDISC */ | ||
315 | 316 | ||
316 | uint32_t ct_flags; | 317 | uint32_t ct_flags; |
317 | #define FC_CT_RFF_ID 0x1 /* RFF_ID accepted by switch */ | 318 | #define FC_CT_RFF_ID 0x1 /* RFF_ID accepted by switch */ |
@@ -440,6 +441,12 @@ enum intr_type_t { | |||
440 | MSIX, | 441 | MSIX, |
441 | }; | 442 | }; |
442 | 443 | ||
444 | struct unsol_rcv_ct_ctx { | ||
445 | uint32_t ctxt_id; | ||
446 | uint32_t SID; | ||
447 | uint32_t oxid; | ||
448 | }; | ||
449 | |||
443 | struct lpfc_hba { | 450 | struct lpfc_hba { |
444 | /* SCSI interface function jump table entries */ | 451 | /* SCSI interface function jump table entries */ |
445 | int (*lpfc_new_scsi_buf) | 452 | int (*lpfc_new_scsi_buf) |
@@ -525,6 +532,8 @@ struct lpfc_hba { | |||
525 | #define FCP_XRI_ABORT_EVENT 0x20 | 532 | #define FCP_XRI_ABORT_EVENT 0x20 |
526 | #define ELS_XRI_ABORT_EVENT 0x40 | 533 | #define ELS_XRI_ABORT_EVENT 0x40 |
527 | #define ASYNC_EVENT 0x80 | 534 | #define ASYNC_EVENT 0x80 |
535 | #define LINK_DISABLED 0x100 /* Link disabled by user */ | ||
536 | #define FCF_DISC_INPROGRESS 0x200 /* FCF discovery in progress */ | ||
528 | struct lpfc_dmabuf slim2p; | 537 | struct lpfc_dmabuf slim2p; |
529 | 538 | ||
530 | MAILBOX_t *mbox; | 539 | MAILBOX_t *mbox; |
@@ -616,6 +625,8 @@ struct lpfc_hba { | |||
616 | uint32_t hbq_count; /* Count of configured HBQs */ | 625 | uint32_t hbq_count; /* Count of configured HBQs */ |
617 | struct hbq_s hbqs[LPFC_MAX_HBQS]; /* local copy of hbq indicies */ | 626 | struct hbq_s hbqs[LPFC_MAX_HBQS]; /* local copy of hbq indicies */ |
618 | 627 | ||
628 | uint32_t fcp_qidx; /* next work queue to post work to */ | ||
629 | |||
619 | unsigned long pci_bar0_map; /* Physical address for PCI BAR0 */ | 630 | unsigned long pci_bar0_map; /* Physical address for PCI BAR0 */ |
620 | unsigned long pci_bar1_map; /* Physical address for PCI BAR1 */ | 631 | unsigned long pci_bar1_map; /* Physical address for PCI BAR1 */ |
621 | unsigned long pci_bar2_map; /* Physical address for PCI BAR2 */ | 632 | unsigned long pci_bar2_map; /* Physical address for PCI BAR2 */ |
@@ -682,6 +693,7 @@ struct lpfc_hba { | |||
682 | struct pci_pool *lpfc_mbuf_pool; | 693 | struct pci_pool *lpfc_mbuf_pool; |
683 | struct pci_pool *lpfc_hrb_pool; /* header receive buffer pool */ | 694 | struct pci_pool *lpfc_hrb_pool; /* header receive buffer pool */ |
684 | struct pci_pool *lpfc_drb_pool; /* data receive buffer pool */ | 695 | struct pci_pool *lpfc_drb_pool; /* data receive buffer pool */ |
696 | struct pci_pool *lpfc_hbq_pool; /* SLI3 hbq buffer pool */ | ||
685 | struct lpfc_dma_pool lpfc_mbuf_safety_pool; | 697 | struct lpfc_dma_pool lpfc_mbuf_safety_pool; |
686 | 698 | ||
687 | mempool_t *mbox_mem_pool; | 699 | mempool_t *mbox_mem_pool; |
@@ -763,11 +775,18 @@ struct lpfc_hba { | |||
763 | /* Maximum number of events that can be outstanding at any time*/ | 775 | /* Maximum number of events that can be outstanding at any time*/ |
764 | #define LPFC_MAX_EVT_COUNT 512 | 776 | #define LPFC_MAX_EVT_COUNT 512 |
765 | atomic_t fast_event_count; | 777 | atomic_t fast_event_count; |
778 | uint32_t fcoe_eventtag; | ||
779 | uint32_t fcoe_eventtag_at_fcf_scan; | ||
766 | struct lpfc_fcf fcf; | 780 | struct lpfc_fcf fcf; |
767 | uint8_t fc_map[3]; | 781 | uint8_t fc_map[3]; |
768 | uint8_t valid_vlan; | 782 | uint8_t valid_vlan; |
769 | uint16_t vlan_id; | 783 | uint16_t vlan_id; |
770 | struct list_head fcf_conn_rec_list; | 784 | struct list_head fcf_conn_rec_list; |
785 | |||
786 | struct mutex ct_event_mutex; /* synchronize access to ct_ev_waiters */ | ||
787 | struct list_head ct_ev_waiters; | ||
788 | struct unsol_rcv_ct_ctx ct_ctx[64]; | ||
789 | uint32_t ctx_idx; | ||
771 | }; | 790 | }; |
772 | 791 | ||
773 | static inline struct Scsi_Host * | 792 | static inline struct Scsi_Host * |
diff --git a/drivers/scsi/lpfc/lpfc_attr.c b/drivers/scsi/lpfc/lpfc_attr.c index fc07be5fbce9..e1a30a16a9fa 100644 --- a/drivers/scsi/lpfc/lpfc_attr.c +++ b/drivers/scsi/lpfc/lpfc_attr.c | |||
@@ -394,7 +394,12 @@ lpfc_link_state_show(struct device *dev, struct device_attribute *attr, | |||
394 | case LPFC_INIT_MBX_CMDS: | 394 | case LPFC_INIT_MBX_CMDS: |
395 | case LPFC_LINK_DOWN: | 395 | case LPFC_LINK_DOWN: |
396 | case LPFC_HBA_ERROR: | 396 | case LPFC_HBA_ERROR: |
397 | len += snprintf(buf + len, PAGE_SIZE-len, "Link Down\n"); | 397 | if (phba->hba_flag & LINK_DISABLED) |
398 | len += snprintf(buf + len, PAGE_SIZE-len, | ||
399 | "Link Down - User disabled\n"); | ||
400 | else | ||
401 | len += snprintf(buf + len, PAGE_SIZE-len, | ||
402 | "Link Down\n"); | ||
398 | break; | 403 | break; |
399 | case LPFC_LINK_UP: | 404 | case LPFC_LINK_UP: |
400 | case LPFC_CLEAR_LA: | 405 | case LPFC_CLEAR_LA: |
@@ -4127,6 +4132,9 @@ struct fc_function_template lpfc_transport_functions = { | |||
4127 | .vport_disable = lpfc_vport_disable, | 4132 | .vport_disable = lpfc_vport_disable, |
4128 | 4133 | ||
4129 | .set_vport_symbolic_name = lpfc_set_vport_symbolic_name, | 4134 | .set_vport_symbolic_name = lpfc_set_vport_symbolic_name, |
4135 | |||
4136 | .bsg_request = lpfc_bsg_request, | ||
4137 | .bsg_timeout = lpfc_bsg_timeout, | ||
4130 | }; | 4138 | }; |
4131 | 4139 | ||
4132 | struct fc_function_template lpfc_vport_transport_functions = { | 4140 | struct fc_function_template lpfc_vport_transport_functions = { |
diff --git a/drivers/scsi/lpfc/lpfc_bsg.c b/drivers/scsi/lpfc/lpfc_bsg.c new file mode 100644 index 000000000000..da6bf5aac9dd --- /dev/null +++ b/drivers/scsi/lpfc/lpfc_bsg.c | |||
@@ -0,0 +1,904 @@ | |||
1 | /******************************************************************* | ||
2 | * This file is part of the Emulex Linux Device Driver for * | ||
3 | * Fibre Channel Host Bus Adapters. * | ||
4 | * Copyright (C) 2009 Emulex. All rights reserved. * | ||
5 | * EMULEX and SLI are trademarks of Emulex. * | ||
6 | * www.emulex.com * | ||
7 | * * | ||
8 | * This program is free software; you can redistribute it and/or * | ||
9 | * modify it under the terms of version 2 of the GNU General * | ||
10 | * Public License as published by the Free Software Foundation. * | ||
11 | * This program is distributed in the hope that it will be useful. * | ||
12 | * ALL EXPRESS OR IMPLIED CONDITIONS, REPRESENTATIONS AND * | ||
13 | * WARRANTIES, INCLUDING ANY IMPLIED WARRANTY OF MERCHANTABILITY, * | ||
14 | * FITNESS FOR A PARTICULAR PURPOSE, OR NON-INFRINGEMENT, ARE * | ||
15 | * DISCLAIMED, EXCEPT TO THE EXTENT THAT SUCH DISCLAIMERS ARE HELD * | ||
16 | * TO BE LEGALLY INVALID. See the GNU General Public License for * | ||
17 | * more details, a copy of which can be found in the file COPYING * | ||
18 | * included with this package. * | ||
19 | *******************************************************************/ | ||
20 | |||
21 | #include <linux/interrupt.h> | ||
22 | #include <linux/mempool.h> | ||
23 | #include <linux/pci.h> | ||
24 | |||
25 | #include <scsi/scsi.h> | ||
26 | #include <scsi/scsi_host.h> | ||
27 | #include <scsi/scsi_transport_fc.h> | ||
28 | #include <scsi/scsi_bsg_fc.h> | ||
29 | |||
30 | #include "lpfc_hw4.h" | ||
31 | #include "lpfc_hw.h" | ||
32 | #include "lpfc_sli.h" | ||
33 | #include "lpfc_sli4.h" | ||
34 | #include "lpfc_nl.h" | ||
35 | #include "lpfc_disc.h" | ||
36 | #include "lpfc_scsi.h" | ||
37 | #include "lpfc.h" | ||
38 | #include "lpfc_logmsg.h" | ||
39 | #include "lpfc_crtn.h" | ||
40 | #include "lpfc_vport.h" | ||
41 | #include "lpfc_version.h" | ||
42 | |||
43 | /** | ||
44 | * lpfc_bsg_rport_ct - send a CT command from a bsg request | ||
45 | * @job: fc_bsg_job to handle | ||
46 | */ | ||
47 | static int | ||
48 | lpfc_bsg_rport_ct(struct fc_bsg_job *job) | ||
49 | { | ||
50 | struct Scsi_Host *shost = job->shost; | ||
51 | struct lpfc_vport *vport = (struct lpfc_vport *)job->shost->hostdata; | ||
52 | struct lpfc_hba *phba = vport->phba; | ||
53 | struct lpfc_rport_data *rdata = job->rport->dd_data; | ||
54 | struct lpfc_nodelist *ndlp = rdata->pnode; | ||
55 | struct ulp_bde64 *bpl = NULL; | ||
56 | uint32_t timeout; | ||
57 | struct lpfc_iocbq *cmdiocbq = NULL; | ||
58 | struct lpfc_iocbq *rspiocbq = NULL; | ||
59 | IOCB_t *cmd; | ||
60 | IOCB_t *rsp; | ||
61 | struct lpfc_dmabuf *bmp = NULL; | ||
62 | int request_nseg; | ||
63 | int reply_nseg; | ||
64 | struct scatterlist *sgel = NULL; | ||
65 | int numbde; | ||
66 | dma_addr_t busaddr; | ||
67 | int rc = 0; | ||
68 | |||
69 | /* in case no data is transferred */ | ||
70 | job->reply->reply_payload_rcv_len = 0; | ||
71 | |||
72 | if (!lpfc_nlp_get(ndlp)) { | ||
73 | job->reply->result = -ENODEV; | ||
74 | return 0; | ||
75 | } | ||
76 | |||
77 | if (ndlp->nlp_flag & NLP_ELS_SND_MASK) { | ||
78 | rc = -ENODEV; | ||
79 | goto free_ndlp_exit; | ||
80 | } | ||
81 | |||
82 | spin_lock_irq(shost->host_lock); | ||
83 | cmdiocbq = lpfc_sli_get_iocbq(phba); | ||
84 | if (!cmdiocbq) { | ||
85 | rc = -ENOMEM; | ||
86 | spin_unlock_irq(shost->host_lock); | ||
87 | goto free_ndlp_exit; | ||
88 | } | ||
89 | cmd = &cmdiocbq->iocb; | ||
90 | |||
91 | rspiocbq = lpfc_sli_get_iocbq(phba); | ||
92 | if (!rspiocbq) { | ||
93 | rc = -ENOMEM; | ||
94 | goto free_cmdiocbq; | ||
95 | } | ||
96 | spin_unlock_irq(shost->host_lock); | ||
97 | |||
98 | rsp = &rspiocbq->iocb; | ||
99 | |||
100 | bmp = kmalloc(sizeof(struct lpfc_dmabuf), GFP_KERNEL); | ||
101 | if (!bmp) { | ||
102 | rc = -ENOMEM; | ||
103 | spin_lock_irq(shost->host_lock); | ||
104 | goto free_rspiocbq; | ||
105 | } | ||
106 | |||
107 | spin_lock_irq(shost->host_lock); | ||
108 | bmp->virt = lpfc_mbuf_alloc(phba, 0, &bmp->phys); | ||
109 | if (!bmp->virt) { | ||
110 | rc = -ENOMEM; | ||
111 | goto free_bmp; | ||
112 | } | ||
113 | spin_unlock_irq(shost->host_lock); | ||
114 | |||
115 | INIT_LIST_HEAD(&bmp->list); | ||
116 | bpl = (struct ulp_bde64 *) bmp->virt; | ||
117 | |||
118 | request_nseg = pci_map_sg(phba->pcidev, job->request_payload.sg_list, | ||
119 | job->request_payload.sg_cnt, DMA_TO_DEVICE); | ||
120 | for_each_sg(job->request_payload.sg_list, sgel, request_nseg, numbde) { | ||
121 | busaddr = sg_dma_address(sgel); | ||
122 | bpl->tus.f.bdeFlags = BUFF_TYPE_BDE_64; | ||
123 | bpl->tus.f.bdeSize = sg_dma_len(sgel); | ||
124 | bpl->tus.w = cpu_to_le32(bpl->tus.w); | ||
125 | bpl->addrLow = cpu_to_le32(putPaddrLow(busaddr)); | ||
126 | bpl->addrHigh = cpu_to_le32(putPaddrHigh(busaddr)); | ||
127 | bpl++; | ||
128 | } | ||
129 | |||
130 | reply_nseg = pci_map_sg(phba->pcidev, job->reply_payload.sg_list, | ||
131 | job->reply_payload.sg_cnt, DMA_FROM_DEVICE); | ||
132 | for_each_sg(job->reply_payload.sg_list, sgel, reply_nseg, numbde) { | ||
133 | busaddr = sg_dma_address(sgel); | ||
134 | bpl->tus.f.bdeFlags = BUFF_TYPE_BDE_64I; | ||
135 | bpl->tus.f.bdeSize = sg_dma_len(sgel); | ||
136 | bpl->tus.w = cpu_to_le32(bpl->tus.w); | ||
137 | bpl->addrLow = cpu_to_le32(putPaddrLow(busaddr)); | ||
138 | bpl->addrHigh = cpu_to_le32(putPaddrHigh(busaddr)); | ||
139 | bpl++; | ||
140 | } | ||
141 | |||
142 | cmd->un.genreq64.bdl.ulpIoTag32 = 0; | ||
143 | cmd->un.genreq64.bdl.addrHigh = putPaddrHigh(bmp->phys); | ||
144 | cmd->un.genreq64.bdl.addrLow = putPaddrLow(bmp->phys); | ||
145 | cmd->un.genreq64.bdl.bdeFlags = BUFF_TYPE_BLP_64; | ||
146 | cmd->un.genreq64.bdl.bdeSize = | ||
147 | (request_nseg + reply_nseg) * sizeof(struct ulp_bde64); | ||
148 | cmd->ulpCommand = CMD_GEN_REQUEST64_CR; | ||
149 | cmd->un.genreq64.w5.hcsw.Fctl = (SI | LA); | ||
150 | cmd->un.genreq64.w5.hcsw.Dfctl = 0; | ||
151 | cmd->un.genreq64.w5.hcsw.Rctl = FC_UNSOL_CTL; | ||
152 | cmd->un.genreq64.w5.hcsw.Type = FC_COMMON_TRANSPORT_ULP; | ||
153 | cmd->ulpBdeCount = 1; | ||
154 | cmd->ulpLe = 1; | ||
155 | cmd->ulpClass = CLASS3; | ||
156 | cmd->ulpContext = ndlp->nlp_rpi; | ||
157 | cmd->ulpOwner = OWN_CHIP; | ||
158 | cmdiocbq->vport = phba->pport; | ||
159 | cmdiocbq->context1 = NULL; | ||
160 | cmdiocbq->context2 = NULL; | ||
161 | cmdiocbq->iocb_flag |= LPFC_IO_LIBDFC; | ||
162 | |||
163 | timeout = phba->fc_ratov * 2; | ||
164 | job->dd_data = cmdiocbq; | ||
165 | |||
166 | rc = lpfc_sli_issue_iocb_wait(phba, LPFC_ELS_RING, cmdiocbq, rspiocbq, | ||
167 | timeout + LPFC_DRVR_TIMEOUT); | ||
168 | |||
169 | if (rc != IOCB_TIMEDOUT) { | ||
170 | pci_unmap_sg(phba->pcidev, job->request_payload.sg_list, | ||
171 | job->request_payload.sg_cnt, DMA_TO_DEVICE); | ||
172 | pci_unmap_sg(phba->pcidev, job->reply_payload.sg_list, | ||
173 | job->reply_payload.sg_cnt, DMA_FROM_DEVICE); | ||
174 | } | ||
175 | |||
176 | if (rc == IOCB_TIMEDOUT) { | ||
177 | lpfc_sli_release_iocbq(phba, rspiocbq); | ||
178 | rc = -EACCES; | ||
179 | goto free_ndlp_exit; | ||
180 | } | ||
181 | |||
182 | if (rc != IOCB_SUCCESS) { | ||
183 | rc = -EACCES; | ||
184 | goto free_outdmp; | ||
185 | } | ||
186 | |||
187 | if (rsp->ulpStatus) { | ||
188 | if (rsp->ulpStatus == IOSTAT_LOCAL_REJECT) { | ||
189 | switch (rsp->un.ulpWord[4] & 0xff) { | ||
190 | case IOERR_SEQUENCE_TIMEOUT: | ||
191 | rc = -ETIMEDOUT; | ||
192 | break; | ||
193 | case IOERR_INVALID_RPI: | ||
194 | rc = -EFAULT; | ||
195 | break; | ||
196 | default: | ||
197 | rc = -EACCES; | ||
198 | break; | ||
199 | } | ||
200 | goto free_outdmp; | ||
201 | } | ||
202 | } else | ||
203 | job->reply->reply_payload_rcv_len = | ||
204 | rsp->un.genreq64.bdl.bdeSize; | ||
205 | |||
206 | free_outdmp: | ||
207 | spin_lock_irq(shost->host_lock); | ||
208 | lpfc_mbuf_free(phba, bmp->virt, bmp->phys); | ||
209 | free_bmp: | ||
210 | kfree(bmp); | ||
211 | free_rspiocbq: | ||
212 | lpfc_sli_release_iocbq(phba, rspiocbq); | ||
213 | free_cmdiocbq: | ||
214 | lpfc_sli_release_iocbq(phba, cmdiocbq); | ||
215 | spin_unlock_irq(shost->host_lock); | ||
216 | free_ndlp_exit: | ||
217 | lpfc_nlp_put(ndlp); | ||
218 | |||
219 | /* make error code available to userspace */ | ||
220 | job->reply->result = rc; | ||
221 | /* complete the job back to userspace */ | ||
222 | job->job_done(job); | ||
223 | |||
224 | return 0; | ||
225 | } | ||
226 | |||
227 | /** | ||
228 | * lpfc_bsg_rport_els - send an ELS command from a bsg request | ||
229 | * @job: fc_bsg_job to handle | ||
230 | */ | ||
231 | static int | ||
232 | lpfc_bsg_rport_els(struct fc_bsg_job *job) | ||
233 | { | ||
234 | struct lpfc_vport *vport = (struct lpfc_vport *)job->shost->hostdata; | ||
235 | struct lpfc_hba *phba = vport->phba; | ||
236 | struct lpfc_rport_data *rdata = job->rport->dd_data; | ||
237 | struct lpfc_nodelist *ndlp = rdata->pnode; | ||
238 | |||
239 | uint32_t elscmd; | ||
240 | uint32_t cmdsize; | ||
241 | uint32_t rspsize; | ||
242 | struct lpfc_iocbq *rspiocbq; | ||
243 | struct lpfc_iocbq *cmdiocbq; | ||
244 | IOCB_t *rsp; | ||
245 | uint16_t rpi = 0; | ||
246 | struct lpfc_dmabuf *pcmd; | ||
247 | struct lpfc_dmabuf *prsp; | ||
248 | struct lpfc_dmabuf *pbuflist = NULL; | ||
249 | struct ulp_bde64 *bpl; | ||
250 | int iocb_status; | ||
251 | int request_nseg; | ||
252 | int reply_nseg; | ||
253 | struct scatterlist *sgel = NULL; | ||
254 | int numbde; | ||
255 | dma_addr_t busaddr; | ||
256 | int rc = 0; | ||
257 | |||
258 | /* in case no data is transferred */ | ||
259 | job->reply->reply_payload_rcv_len = 0; | ||
260 | |||
261 | if (!lpfc_nlp_get(ndlp)) { | ||
262 | rc = -ENODEV; | ||
263 | goto out; | ||
264 | } | ||
265 | |||
266 | elscmd = job->request->rqst_data.r_els.els_code; | ||
267 | cmdsize = job->request_payload.payload_len; | ||
268 | rspsize = job->reply_payload.payload_len; | ||
269 | rspiocbq = lpfc_sli_get_iocbq(phba); | ||
270 | if (!rspiocbq) { | ||
271 | lpfc_nlp_put(ndlp); | ||
272 | rc = -ENOMEM; | ||
273 | goto out; | ||
274 | } | ||
275 | |||
276 | rsp = &rspiocbq->iocb; | ||
277 | rpi = ndlp->nlp_rpi; | ||
278 | |||
279 | cmdiocbq = lpfc_prep_els_iocb(phba->pport, 1, cmdsize, 0, ndlp, | ||
280 | ndlp->nlp_DID, elscmd); | ||
281 | |||
282 | if (!cmdiocbq) { | ||
283 | lpfc_sli_release_iocbq(phba, rspiocbq); | ||
284 | return -EIO; | ||
285 | } | ||
286 | |||
287 | job->dd_data = cmdiocbq; | ||
288 | pcmd = (struct lpfc_dmabuf *) cmdiocbq->context2; | ||
289 | prsp = (struct lpfc_dmabuf *) pcmd->list.next; | ||
290 | |||
291 | lpfc_mbuf_free(phba, pcmd->virt, pcmd->phys); | ||
292 | kfree(pcmd); | ||
293 | lpfc_mbuf_free(phba, prsp->virt, prsp->phys); | ||
294 | kfree(prsp); | ||
295 | cmdiocbq->context2 = NULL; | ||
296 | |||
297 | pbuflist = (struct lpfc_dmabuf *) cmdiocbq->context3; | ||
298 | bpl = (struct ulp_bde64 *) pbuflist->virt; | ||
299 | |||
300 | request_nseg = pci_map_sg(phba->pcidev, job->request_payload.sg_list, | ||
301 | job->request_payload.sg_cnt, DMA_TO_DEVICE); | ||
302 | |||
303 | for_each_sg(job->request_payload.sg_list, sgel, request_nseg, numbde) { | ||
304 | busaddr = sg_dma_address(sgel); | ||
305 | bpl->tus.f.bdeFlags = BUFF_TYPE_BDE_64; | ||
306 | bpl->tus.f.bdeSize = sg_dma_len(sgel); | ||
307 | bpl->tus.w = cpu_to_le32(bpl->tus.w); | ||
308 | bpl->addrLow = cpu_to_le32(putPaddrLow(busaddr)); | ||
309 | bpl->addrHigh = cpu_to_le32(putPaddrHigh(busaddr)); | ||
310 | bpl++; | ||
311 | } | ||
312 | |||
313 | reply_nseg = pci_map_sg(phba->pcidev, job->reply_payload.sg_list, | ||
314 | job->reply_payload.sg_cnt, DMA_FROM_DEVICE); | ||
315 | for_each_sg(job->reply_payload.sg_list, sgel, reply_nseg, numbde) { | ||
316 | busaddr = sg_dma_address(sgel); | ||
317 | bpl->tus.f.bdeFlags = BUFF_TYPE_BDE_64I; | ||
318 | bpl->tus.f.bdeSize = sg_dma_len(sgel); | ||
319 | bpl->tus.w = cpu_to_le32(bpl->tus.w); | ||
320 | bpl->addrLow = cpu_to_le32(putPaddrLow(busaddr)); | ||
321 | bpl->addrHigh = cpu_to_le32(putPaddrHigh(busaddr)); | ||
322 | bpl++; | ||
323 | } | ||
324 | |||
325 | cmdiocbq->iocb.un.elsreq64.bdl.bdeSize = | ||
326 | (request_nseg + reply_nseg) * sizeof(struct ulp_bde64); | ||
327 | cmdiocbq->iocb.ulpContext = rpi; | ||
328 | cmdiocbq->iocb_flag |= LPFC_IO_LIBDFC; | ||
329 | cmdiocbq->context1 = NULL; | ||
330 | cmdiocbq->context2 = NULL; | ||
331 | |||
332 | iocb_status = lpfc_sli_issue_iocb_wait(phba, LPFC_ELS_RING, cmdiocbq, | ||
333 | rspiocbq, (phba->fc_ratov * 2) | ||
334 | + LPFC_DRVR_TIMEOUT); | ||
335 | |||
336 | /* release the new ndlp once the iocb completes */ | ||
337 | lpfc_nlp_put(ndlp); | ||
338 | if (iocb_status != IOCB_TIMEDOUT) { | ||
339 | pci_unmap_sg(phba->pcidev, job->request_payload.sg_list, | ||
340 | job->request_payload.sg_cnt, DMA_TO_DEVICE); | ||
341 | pci_unmap_sg(phba->pcidev, job->reply_payload.sg_list, | ||
342 | job->reply_payload.sg_cnt, DMA_FROM_DEVICE); | ||
343 | } | ||
344 | |||
345 | if (iocb_status == IOCB_SUCCESS) { | ||
346 | if (rsp->ulpStatus == IOSTAT_SUCCESS) { | ||
347 | job->reply->reply_payload_rcv_len = | ||
348 | rsp->un.elsreq64.bdl.bdeSize; | ||
349 | rc = 0; | ||
350 | } else if (rsp->ulpStatus == IOSTAT_LS_RJT) { | ||
351 | struct fc_bsg_ctels_reply *els_reply; | ||
352 | /* LS_RJT data returned in word 4 */ | ||
353 | uint8_t *rjt_data = (uint8_t *)&rsp->un.ulpWord[4]; | ||
354 | |||
355 | els_reply = &job->reply->reply_data.ctels_reply; | ||
356 | job->reply->result = 0; | ||
357 | els_reply->status = FC_CTELS_STATUS_REJECT; | ||
358 | els_reply->rjt_data.action = rjt_data[0]; | ||
359 | els_reply->rjt_data.reason_code = rjt_data[1]; | ||
360 | els_reply->rjt_data.reason_explanation = rjt_data[2]; | ||
361 | els_reply->rjt_data.vendor_unique = rjt_data[3]; | ||
362 | } else | ||
363 | rc = -EIO; | ||
364 | } else | ||
365 | rc = -EIO; | ||
366 | |||
367 | if (iocb_status != IOCB_TIMEDOUT) | ||
368 | lpfc_els_free_iocb(phba, cmdiocbq); | ||
369 | |||
370 | lpfc_sli_release_iocbq(phba, rspiocbq); | ||
371 | |||
372 | out: | ||
373 | /* make error code available to userspace */ | ||
374 | job->reply->result = rc; | ||
375 | /* complete the job back to userspace */ | ||
376 | job->job_done(job); | ||
377 | |||
378 | return 0; | ||
379 | } | ||
380 | |||
381 | struct lpfc_ct_event { | ||
382 | struct list_head node; | ||
383 | int ref; | ||
384 | wait_queue_head_t wq; | ||
385 | |||
386 | /* Event type and waiter identifiers */ | ||
387 | uint32_t type_mask; | ||
388 | uint32_t req_id; | ||
389 | uint32_t reg_id; | ||
390 | |||
391 | /* next two flags are here for the auto-delete logic */ | ||
392 | unsigned long wait_time_stamp; | ||
393 | int waiting; | ||
394 | |||
395 | /* seen and not seen events */ | ||
396 | struct list_head events_to_get; | ||
397 | struct list_head events_to_see; | ||
398 | }; | ||
399 | |||
400 | struct event_data { | ||
401 | struct list_head node; | ||
402 | uint32_t type; | ||
403 | uint32_t immed_dat; | ||
404 | void *data; | ||
405 | uint32_t len; | ||
406 | }; | ||
407 | |||
408 | static struct lpfc_ct_event * | ||
409 | lpfc_ct_event_new(int ev_reg_id, uint32_t ev_req_id) | ||
410 | { | ||
411 | struct lpfc_ct_event *evt = kzalloc(sizeof(*evt), GFP_KERNEL); | ||
412 | if (!evt) | ||
413 | return NULL; | ||
414 | |||
415 | INIT_LIST_HEAD(&evt->events_to_get); | ||
416 | INIT_LIST_HEAD(&evt->events_to_see); | ||
417 | evt->req_id = ev_req_id; | ||
418 | evt->reg_id = ev_reg_id; | ||
419 | evt->wait_time_stamp = jiffies; | ||
420 | init_waitqueue_head(&evt->wq); | ||
421 | |||
422 | return evt; | ||
423 | } | ||
424 | |||
425 | static void | ||
426 | lpfc_ct_event_free(struct lpfc_ct_event *evt) | ||
427 | { | ||
428 | struct event_data *ed; | ||
429 | |||
430 | list_del(&evt->node); | ||
431 | |||
432 | while (!list_empty(&evt->events_to_get)) { | ||
433 | ed = list_entry(evt->events_to_get.next, typeof(*ed), node); | ||
434 | list_del(&ed->node); | ||
435 | kfree(ed->data); | ||
436 | kfree(ed); | ||
437 | } | ||
438 | |||
439 | while (!list_empty(&evt->events_to_see)) { | ||
440 | ed = list_entry(evt->events_to_see.next, typeof(*ed), node); | ||
441 | list_del(&ed->node); | ||
442 | kfree(ed->data); | ||
443 | kfree(ed); | ||
444 | } | ||
445 | |||
446 | kfree(evt); | ||
447 | } | ||
448 | |||
449 | static inline void | ||
450 | lpfc_ct_event_ref(struct lpfc_ct_event *evt) | ||
451 | { | ||
452 | evt->ref++; | ||
453 | } | ||
454 | |||
455 | static inline void | ||
456 | lpfc_ct_event_unref(struct lpfc_ct_event *evt) | ||
457 | { | ||
458 | if (--evt->ref < 0) | ||
459 | lpfc_ct_event_free(evt); | ||
460 | } | ||
461 | |||
462 | #define SLI_CT_ELX_LOOPBACK 0x10 | ||
463 | |||
464 | enum ELX_LOOPBACK_CMD { | ||
465 | ELX_LOOPBACK_XRI_SETUP, | ||
466 | ELX_LOOPBACK_DATA, | ||
467 | }; | ||
468 | |||
469 | /** | ||
470 | * lpfc_bsg_ct_unsol_event - process an unsolicited CT command | ||
471 | * @phba: | ||
472 | * @pring: | ||
473 | * @piocbq: | ||
474 | * | ||
475 | * This function is called when an unsolicited CT command is received. It | ||
476 | * forwards the event to any processes registerd to receive CT events. | ||
477 | */ | ||
478 | void | ||
479 | lpfc_bsg_ct_unsol_event(struct lpfc_hba *phba, struct lpfc_sli_ring *pring, | ||
480 | struct lpfc_iocbq *piocbq) | ||
481 | { | ||
482 | uint32_t evt_req_id = 0; | ||
483 | uint32_t cmd; | ||
484 | uint32_t len; | ||
485 | struct lpfc_dmabuf *dmabuf = NULL; | ||
486 | struct lpfc_ct_event *evt; | ||
487 | struct event_data *evt_dat = NULL; | ||
488 | struct lpfc_iocbq *iocbq; | ||
489 | size_t offset = 0; | ||
490 | struct list_head head; | ||
491 | struct ulp_bde64 *bde; | ||
492 | dma_addr_t dma_addr; | ||
493 | int i; | ||
494 | struct lpfc_dmabuf *bdeBuf1 = piocbq->context2; | ||
495 | struct lpfc_dmabuf *bdeBuf2 = piocbq->context3; | ||
496 | struct lpfc_hbq_entry *hbqe; | ||
497 | struct lpfc_sli_ct_request *ct_req; | ||
498 | |||
499 | INIT_LIST_HEAD(&head); | ||
500 | list_add_tail(&head, &piocbq->list); | ||
501 | |||
502 | if (piocbq->iocb.ulpBdeCount == 0 || | ||
503 | piocbq->iocb.un.cont64[0].tus.f.bdeSize == 0) | ||
504 | goto error_ct_unsol_exit; | ||
505 | |||
506 | if (phba->sli3_options & LPFC_SLI3_HBQ_ENABLED) | ||
507 | dmabuf = bdeBuf1; | ||
508 | else { | ||
509 | dma_addr = getPaddr(piocbq->iocb.un.cont64[0].addrHigh, | ||
510 | piocbq->iocb.un.cont64[0].addrLow); | ||
511 | dmabuf = lpfc_sli_ringpostbuf_get(phba, pring, dma_addr); | ||
512 | } | ||
513 | |||
514 | ct_req = (struct lpfc_sli_ct_request *)dmabuf->virt; | ||
515 | evt_req_id = ct_req->FsType; | ||
516 | cmd = ct_req->CommandResponse.bits.CmdRsp; | ||
517 | len = ct_req->CommandResponse.bits.Size; | ||
518 | if (!(phba->sli3_options & LPFC_SLI3_HBQ_ENABLED)) | ||
519 | lpfc_sli_ringpostbuf_put(phba, pring, dmabuf); | ||
520 | |||
521 | mutex_lock(&phba->ct_event_mutex); | ||
522 | list_for_each_entry(evt, &phba->ct_ev_waiters, node) { | ||
523 | if (evt->req_id != evt_req_id) | ||
524 | continue; | ||
525 | |||
526 | lpfc_ct_event_ref(evt); | ||
527 | |||
528 | evt_dat = kzalloc(sizeof(*evt_dat), GFP_KERNEL); | ||
529 | if (!evt_dat) { | ||
530 | lpfc_ct_event_unref(evt); | ||
531 | lpfc_printf_log(phba, KERN_WARNING, LOG_LIBDFC, | ||
532 | "2614 Memory allocation failed for " | ||
533 | "CT event\n"); | ||
534 | break; | ||
535 | } | ||
536 | |||
537 | mutex_unlock(&phba->ct_event_mutex); | ||
538 | |||
539 | if (phba->sli3_options & LPFC_SLI3_HBQ_ENABLED) { | ||
540 | /* take accumulated byte count from the last iocbq */ | ||
541 | iocbq = list_entry(head.prev, typeof(*iocbq), list); | ||
542 | evt_dat->len = iocbq->iocb.unsli3.rcvsli3.acc_len; | ||
543 | } else { | ||
544 | list_for_each_entry(iocbq, &head, list) { | ||
545 | for (i = 0; i < iocbq->iocb.ulpBdeCount; i++) | ||
546 | evt_dat->len += | ||
547 | iocbq->iocb.un.cont64[i].tus.f.bdeSize; | ||
548 | } | ||
549 | } | ||
550 | |||
551 | evt_dat->data = kzalloc(evt_dat->len, GFP_KERNEL); | ||
552 | if (!evt_dat->data) { | ||
553 | lpfc_printf_log(phba, KERN_WARNING, LOG_LIBDFC, | ||
554 | "2615 Memory allocation failed for " | ||
555 | "CT event data, size %d\n", | ||
556 | evt_dat->len); | ||
557 | kfree(evt_dat); | ||
558 | mutex_lock(&phba->ct_event_mutex); | ||
559 | lpfc_ct_event_unref(evt); | ||
560 | mutex_unlock(&phba->ct_event_mutex); | ||
561 | goto error_ct_unsol_exit; | ||
562 | } | ||
563 | |||
564 | list_for_each_entry(iocbq, &head, list) { | ||
565 | if (phba->sli3_options & LPFC_SLI3_HBQ_ENABLED) { | ||
566 | bdeBuf1 = iocbq->context2; | ||
567 | bdeBuf2 = iocbq->context3; | ||
568 | } | ||
569 | for (i = 0; i < iocbq->iocb.ulpBdeCount; i++) { | ||
570 | int size = 0; | ||
571 | if (phba->sli3_options & | ||
572 | LPFC_SLI3_HBQ_ENABLED) { | ||
573 | if (i == 0) { | ||
574 | hbqe = (struct lpfc_hbq_entry *) | ||
575 | &iocbq->iocb.un.ulpWord[0]; | ||
576 | size = hbqe->bde.tus.f.bdeSize; | ||
577 | dmabuf = bdeBuf1; | ||
578 | } else if (i == 1) { | ||
579 | hbqe = (struct lpfc_hbq_entry *) | ||
580 | &iocbq->iocb.unsli3. | ||
581 | sli3Words[4]; | ||
582 | size = hbqe->bde.tus.f.bdeSize; | ||
583 | dmabuf = bdeBuf2; | ||
584 | } | ||
585 | if ((offset + size) > evt_dat->len) | ||
586 | size = evt_dat->len - offset; | ||
587 | } else { | ||
588 | size = iocbq->iocb.un.cont64[i]. | ||
589 | tus.f.bdeSize; | ||
590 | bde = &iocbq->iocb.un.cont64[i]; | ||
591 | dma_addr = getPaddr(bde->addrHigh, | ||
592 | bde->addrLow); | ||
593 | dmabuf = lpfc_sli_ringpostbuf_get(phba, | ||
594 | pring, dma_addr); | ||
595 | } | ||
596 | if (!dmabuf) { | ||
597 | lpfc_printf_log(phba, KERN_ERR, | ||
598 | LOG_LIBDFC, "2616 No dmabuf " | ||
599 | "found for iocbq 0x%p\n", | ||
600 | iocbq); | ||
601 | kfree(evt_dat->data); | ||
602 | kfree(evt_dat); | ||
603 | mutex_lock(&phba->ct_event_mutex); | ||
604 | lpfc_ct_event_unref(evt); | ||
605 | mutex_unlock(&phba->ct_event_mutex); | ||
606 | goto error_ct_unsol_exit; | ||
607 | } | ||
608 | memcpy((char *)(evt_dat->data) + offset, | ||
609 | dmabuf->virt, size); | ||
610 | offset += size; | ||
611 | if (evt_req_id != SLI_CT_ELX_LOOPBACK && | ||
612 | !(phba->sli3_options & | ||
613 | LPFC_SLI3_HBQ_ENABLED)) { | ||
614 | lpfc_sli_ringpostbuf_put(phba, pring, | ||
615 | dmabuf); | ||
616 | } else { | ||
617 | switch (cmd) { | ||
618 | case ELX_LOOPBACK_XRI_SETUP: | ||
619 | if (!(phba->sli3_options & | ||
620 | LPFC_SLI3_HBQ_ENABLED)) | ||
621 | lpfc_post_buffer(phba, | ||
622 | pring, | ||
623 | 1); | ||
624 | else | ||
625 | lpfc_in_buf_free(phba, | ||
626 | dmabuf); | ||
627 | break; | ||
628 | default: | ||
629 | if (!(phba->sli3_options & | ||
630 | LPFC_SLI3_HBQ_ENABLED)) | ||
631 | lpfc_post_buffer(phba, | ||
632 | pring, | ||
633 | 1); | ||
634 | break; | ||
635 | } | ||
636 | } | ||
637 | } | ||
638 | } | ||
639 | |||
640 | mutex_lock(&phba->ct_event_mutex); | ||
641 | if (phba->sli_rev == LPFC_SLI_REV4) { | ||
642 | evt_dat->immed_dat = phba->ctx_idx; | ||
643 | phba->ctx_idx = (phba->ctx_idx + 1) % 64; | ||
644 | phba->ct_ctx[evt_dat->immed_dat].oxid = | ||
645 | piocbq->iocb.ulpContext; | ||
646 | phba->ct_ctx[evt_dat->immed_dat].SID = | ||
647 | piocbq->iocb.un.rcvels.remoteID; | ||
648 | } else | ||
649 | evt_dat->immed_dat = piocbq->iocb.ulpContext; | ||
650 | |||
651 | evt_dat->type = FC_REG_CT_EVENT; | ||
652 | list_add(&evt_dat->node, &evt->events_to_see); | ||
653 | wake_up_interruptible(&evt->wq); | ||
654 | lpfc_ct_event_unref(evt); | ||
655 | if (evt_req_id == SLI_CT_ELX_LOOPBACK) | ||
656 | break; | ||
657 | } | ||
658 | mutex_unlock(&phba->ct_event_mutex); | ||
659 | |||
660 | error_ct_unsol_exit: | ||
661 | if (!list_empty(&head)) | ||
662 | list_del(&head); | ||
663 | |||
664 | return; | ||
665 | } | ||
666 | |||
667 | /** | ||
668 | * lpfc_bsg_set_event - process a SET_EVENT bsg vendor command | ||
669 | * @job: SET_EVENT fc_bsg_job | ||
670 | */ | ||
671 | static int | ||
672 | lpfc_bsg_set_event(struct fc_bsg_job *job) | ||
673 | { | ||
674 | struct lpfc_vport *vport = (struct lpfc_vport *)job->shost->hostdata; | ||
675 | struct lpfc_hba *phba = vport->phba; | ||
676 | struct set_ct_event *event_req; | ||
677 | struct lpfc_ct_event *evt; | ||
678 | int rc = 0; | ||
679 | |||
680 | if (job->request_len < | ||
681 | sizeof(struct fc_bsg_request) + sizeof(struct set_ct_event)) { | ||
682 | lpfc_printf_log(phba, KERN_WARNING, LOG_LIBDFC, | ||
683 | "2612 Received SET_CT_EVENT below minimum " | ||
684 | "size\n"); | ||
685 | return -EINVAL; | ||
686 | } | ||
687 | |||
688 | event_req = (struct set_ct_event *) | ||
689 | job->request->rqst_data.h_vendor.vendor_cmd; | ||
690 | |||
691 | mutex_lock(&phba->ct_event_mutex); | ||
692 | list_for_each_entry(evt, &phba->ct_ev_waiters, node) { | ||
693 | if (evt->reg_id == event_req->ev_reg_id) { | ||
694 | lpfc_ct_event_ref(evt); | ||
695 | evt->wait_time_stamp = jiffies; | ||
696 | break; | ||
697 | } | ||
698 | } | ||
699 | mutex_unlock(&phba->ct_event_mutex); | ||
700 | |||
701 | if (&evt->node == &phba->ct_ev_waiters) { | ||
702 | /* no event waiting struct yet - first call */ | ||
703 | evt = lpfc_ct_event_new(event_req->ev_reg_id, | ||
704 | event_req->ev_req_id); | ||
705 | if (!evt) { | ||
706 | lpfc_printf_log(phba, KERN_WARNING, LOG_LIBDFC, | ||
707 | "2617 Failed allocation of event " | ||
708 | "waiter\n"); | ||
709 | return -ENOMEM; | ||
710 | } | ||
711 | |||
712 | mutex_lock(&phba->ct_event_mutex); | ||
713 | list_add(&evt->node, &phba->ct_ev_waiters); | ||
714 | lpfc_ct_event_ref(evt); | ||
715 | mutex_unlock(&phba->ct_event_mutex); | ||
716 | } | ||
717 | |||
718 | evt->waiting = 1; | ||
719 | if (wait_event_interruptible(evt->wq, | ||
720 | !list_empty(&evt->events_to_see))) { | ||
721 | mutex_lock(&phba->ct_event_mutex); | ||
722 | lpfc_ct_event_unref(evt); /* release ref */ | ||
723 | lpfc_ct_event_unref(evt); /* delete */ | ||
724 | mutex_unlock(&phba->ct_event_mutex); | ||
725 | rc = -EINTR; | ||
726 | goto set_event_out; | ||
727 | } | ||
728 | |||
729 | evt->wait_time_stamp = jiffies; | ||
730 | evt->waiting = 0; | ||
731 | |||
732 | mutex_lock(&phba->ct_event_mutex); | ||
733 | list_move(evt->events_to_see.prev, &evt->events_to_get); | ||
734 | lpfc_ct_event_unref(evt); /* release ref */ | ||
735 | mutex_unlock(&phba->ct_event_mutex); | ||
736 | |||
737 | set_event_out: | ||
738 | /* set_event carries no reply payload */ | ||
739 | job->reply->reply_payload_rcv_len = 0; | ||
740 | /* make error code available to userspace */ | ||
741 | job->reply->result = rc; | ||
742 | /* complete the job back to userspace */ | ||
743 | job->job_done(job); | ||
744 | |||
745 | return 0; | ||
746 | } | ||
747 | |||
748 | /** | ||
749 | * lpfc_bsg_get_event - process a GET_EVENT bsg vendor command | ||
750 | * @job: GET_EVENT fc_bsg_job | ||
751 | */ | ||
752 | static int | ||
753 | lpfc_bsg_get_event(struct fc_bsg_job *job) | ||
754 | { | ||
755 | struct lpfc_vport *vport = (struct lpfc_vport *)job->shost->hostdata; | ||
756 | struct lpfc_hba *phba = vport->phba; | ||
757 | struct get_ct_event *event_req; | ||
758 | struct get_ct_event_reply *event_reply; | ||
759 | struct lpfc_ct_event *evt; | ||
760 | struct event_data *evt_dat = NULL; | ||
761 | int rc = 0; | ||
762 | |||
763 | if (job->request_len < | ||
764 | sizeof(struct fc_bsg_request) + sizeof(struct get_ct_event)) { | ||
765 | lpfc_printf_log(phba, KERN_WARNING, LOG_LIBDFC, | ||
766 | "2613 Received GET_CT_EVENT request below " | ||
767 | "minimum size\n"); | ||
768 | return -EINVAL; | ||
769 | } | ||
770 | |||
771 | event_req = (struct get_ct_event *) | ||
772 | job->request->rqst_data.h_vendor.vendor_cmd; | ||
773 | |||
774 | event_reply = (struct get_ct_event_reply *) | ||
775 | job->reply->reply_data.vendor_reply.vendor_rsp; | ||
776 | |||
777 | mutex_lock(&phba->ct_event_mutex); | ||
778 | list_for_each_entry(evt, &phba->ct_ev_waiters, node) { | ||
779 | if (evt->reg_id == event_req->ev_reg_id) { | ||
780 | if (list_empty(&evt->events_to_get)) | ||
781 | break; | ||
782 | lpfc_ct_event_ref(evt); | ||
783 | evt->wait_time_stamp = jiffies; | ||
784 | evt_dat = list_entry(evt->events_to_get.prev, | ||
785 | struct event_data, node); | ||
786 | list_del(&evt_dat->node); | ||
787 | break; | ||
788 | } | ||
789 | } | ||
790 | mutex_unlock(&phba->ct_event_mutex); | ||
791 | |||
792 | if (!evt_dat) { | ||
793 | job->reply->reply_payload_rcv_len = 0; | ||
794 | rc = -ENOENT; | ||
795 | goto error_get_event_exit; | ||
796 | } | ||
797 | |||
798 | if (evt_dat->len > job->reply_payload.payload_len) { | ||
799 | evt_dat->len = job->reply_payload.payload_len; | ||
800 | lpfc_printf_log(phba, KERN_WARNING, LOG_LIBDFC, | ||
801 | "2618 Truncated event data at %d " | ||
802 | "bytes\n", | ||
803 | job->reply_payload.payload_len); | ||
804 | } | ||
805 | |||
806 | event_reply->immed_data = evt_dat->immed_dat; | ||
807 | |||
808 | if (evt_dat->len > 0) | ||
809 | job->reply->reply_payload_rcv_len = | ||
810 | sg_copy_from_buffer(job->reply_payload.sg_list, | ||
811 | job->reply_payload.sg_cnt, | ||
812 | evt_dat->data, evt_dat->len); | ||
813 | else | ||
814 | job->reply->reply_payload_rcv_len = 0; | ||
815 | rc = 0; | ||
816 | |||
817 | if (evt_dat) | ||
818 | kfree(evt_dat->data); | ||
819 | kfree(evt_dat); | ||
820 | mutex_lock(&phba->ct_event_mutex); | ||
821 | lpfc_ct_event_unref(evt); | ||
822 | mutex_unlock(&phba->ct_event_mutex); | ||
823 | |||
824 | error_get_event_exit: | ||
825 | /* make error code available to userspace */ | ||
826 | job->reply->result = rc; | ||
827 | /* complete the job back to userspace */ | ||
828 | job->job_done(job); | ||
829 | |||
830 | return rc; | ||
831 | } | ||
832 | |||
833 | /** | ||
834 | * lpfc_bsg_hst_vendor - process a vendor-specific fc_bsg_job | ||
835 | * @job: fc_bsg_job to handle | ||
836 | */ | ||
837 | static int | ||
838 | lpfc_bsg_hst_vendor(struct fc_bsg_job *job) | ||
839 | { | ||
840 | int command = job->request->rqst_data.h_vendor.vendor_cmd[0]; | ||
841 | |||
842 | switch (command) { | ||
843 | case LPFC_BSG_VENDOR_SET_CT_EVENT: | ||
844 | return lpfc_bsg_set_event(job); | ||
845 | break; | ||
846 | |||
847 | case LPFC_BSG_VENDOR_GET_CT_EVENT: | ||
848 | return lpfc_bsg_get_event(job); | ||
849 | break; | ||
850 | |||
851 | default: | ||
852 | return -EINVAL; | ||
853 | } | ||
854 | } | ||
855 | |||
856 | /** | ||
857 | * lpfc_bsg_request - handle a bsg request from the FC transport | ||
858 | * @job: fc_bsg_job to handle | ||
859 | */ | ||
860 | int | ||
861 | lpfc_bsg_request(struct fc_bsg_job *job) | ||
862 | { | ||
863 | uint32_t msgcode; | ||
864 | int rc = -EINVAL; | ||
865 | |||
866 | msgcode = job->request->msgcode; | ||
867 | |||
868 | switch (msgcode) { | ||
869 | case FC_BSG_HST_VENDOR: | ||
870 | rc = lpfc_bsg_hst_vendor(job); | ||
871 | break; | ||
872 | case FC_BSG_RPT_ELS: | ||
873 | rc = lpfc_bsg_rport_els(job); | ||
874 | break; | ||
875 | case FC_BSG_RPT_CT: | ||
876 | rc = lpfc_bsg_rport_ct(job); | ||
877 | break; | ||
878 | default: | ||
879 | break; | ||
880 | } | ||
881 | |||
882 | return rc; | ||
883 | } | ||
884 | |||
885 | /** | ||
886 | * lpfc_bsg_timeout - handle timeout of a bsg request from the FC transport | ||
887 | * @job: fc_bsg_job that has timed out | ||
888 | * | ||
889 | * This function just aborts the job's IOCB. The aborted IOCB will return to | ||
890 | * the waiting function which will handle passing the error back to userspace | ||
891 | */ | ||
892 | int | ||
893 | lpfc_bsg_timeout(struct fc_bsg_job *job) | ||
894 | { | ||
895 | struct lpfc_vport *vport = (struct lpfc_vport *)job->shost->hostdata; | ||
896 | struct lpfc_hba *phba = vport->phba; | ||
897 | struct lpfc_iocbq *cmdiocb = (struct lpfc_iocbq *)job->dd_data; | ||
898 | struct lpfc_sli_ring *pring = &phba->sli.ring[LPFC_ELS_RING]; | ||
899 | |||
900 | if (cmdiocb) | ||
901 | lpfc_sli_issue_abort_iotag(phba, pring, cmdiocb); | ||
902 | |||
903 | return 0; | ||
904 | } | ||
diff --git a/drivers/scsi/lpfc/lpfc_crtn.h b/drivers/scsi/lpfc/lpfc_crtn.h index d2a922997c0f..0830f37409a3 100644 --- a/drivers/scsi/lpfc/lpfc_crtn.h +++ b/drivers/scsi/lpfc/lpfc_crtn.h | |||
@@ -21,9 +21,11 @@ | |||
21 | typedef int (*node_filter)(struct lpfc_nodelist *, void *); | 21 | typedef int (*node_filter)(struct lpfc_nodelist *, void *); |
22 | 22 | ||
23 | struct fc_rport; | 23 | struct fc_rport; |
24 | void lpfc_dump_mem(struct lpfc_hba *, LPFC_MBOXQ_t *, uint16_t); | 24 | void lpfc_down_link(struct lpfc_hba *, LPFC_MBOXQ_t *); |
25 | void lpfc_sli_read_link_ste(struct lpfc_hba *); | ||
26 | void lpfc_dump_mem(struct lpfc_hba *, LPFC_MBOXQ_t *, uint16_t, uint16_t); | ||
25 | void lpfc_dump_wakeup_param(struct lpfc_hba *, LPFC_MBOXQ_t *); | 27 | void lpfc_dump_wakeup_param(struct lpfc_hba *, LPFC_MBOXQ_t *); |
26 | void lpfc_dump_static_vport(struct lpfc_hba *, LPFC_MBOXQ_t *, uint16_t); | 28 | int lpfc_dump_static_vport(struct lpfc_hba *, LPFC_MBOXQ_t *, uint16_t); |
27 | int lpfc_dump_fcoe_param(struct lpfc_hba *, struct lpfcMboxq *); | 29 | int lpfc_dump_fcoe_param(struct lpfc_hba *, struct lpfcMboxq *); |
28 | void lpfc_read_nv(struct lpfc_hba *, LPFC_MBOXQ_t *); | 30 | void lpfc_read_nv(struct lpfc_hba *, LPFC_MBOXQ_t *); |
29 | void lpfc_config_async(struct lpfc_hba *, LPFC_MBOXQ_t *, uint32_t); | 31 | void lpfc_config_async(struct lpfc_hba *, LPFC_MBOXQ_t *, uint32_t); |
@@ -135,6 +137,9 @@ int lpfc_els_disc_adisc(struct lpfc_vport *); | |||
135 | int lpfc_els_disc_plogi(struct lpfc_vport *); | 137 | int lpfc_els_disc_plogi(struct lpfc_vport *); |
136 | void lpfc_els_timeout(unsigned long); | 138 | void lpfc_els_timeout(unsigned long); |
137 | void lpfc_els_timeout_handler(struct lpfc_vport *); | 139 | void lpfc_els_timeout_handler(struct lpfc_vport *); |
140 | struct lpfc_iocbq *lpfc_prep_els_iocb(struct lpfc_vport *, uint8_t, uint16_t, | ||
141 | uint8_t, struct lpfc_nodelist *, | ||
142 | uint32_t, uint32_t); | ||
138 | void lpfc_hb_timeout_handler(struct lpfc_hba *); | 143 | void lpfc_hb_timeout_handler(struct lpfc_hba *); |
139 | 144 | ||
140 | void lpfc_ct_unsol_event(struct lpfc_hba *, struct lpfc_sli_ring *, | 145 | void lpfc_ct_unsol_event(struct lpfc_hba *, struct lpfc_sli_ring *, |
@@ -182,11 +187,12 @@ int lpfc_mbox_dev_check(struct lpfc_hba *); | |||
182 | int lpfc_mbox_tmo_val(struct lpfc_hba *, int); | 187 | int lpfc_mbox_tmo_val(struct lpfc_hba *, int); |
183 | void lpfc_init_vfi(struct lpfcMboxq *, struct lpfc_vport *); | 188 | void lpfc_init_vfi(struct lpfcMboxq *, struct lpfc_vport *); |
184 | void lpfc_reg_vfi(struct lpfcMboxq *, struct lpfc_vport *, dma_addr_t); | 189 | void lpfc_reg_vfi(struct lpfcMboxq *, struct lpfc_vport *, dma_addr_t); |
185 | void lpfc_init_vpi(struct lpfcMboxq *, uint16_t); | 190 | void lpfc_init_vpi(struct lpfc_hba *, struct lpfcMboxq *, uint16_t); |
186 | void lpfc_unreg_vfi(struct lpfcMboxq *, uint16_t); | 191 | void lpfc_unreg_vfi(struct lpfcMboxq *, uint16_t); |
187 | void lpfc_reg_fcfi(struct lpfc_hba *, struct lpfcMboxq *); | 192 | void lpfc_reg_fcfi(struct lpfc_hba *, struct lpfcMboxq *); |
188 | void lpfc_unreg_fcfi(struct lpfcMboxq *, uint16_t); | 193 | void lpfc_unreg_fcfi(struct lpfcMboxq *, uint16_t); |
189 | void lpfc_resume_rpi(struct lpfcMboxq *, struct lpfc_nodelist *); | 194 | void lpfc_resume_rpi(struct lpfcMboxq *, struct lpfc_nodelist *); |
195 | int lpfc_check_pending_fcoe_event(struct lpfc_hba *, uint8_t); | ||
190 | 196 | ||
191 | void lpfc_config_hbq(struct lpfc_hba *, uint32_t, struct lpfc_hbq_init *, | 197 | void lpfc_config_hbq(struct lpfc_hba *, uint32_t, struct lpfc_hbq_init *, |
192 | uint32_t , LPFC_MBOXQ_t *); | 198 | uint32_t , LPFC_MBOXQ_t *); |
@@ -234,6 +240,7 @@ void lpfc_sli_def_mbox_cmpl(struct lpfc_hba *, LPFC_MBOXQ_t *); | |||
234 | int lpfc_sli_issue_iocb(struct lpfc_hba *, uint32_t, | 240 | int lpfc_sli_issue_iocb(struct lpfc_hba *, uint32_t, |
235 | struct lpfc_iocbq *, uint32_t); | 241 | struct lpfc_iocbq *, uint32_t); |
236 | void lpfc_sli_pcimem_bcopy(void *, void *, uint32_t); | 242 | void lpfc_sli_pcimem_bcopy(void *, void *, uint32_t); |
243 | void lpfc_sli_bemem_bcopy(void *, void *, uint32_t); | ||
237 | void lpfc_sli_abort_iocb_ring(struct lpfc_hba *, struct lpfc_sli_ring *); | 244 | void lpfc_sli_abort_iocb_ring(struct lpfc_hba *, struct lpfc_sli_ring *); |
238 | void lpfc_sli_flush_fcp_rings(struct lpfc_hba *); | 245 | void lpfc_sli_flush_fcp_rings(struct lpfc_hba *); |
239 | int lpfc_sli_ringpostbuf_put(struct lpfc_hba *, struct lpfc_sli_ring *, | 246 | int lpfc_sli_ringpostbuf_put(struct lpfc_hba *, struct lpfc_sli_ring *, |
@@ -360,3 +367,8 @@ void lpfc_start_fdiscs(struct lpfc_hba *phba); | |||
360 | #define HBA_EVENT_LINK_UP 2 | 367 | #define HBA_EVENT_LINK_UP 2 |
361 | #define HBA_EVENT_LINK_DOWN 3 | 368 | #define HBA_EVENT_LINK_DOWN 3 |
362 | 369 | ||
370 | /* functions to support SGIOv4/bsg interface */ | ||
371 | int lpfc_bsg_request(struct fc_bsg_job *); | ||
372 | int lpfc_bsg_timeout(struct fc_bsg_job *); | ||
373 | void lpfc_bsg_ct_unsol_event(struct lpfc_hba *, struct lpfc_sli_ring *, | ||
374 | struct lpfc_iocbq *); | ||
diff --git a/drivers/scsi/lpfc/lpfc_ct.c b/drivers/scsi/lpfc/lpfc_ct.c index 0e532f072eb3..9df7ed38e1be 100644 --- a/drivers/scsi/lpfc/lpfc_ct.c +++ b/drivers/scsi/lpfc/lpfc_ct.c | |||
@@ -97,6 +97,8 @@ lpfc_ct_unsol_event(struct lpfc_hba *phba, struct lpfc_sli_ring *pring, | |||
97 | struct list_head head; | 97 | struct list_head head; |
98 | struct lpfc_dmabuf *bdeBuf; | 98 | struct lpfc_dmabuf *bdeBuf; |
99 | 99 | ||
100 | lpfc_bsg_ct_unsol_event(phba, pring, piocbq); | ||
101 | |||
100 | if (unlikely(icmd->ulpStatus == IOSTAT_NEED_BUFFER)) { | 102 | if (unlikely(icmd->ulpStatus == IOSTAT_NEED_BUFFER)) { |
101 | lpfc_sli_hbqbuf_add_hbqs(phba, LPFC_ELS_HBQ); | 103 | lpfc_sli_hbqbuf_add_hbqs(phba, LPFC_ELS_HBQ); |
102 | } else if ((icmd->ulpStatus == IOSTAT_LOCAL_REJECT) && | 104 | } else if ((icmd->ulpStatus == IOSTAT_LOCAL_REJECT) && |
diff --git a/drivers/scsi/lpfc/lpfc_els.c b/drivers/scsi/lpfc/lpfc_els.c index f72fdf23bf1b..45337cd23feb 100644 --- a/drivers/scsi/lpfc/lpfc_els.c +++ b/drivers/scsi/lpfc/lpfc_els.c | |||
@@ -146,7 +146,7 @@ lpfc_els_chk_latt(struct lpfc_vport *vport) | |||
146 | * Pointer to the newly allocated/prepared els iocb data structure | 146 | * Pointer to the newly allocated/prepared els iocb data structure |
147 | * NULL - when els iocb data structure allocation/preparation failed | 147 | * NULL - when els iocb data structure allocation/preparation failed |
148 | **/ | 148 | **/ |
149 | static struct lpfc_iocbq * | 149 | struct lpfc_iocbq * |
150 | lpfc_prep_els_iocb(struct lpfc_vport *vport, uint8_t expectRsp, | 150 | lpfc_prep_els_iocb(struct lpfc_vport *vport, uint8_t expectRsp, |
151 | uint16_t cmdSize, uint8_t retry, | 151 | uint16_t cmdSize, uint8_t retry, |
152 | struct lpfc_nodelist *ndlp, uint32_t did, | 152 | struct lpfc_nodelist *ndlp, uint32_t did, |
diff --git a/drivers/scsi/lpfc/lpfc_hbadisc.c b/drivers/scsi/lpfc/lpfc_hbadisc.c index ed46b24a3380..e6a47e25b218 100644 --- a/drivers/scsi/lpfc/lpfc_hbadisc.c +++ b/drivers/scsi/lpfc/lpfc_hbadisc.c | |||
@@ -61,6 +61,7 @@ static uint8_t lpfcAlpaArray[] = { | |||
61 | 61 | ||
62 | static void lpfc_disc_timeout_handler(struct lpfc_vport *); | 62 | static void lpfc_disc_timeout_handler(struct lpfc_vport *); |
63 | static void lpfc_disc_flush_list(struct lpfc_vport *vport); | 63 | static void lpfc_disc_flush_list(struct lpfc_vport *vport); |
64 | static void lpfc_unregister_fcfi_cmpl(struct lpfc_hba *, LPFC_MBOXQ_t *); | ||
64 | 65 | ||
65 | void | 66 | void |
66 | lpfc_terminate_rport_io(struct fc_rport *rport) | 67 | lpfc_terminate_rport_io(struct fc_rport *rport) |
@@ -1009,9 +1010,15 @@ lpfc_mbx_cmpl_reg_fcfi(struct lpfc_hba *phba, LPFC_MBOXQ_t *mboxq) | |||
1009 | spin_lock_irqsave(&phba->hbalock, flags); | 1010 | spin_lock_irqsave(&phba->hbalock, flags); |
1010 | phba->fcf.fcf_flag |= FCF_REGISTERED; | 1011 | phba->fcf.fcf_flag |= FCF_REGISTERED; |
1011 | spin_unlock_irqrestore(&phba->hbalock, flags); | 1012 | spin_unlock_irqrestore(&phba->hbalock, flags); |
1013 | /* If there is a pending FCoE event, restart FCF table scan. */ | ||
1014 | if (lpfc_check_pending_fcoe_event(phba, 1)) { | ||
1015 | mempool_free(mboxq, phba->mbox_mem_pool); | ||
1016 | return; | ||
1017 | } | ||
1012 | if (vport->port_state != LPFC_FLOGI) { | 1018 | if (vport->port_state != LPFC_FLOGI) { |
1013 | spin_lock_irqsave(&phba->hbalock, flags); | 1019 | spin_lock_irqsave(&phba->hbalock, flags); |
1014 | phba->fcf.fcf_flag |= (FCF_DISCOVERED | FCF_IN_USE); | 1020 | phba->fcf.fcf_flag |= (FCF_DISCOVERED | FCF_IN_USE); |
1021 | phba->hba_flag &= ~FCF_DISC_INPROGRESS; | ||
1015 | spin_unlock_irqrestore(&phba->hbalock, flags); | 1022 | spin_unlock_irqrestore(&phba->hbalock, flags); |
1016 | lpfc_initial_flogi(vport); | 1023 | lpfc_initial_flogi(vport); |
1017 | } | 1024 | } |
@@ -1054,6 +1061,39 @@ lpfc_fab_name_match(uint8_t *fab_name, struct fcf_record *new_fcf_record) | |||
1054 | } | 1061 | } |
1055 | 1062 | ||
1056 | /** | 1063 | /** |
1064 | * lpfc_sw_name_match - Check if the fcf switch name match. | ||
1065 | * @fab_name: pointer to fabric name. | ||
1066 | * @new_fcf_record: pointer to fcf record. | ||
1067 | * | ||
1068 | * This routine compare the fcf record's switch name with provided | ||
1069 | * switch name. If the switch name are identical this function | ||
1070 | * returns 1 else return 0. | ||
1071 | **/ | ||
1072 | static uint32_t | ||
1073 | lpfc_sw_name_match(uint8_t *sw_name, struct fcf_record *new_fcf_record) | ||
1074 | { | ||
1075 | if ((sw_name[0] == | ||
1076 | bf_get(lpfc_fcf_record_switch_name_0, new_fcf_record)) && | ||
1077 | (sw_name[1] == | ||
1078 | bf_get(lpfc_fcf_record_switch_name_1, new_fcf_record)) && | ||
1079 | (sw_name[2] == | ||
1080 | bf_get(lpfc_fcf_record_switch_name_2, new_fcf_record)) && | ||
1081 | (sw_name[3] == | ||
1082 | bf_get(lpfc_fcf_record_switch_name_3, new_fcf_record)) && | ||
1083 | (sw_name[4] == | ||
1084 | bf_get(lpfc_fcf_record_switch_name_4, new_fcf_record)) && | ||
1085 | (sw_name[5] == | ||
1086 | bf_get(lpfc_fcf_record_switch_name_5, new_fcf_record)) && | ||
1087 | (sw_name[6] == | ||
1088 | bf_get(lpfc_fcf_record_switch_name_6, new_fcf_record)) && | ||
1089 | (sw_name[7] == | ||
1090 | bf_get(lpfc_fcf_record_switch_name_7, new_fcf_record))) | ||
1091 | return 1; | ||
1092 | else | ||
1093 | return 0; | ||
1094 | } | ||
1095 | |||
1096 | /** | ||
1057 | * lpfc_mac_addr_match - Check if the fcf mac address match. | 1097 | * lpfc_mac_addr_match - Check if the fcf mac address match. |
1058 | * @phba: pointer to lpfc hba data structure. | 1098 | * @phba: pointer to lpfc hba data structure. |
1059 | * @new_fcf_record: pointer to fcf record. | 1099 | * @new_fcf_record: pointer to fcf record. |
@@ -1123,6 +1163,22 @@ lpfc_copy_fcf_record(struct lpfc_hba *phba, struct fcf_record *new_fcf_record) | |||
1123 | bf_get(lpfc_fcf_record_mac_5, new_fcf_record); | 1163 | bf_get(lpfc_fcf_record_mac_5, new_fcf_record); |
1124 | phba->fcf.fcf_indx = bf_get(lpfc_fcf_record_fcf_index, new_fcf_record); | 1164 | phba->fcf.fcf_indx = bf_get(lpfc_fcf_record_fcf_index, new_fcf_record); |
1125 | phba->fcf.priority = new_fcf_record->fip_priority; | 1165 | phba->fcf.priority = new_fcf_record->fip_priority; |
1166 | phba->fcf.switch_name[0] = | ||
1167 | bf_get(lpfc_fcf_record_switch_name_0, new_fcf_record); | ||
1168 | phba->fcf.switch_name[1] = | ||
1169 | bf_get(lpfc_fcf_record_switch_name_1, new_fcf_record); | ||
1170 | phba->fcf.switch_name[2] = | ||
1171 | bf_get(lpfc_fcf_record_switch_name_2, new_fcf_record); | ||
1172 | phba->fcf.switch_name[3] = | ||
1173 | bf_get(lpfc_fcf_record_switch_name_3, new_fcf_record); | ||
1174 | phba->fcf.switch_name[4] = | ||
1175 | bf_get(lpfc_fcf_record_switch_name_4, new_fcf_record); | ||
1176 | phba->fcf.switch_name[5] = | ||
1177 | bf_get(lpfc_fcf_record_switch_name_5, new_fcf_record); | ||
1178 | phba->fcf.switch_name[6] = | ||
1179 | bf_get(lpfc_fcf_record_switch_name_6, new_fcf_record); | ||
1180 | phba->fcf.switch_name[7] = | ||
1181 | bf_get(lpfc_fcf_record_switch_name_7, new_fcf_record); | ||
1126 | } | 1182 | } |
1127 | 1183 | ||
1128 | /** | 1184 | /** |
@@ -1150,6 +1206,7 @@ lpfc_register_fcf(struct lpfc_hba *phba) | |||
1150 | /* The FCF is already registered, start discovery */ | 1206 | /* The FCF is already registered, start discovery */ |
1151 | if (phba->fcf.fcf_flag & FCF_REGISTERED) { | 1207 | if (phba->fcf.fcf_flag & FCF_REGISTERED) { |
1152 | phba->fcf.fcf_flag |= (FCF_DISCOVERED | FCF_IN_USE); | 1208 | phba->fcf.fcf_flag |= (FCF_DISCOVERED | FCF_IN_USE); |
1209 | phba->hba_flag &= ~FCF_DISC_INPROGRESS; | ||
1153 | spin_unlock_irqrestore(&phba->hbalock, flags); | 1210 | spin_unlock_irqrestore(&phba->hbalock, flags); |
1154 | if (phba->pport->port_state != LPFC_FLOGI) | 1211 | if (phba->pport->port_state != LPFC_FLOGI) |
1155 | lpfc_initial_flogi(phba->pport); | 1212 | lpfc_initial_flogi(phba->pport); |
@@ -1239,9 +1296,12 @@ lpfc_match_fcf_conn_list(struct lpfc_hba *phba, | |||
1239 | 1296 | ||
1240 | if ((conn_entry->conn_rec.flags & FCFCNCT_FBNM_VALID) && | 1297 | if ((conn_entry->conn_rec.flags & FCFCNCT_FBNM_VALID) && |
1241 | !lpfc_fab_name_match(conn_entry->conn_rec.fabric_name, | 1298 | !lpfc_fab_name_match(conn_entry->conn_rec.fabric_name, |
1242 | new_fcf_record)) | 1299 | new_fcf_record)) |
1300 | continue; | ||
1301 | if ((conn_entry->conn_rec.flags & FCFCNCT_SWNM_VALID) && | ||
1302 | !lpfc_sw_name_match(conn_entry->conn_rec.switch_name, | ||
1303 | new_fcf_record)) | ||
1243 | continue; | 1304 | continue; |
1244 | |||
1245 | if (conn_entry->conn_rec.flags & FCFCNCT_VLAN_VALID) { | 1305 | if (conn_entry->conn_rec.flags & FCFCNCT_VLAN_VALID) { |
1246 | /* | 1306 | /* |
1247 | * If the vlan bit map does not have the bit set for the | 1307 | * If the vlan bit map does not have the bit set for the |
@@ -1336,6 +1396,60 @@ lpfc_match_fcf_conn_list(struct lpfc_hba *phba, | |||
1336 | } | 1396 | } |
1337 | 1397 | ||
1338 | /** | 1398 | /** |
1399 | * lpfc_check_pending_fcoe_event - Check if there is pending fcoe event. | ||
1400 | * @phba: pointer to lpfc hba data structure. | ||
1401 | * @unreg_fcf: Unregister FCF if FCF table need to be re-scaned. | ||
1402 | * | ||
1403 | * This function check if there is any fcoe event pending while driver | ||
1404 | * scan FCF entries. If there is any pending event, it will restart the | ||
1405 | * FCF saning and return 1 else return 0. | ||
1406 | */ | ||
1407 | int | ||
1408 | lpfc_check_pending_fcoe_event(struct lpfc_hba *phba, uint8_t unreg_fcf) | ||
1409 | { | ||
1410 | LPFC_MBOXQ_t *mbox; | ||
1411 | int rc; | ||
1412 | /* | ||
1413 | * If the Link is up and no FCoE events while in the | ||
1414 | * FCF discovery, no need to restart FCF discovery. | ||
1415 | */ | ||
1416 | if ((phba->link_state >= LPFC_LINK_UP) && | ||
1417 | (phba->fcoe_eventtag == phba->fcoe_eventtag_at_fcf_scan)) | ||
1418 | return 0; | ||
1419 | |||
1420 | spin_lock_irq(&phba->hbalock); | ||
1421 | phba->fcf.fcf_flag &= ~FCF_AVAILABLE; | ||
1422 | spin_unlock_irq(&phba->hbalock); | ||
1423 | |||
1424 | if (phba->link_state >= LPFC_LINK_UP) | ||
1425 | lpfc_sli4_read_fcf_record(phba, LPFC_FCOE_FCF_GET_FIRST); | ||
1426 | |||
1427 | if (unreg_fcf) { | ||
1428 | spin_lock_irq(&phba->hbalock); | ||
1429 | phba->fcf.fcf_flag &= ~FCF_REGISTERED; | ||
1430 | spin_unlock_irq(&phba->hbalock); | ||
1431 | mbox = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); | ||
1432 | if (!mbox) { | ||
1433 | lpfc_printf_log(phba, KERN_ERR, | ||
1434 | LOG_DISCOVERY|LOG_MBOX, | ||
1435 | "2610 UNREG_FCFI mbox allocation failed\n"); | ||
1436 | return 1; | ||
1437 | } | ||
1438 | lpfc_unreg_fcfi(mbox, phba->fcf.fcfi); | ||
1439 | mbox->vport = phba->pport; | ||
1440 | mbox->mbox_cmpl = lpfc_unregister_fcfi_cmpl; | ||
1441 | rc = lpfc_sli_issue_mbox(phba, mbox, MBX_NOWAIT); | ||
1442 | if (rc == MBX_NOT_FINISHED) { | ||
1443 | lpfc_printf_log(phba, KERN_ERR, LOG_DISCOVERY|LOG_MBOX, | ||
1444 | "2611 UNREG_FCFI issue mbox failed\n"); | ||
1445 | mempool_free(mbox, phba->mbox_mem_pool); | ||
1446 | } | ||
1447 | } | ||
1448 | |||
1449 | return 1; | ||
1450 | } | ||
1451 | |||
1452 | /** | ||
1339 | * lpfc_mbx_cmpl_read_fcf_record - Completion handler for read_fcf mbox. | 1453 | * lpfc_mbx_cmpl_read_fcf_record - Completion handler for read_fcf mbox. |
1340 | * @phba: pointer to lpfc hba data structure. | 1454 | * @phba: pointer to lpfc hba data structure. |
1341 | * @mboxq: pointer to mailbox object. | 1455 | * @mboxq: pointer to mailbox object. |
@@ -1367,6 +1481,12 @@ lpfc_mbx_cmpl_read_fcf_record(struct lpfc_hba *phba, LPFC_MBOXQ_t *mboxq) | |||
1367 | unsigned long flags; | 1481 | unsigned long flags; |
1368 | uint16_t vlan_id; | 1482 | uint16_t vlan_id; |
1369 | 1483 | ||
1484 | /* If there is pending FCoE event restart FCF table scan */ | ||
1485 | if (lpfc_check_pending_fcoe_event(phba, 0)) { | ||
1486 | lpfc_sli4_mbox_cmd_free(phba, mboxq); | ||
1487 | return; | ||
1488 | } | ||
1489 | |||
1370 | /* Get the first SGE entry from the non-embedded DMA memory. This | 1490 | /* Get the first SGE entry from the non-embedded DMA memory. This |
1371 | * routine only uses a single SGE. | 1491 | * routine only uses a single SGE. |
1372 | */ | 1492 | */ |
@@ -1424,7 +1544,9 @@ lpfc_mbx_cmpl_read_fcf_record(struct lpfc_hba *phba, LPFC_MBOXQ_t *mboxq) | |||
1424 | spin_lock_irqsave(&phba->hbalock, flags); | 1544 | spin_lock_irqsave(&phba->hbalock, flags); |
1425 | if (phba->fcf.fcf_flag & FCF_IN_USE) { | 1545 | if (phba->fcf.fcf_flag & FCF_IN_USE) { |
1426 | if (lpfc_fab_name_match(phba->fcf.fabric_name, | 1546 | if (lpfc_fab_name_match(phba->fcf.fabric_name, |
1427 | new_fcf_record) && | 1547 | new_fcf_record) && |
1548 | lpfc_sw_name_match(phba->fcf.switch_name, | ||
1549 | new_fcf_record) && | ||
1428 | lpfc_mac_addr_match(phba, new_fcf_record)) { | 1550 | lpfc_mac_addr_match(phba, new_fcf_record)) { |
1429 | phba->fcf.fcf_flag |= FCF_AVAILABLE; | 1551 | phba->fcf.fcf_flag |= FCF_AVAILABLE; |
1430 | spin_unlock_irqrestore(&phba->hbalock, flags); | 1552 | spin_unlock_irqrestore(&phba->hbalock, flags); |
@@ -1464,9 +1586,9 @@ lpfc_mbx_cmpl_read_fcf_record(struct lpfc_hba *phba, LPFC_MBOXQ_t *mboxq) | |||
1464 | * If there is a record with lower priority value for | 1586 | * If there is a record with lower priority value for |
1465 | * the current FCF, use that record. | 1587 | * the current FCF, use that record. |
1466 | */ | 1588 | */ |
1467 | if (lpfc_fab_name_match(phba->fcf.fabric_name, new_fcf_record) | 1589 | if (lpfc_fab_name_match(phba->fcf.fabric_name, |
1468 | && (new_fcf_record->fip_priority < | 1590 | new_fcf_record) && |
1469 | phba->fcf.priority)) { | 1591 | (new_fcf_record->fip_priority < phba->fcf.priority)) { |
1470 | /* Use this FCF record */ | 1592 | /* Use this FCF record */ |
1471 | lpfc_copy_fcf_record(phba, new_fcf_record); | 1593 | lpfc_copy_fcf_record(phba, new_fcf_record); |
1472 | phba->fcf.addr_mode = addr_mode; | 1594 | phba->fcf.addr_mode = addr_mode; |
@@ -1512,6 +1634,39 @@ out: | |||
1512 | } | 1634 | } |
1513 | 1635 | ||
1514 | /** | 1636 | /** |
1637 | * lpfc_init_vpi_cmpl - Completion handler for init_vpi mbox command. | ||
1638 | * @phba: pointer to lpfc hba data structure. | ||
1639 | * @mboxq: pointer to mailbox data structure. | ||
1640 | * | ||
1641 | * This function handles completion of init vpi mailbox command. | ||
1642 | */ | ||
1643 | static void | ||
1644 | lpfc_init_vpi_cmpl(struct lpfc_hba *phba, LPFC_MBOXQ_t *mboxq) | ||
1645 | { | ||
1646 | struct lpfc_vport *vport = mboxq->vport; | ||
1647 | if (mboxq->u.mb.mbxStatus) { | ||
1648 | lpfc_printf_vlog(vport, KERN_ERR, | ||
1649 | LOG_MBOX, | ||
1650 | "2609 Init VPI mailbox failed 0x%x\n", | ||
1651 | mboxq->u.mb.mbxStatus); | ||
1652 | mempool_free(mboxq, phba->mbox_mem_pool); | ||
1653 | lpfc_vport_set_state(vport, FC_VPORT_FAILED); | ||
1654 | return; | ||
1655 | } | ||
1656 | vport->fc_flag &= ~FC_VPORT_NEEDS_INIT_VPI; | ||
1657 | |||
1658 | if (phba->link_flag & LS_NPIV_FAB_SUPPORTED) | ||
1659 | lpfc_initial_fdisc(vport); | ||
1660 | else { | ||
1661 | lpfc_vport_set_state(vport, FC_VPORT_NO_FABRIC_SUPP); | ||
1662 | lpfc_printf_vlog(vport, KERN_ERR, | ||
1663 | LOG_ELS, | ||
1664 | "2606 No NPIV Fabric support\n"); | ||
1665 | } | ||
1666 | return; | ||
1667 | } | ||
1668 | |||
1669 | /** | ||
1515 | * lpfc_start_fdiscs - send fdiscs for each vports on this port. | 1670 | * lpfc_start_fdiscs - send fdiscs for each vports on this port. |
1516 | * @phba: pointer to lpfc hba data structure. | 1671 | * @phba: pointer to lpfc hba data structure. |
1517 | * | 1672 | * |
@@ -1523,6 +1678,8 @@ lpfc_start_fdiscs(struct lpfc_hba *phba) | |||
1523 | { | 1678 | { |
1524 | struct lpfc_vport **vports; | 1679 | struct lpfc_vport **vports; |
1525 | int i; | 1680 | int i; |
1681 | LPFC_MBOXQ_t *mboxq; | ||
1682 | int rc; | ||
1526 | 1683 | ||
1527 | vports = lpfc_create_vport_work_array(phba); | 1684 | vports = lpfc_create_vport_work_array(phba); |
1528 | if (vports != NULL) { | 1685 | if (vports != NULL) { |
@@ -1540,6 +1697,29 @@ lpfc_start_fdiscs(struct lpfc_hba *phba) | |||
1540 | FC_VPORT_LINKDOWN); | 1697 | FC_VPORT_LINKDOWN); |
1541 | continue; | 1698 | continue; |
1542 | } | 1699 | } |
1700 | if (vports[i]->fc_flag & FC_VPORT_NEEDS_INIT_VPI) { | ||
1701 | mboxq = mempool_alloc(phba->mbox_mem_pool, | ||
1702 | GFP_KERNEL); | ||
1703 | if (!mboxq) { | ||
1704 | lpfc_printf_vlog(vports[i], KERN_ERR, | ||
1705 | LOG_MBOX, "2607 Failed to allocate " | ||
1706 | "init_vpi mailbox\n"); | ||
1707 | continue; | ||
1708 | } | ||
1709 | lpfc_init_vpi(phba, mboxq, vports[i]->vpi); | ||
1710 | mboxq->vport = vports[i]; | ||
1711 | mboxq->mbox_cmpl = lpfc_init_vpi_cmpl; | ||
1712 | rc = lpfc_sli_issue_mbox(phba, mboxq, | ||
1713 | MBX_NOWAIT); | ||
1714 | if (rc == MBX_NOT_FINISHED) { | ||
1715 | lpfc_printf_vlog(vports[i], KERN_ERR, | ||
1716 | LOG_MBOX, "2608 Failed to issue " | ||
1717 | "init_vpi mailbox\n"); | ||
1718 | mempool_free(mboxq, | ||
1719 | phba->mbox_mem_pool); | ||
1720 | } | ||
1721 | continue; | ||
1722 | } | ||
1543 | if (phba->link_flag & LS_NPIV_FAB_SUPPORTED) | 1723 | if (phba->link_flag & LS_NPIV_FAB_SUPPORTED) |
1544 | lpfc_initial_fdisc(vports[i]); | 1724 | lpfc_initial_fdisc(vports[i]); |
1545 | else { | 1725 | else { |
@@ -1769,6 +1949,7 @@ lpfc_mbx_process_link_up(struct lpfc_hba *phba, READ_LA_VAR *la) | |||
1769 | goto out; | 1949 | goto out; |
1770 | } | 1950 | } |
1771 | } else { | 1951 | } else { |
1952 | vport->port_state = LPFC_VPORT_UNKNOWN; | ||
1772 | /* | 1953 | /* |
1773 | * Add the driver's default FCF record at FCF index 0 now. This | 1954 | * Add the driver's default FCF record at FCF index 0 now. This |
1774 | * is phase 1 implementation that support FCF index 0 and driver | 1955 | * is phase 1 implementation that support FCF index 0 and driver |
@@ -1804,6 +1985,12 @@ lpfc_mbx_process_link_up(struct lpfc_hba *phba, READ_LA_VAR *la) | |||
1804 | * The driver is expected to do FIP/FCF. Call the port | 1985 | * The driver is expected to do FIP/FCF. Call the port |
1805 | * and get the FCF Table. | 1986 | * and get the FCF Table. |
1806 | */ | 1987 | */ |
1988 | spin_lock_irq(&phba->hbalock); | ||
1989 | if (phba->hba_flag & FCF_DISC_INPROGRESS) { | ||
1990 | spin_unlock_irq(&phba->hbalock); | ||
1991 | return; | ||
1992 | } | ||
1993 | spin_unlock_irq(&phba->hbalock); | ||
1807 | rc = lpfc_sli4_read_fcf_record(phba, | 1994 | rc = lpfc_sli4_read_fcf_record(phba, |
1808 | LPFC_FCOE_FCF_GET_FIRST); | 1995 | LPFC_FCOE_FCF_GET_FIRST); |
1809 | if (rc) | 1996 | if (rc) |
@@ -2113,13 +2300,15 @@ lpfc_create_static_vport(struct lpfc_hba *phba) | |||
2113 | LPFC_MBOXQ_t *pmb = NULL; | 2300 | LPFC_MBOXQ_t *pmb = NULL; |
2114 | MAILBOX_t *mb; | 2301 | MAILBOX_t *mb; |
2115 | struct static_vport_info *vport_info; | 2302 | struct static_vport_info *vport_info; |
2116 | int rc, i; | 2303 | int rc = 0, i; |
2117 | struct fc_vport_identifiers vport_id; | 2304 | struct fc_vport_identifiers vport_id; |
2118 | struct fc_vport *new_fc_vport; | 2305 | struct fc_vport *new_fc_vport; |
2119 | struct Scsi_Host *shost; | 2306 | struct Scsi_Host *shost; |
2120 | struct lpfc_vport *vport; | 2307 | struct lpfc_vport *vport; |
2121 | uint16_t offset = 0; | 2308 | uint16_t offset = 0; |
2122 | uint8_t *vport_buff; | 2309 | uint8_t *vport_buff; |
2310 | struct lpfc_dmabuf *mp; | ||
2311 | uint32_t byte_count = 0; | ||
2123 | 2312 | ||
2124 | pmb = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); | 2313 | pmb = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); |
2125 | if (!pmb) { | 2314 | if (!pmb) { |
@@ -2142,7 +2331,9 @@ lpfc_create_static_vport(struct lpfc_hba *phba) | |||
2142 | 2331 | ||
2143 | vport_buff = (uint8_t *) vport_info; | 2332 | vport_buff = (uint8_t *) vport_info; |
2144 | do { | 2333 | do { |
2145 | lpfc_dump_static_vport(phba, pmb, offset); | 2334 | if (lpfc_dump_static_vport(phba, pmb, offset)) |
2335 | goto out; | ||
2336 | |||
2146 | pmb->vport = phba->pport; | 2337 | pmb->vport = phba->pport; |
2147 | rc = lpfc_sli_issue_mbox_wait(phba, pmb, LPFC_MBOX_TMO); | 2338 | rc = lpfc_sli_issue_mbox_wait(phba, pmb, LPFC_MBOX_TMO); |
2148 | 2339 | ||
@@ -2155,17 +2346,30 @@ lpfc_create_static_vport(struct lpfc_hba *phba) | |||
2155 | goto out; | 2346 | goto out; |
2156 | } | 2347 | } |
2157 | 2348 | ||
2158 | if (mb->un.varDmp.word_cnt > | 2349 | if (phba->sli_rev == LPFC_SLI_REV4) { |
2159 | sizeof(struct static_vport_info) - offset) | 2350 | byte_count = pmb->u.mqe.un.mb_words[5]; |
2160 | mb->un.varDmp.word_cnt = | 2351 | mp = (struct lpfc_dmabuf *) pmb->context2; |
2161 | sizeof(struct static_vport_info) - offset; | 2352 | if (byte_count > sizeof(struct static_vport_info) - |
2162 | 2353 | offset) | |
2163 | lpfc_sli_pcimem_bcopy(((uint8_t *)mb) + DMP_RSP_OFFSET, | 2354 | byte_count = sizeof(struct static_vport_info) |
2164 | vport_buff + offset, | 2355 | - offset; |
2165 | mb->un.varDmp.word_cnt); | 2356 | memcpy(vport_buff + offset, mp->virt, byte_count); |
2166 | offset += mb->un.varDmp.word_cnt; | 2357 | offset += byte_count; |
2358 | } else { | ||
2359 | if (mb->un.varDmp.word_cnt > | ||
2360 | sizeof(struct static_vport_info) - offset) | ||
2361 | mb->un.varDmp.word_cnt = | ||
2362 | sizeof(struct static_vport_info) | ||
2363 | - offset; | ||
2364 | byte_count = mb->un.varDmp.word_cnt; | ||
2365 | lpfc_sli_pcimem_bcopy(((uint8_t *)mb) + DMP_RSP_OFFSET, | ||
2366 | vport_buff + offset, | ||
2367 | byte_count); | ||
2368 | |||
2369 | offset += byte_count; | ||
2370 | } | ||
2167 | 2371 | ||
2168 | } while (mb->un.varDmp.word_cnt && | 2372 | } while (byte_count && |
2169 | offset < sizeof(struct static_vport_info)); | 2373 | offset < sizeof(struct static_vport_info)); |
2170 | 2374 | ||
2171 | 2375 | ||
@@ -2198,7 +2402,7 @@ lpfc_create_static_vport(struct lpfc_hba *phba) | |||
2198 | if (!new_fc_vport) { | 2402 | if (!new_fc_vport) { |
2199 | lpfc_printf_log(phba, KERN_WARNING, LOG_INIT, | 2403 | lpfc_printf_log(phba, KERN_WARNING, LOG_INIT, |
2200 | "0546 lpfc_create_static_vport failed to" | 2404 | "0546 lpfc_create_static_vport failed to" |
2201 | " create vport \n"); | 2405 | " create vport\n"); |
2202 | continue; | 2406 | continue; |
2203 | } | 2407 | } |
2204 | 2408 | ||
@@ -2207,16 +2411,15 @@ lpfc_create_static_vport(struct lpfc_hba *phba) | |||
2207 | } | 2411 | } |
2208 | 2412 | ||
2209 | out: | 2413 | out: |
2210 | /* | ||
2211 | * If this is timed out command, setting NULL to context2 tell SLI | ||
2212 | * layer not to use this buffer. | ||
2213 | */ | ||
2214 | spin_lock_irq(&phba->hbalock); | ||
2215 | pmb->context2 = NULL; | ||
2216 | spin_unlock_irq(&phba->hbalock); | ||
2217 | kfree(vport_info); | 2414 | kfree(vport_info); |
2218 | if (rc != MBX_TIMEOUT) | 2415 | if (rc != MBX_TIMEOUT) { |
2416 | if (pmb->context2) { | ||
2417 | mp = (struct lpfc_dmabuf *) pmb->context2; | ||
2418 | lpfc_mbuf_free(phba, mp->virt, mp->phys); | ||
2419 | kfree(mp); | ||
2420 | } | ||
2219 | mempool_free(pmb, phba->mbox_mem_pool); | 2421 | mempool_free(pmb, phba->mbox_mem_pool); |
2422 | } | ||
2220 | 2423 | ||
2221 | return; | 2424 | return; |
2222 | } | 2425 | } |
@@ -4360,7 +4563,7 @@ lpfc_read_fcoe_param(struct lpfc_hba *phba, | |||
4360 | fcoe_param_hdr = (struct lpfc_fip_param_hdr *) | 4563 | fcoe_param_hdr = (struct lpfc_fip_param_hdr *) |
4361 | buff; | 4564 | buff; |
4362 | fcoe_param = (struct lpfc_fcoe_params *) | 4565 | fcoe_param = (struct lpfc_fcoe_params *) |
4363 | buff + sizeof(struct lpfc_fip_param_hdr); | 4566 | (buff + sizeof(struct lpfc_fip_param_hdr)); |
4364 | 4567 | ||
4365 | if ((fcoe_param_hdr->parm_version != FIPP_VERSION) || | 4568 | if ((fcoe_param_hdr->parm_version != FIPP_VERSION) || |
4366 | (fcoe_param_hdr->length != FCOE_PARAM_LENGTH)) | 4569 | (fcoe_param_hdr->length != FCOE_PARAM_LENGTH)) |
diff --git a/drivers/scsi/lpfc/lpfc_hw.h b/drivers/scsi/lpfc/lpfc_hw.h index 8a3a026667e4..ccb26724dc53 100644 --- a/drivers/scsi/lpfc/lpfc_hw.h +++ b/drivers/scsi/lpfc/lpfc_hw.h | |||
@@ -2496,8 +2496,8 @@ typedef struct { | |||
2496 | #define DMP_VPORT_REGION_SIZE 0x200 | 2496 | #define DMP_VPORT_REGION_SIZE 0x200 |
2497 | #define DMP_MBOX_OFFSET_WORD 0x5 | 2497 | #define DMP_MBOX_OFFSET_WORD 0x5 |
2498 | 2498 | ||
2499 | #define DMP_REGION_FCOEPARAM 0x17 /* fcoe param region */ | 2499 | #define DMP_REGION_23 0x17 /* fcoe param and port state region */ |
2500 | #define DMP_FCOEPARAM_RGN_SIZE 0x400 | 2500 | #define DMP_RGN23_SIZE 0x400 |
2501 | 2501 | ||
2502 | #define WAKE_UP_PARMS_REGION_ID 4 | 2502 | #define WAKE_UP_PARMS_REGION_ID 4 |
2503 | #define WAKE_UP_PARMS_WORD_SIZE 15 | 2503 | #define WAKE_UP_PARMS_WORD_SIZE 15 |
diff --git a/drivers/scsi/lpfc/lpfc_hw4.h b/drivers/scsi/lpfc/lpfc_hw4.h index 2995d128f07f..3689eee04535 100644 --- a/drivers/scsi/lpfc/lpfc_hw4.h +++ b/drivers/scsi/lpfc/lpfc_hw4.h | |||
@@ -52,6 +52,31 @@ struct dma_address { | |||
52 | uint32_t addr_hi; | 52 | uint32_t addr_hi; |
53 | }; | 53 | }; |
54 | 54 | ||
55 | #define LPFC_SLIREV_CONF_WORD 0x58 | ||
56 | struct lpfc_sli_intf { | ||
57 | uint32_t word0; | ||
58 | #define lpfc_sli_intf_iftype_MASK 0x00000007 | ||
59 | #define lpfc_sli_intf_iftype_SHIFT 0 | ||
60 | #define lpfc_sli_intf_iftype_WORD word0 | ||
61 | #define lpfc_sli_intf_rev_MASK 0x0000000f | ||
62 | #define lpfc_sli_intf_rev_SHIFT 4 | ||
63 | #define lpfc_sli_intf_rev_WORD word0 | ||
64 | #define LPFC_SLIREV_CONF_SLI4 4 | ||
65 | #define lpfc_sli_intf_family_MASK 0x000000ff | ||
66 | #define lpfc_sli_intf_family_SHIFT 8 | ||
67 | #define lpfc_sli_intf_family_WORD word0 | ||
68 | #define lpfc_sli_intf_feat1_MASK 0x000000ff | ||
69 | #define lpfc_sli_intf_feat1_SHIFT 16 | ||
70 | #define lpfc_sli_intf_feat1_WORD word0 | ||
71 | #define lpfc_sli_intf_feat2_MASK 0x0000001f | ||
72 | #define lpfc_sli_intf_feat2_SHIFT 24 | ||
73 | #define lpfc_sli_intf_feat2_WORD word0 | ||
74 | #define lpfc_sli_intf_valid_MASK 0x00000007 | ||
75 | #define lpfc_sli_intf_valid_SHIFT 29 | ||
76 | #define lpfc_sli_intf_valid_WORD word0 | ||
77 | #define LPFC_SLI_INTF_VALID 6 | ||
78 | }; | ||
79 | |||
55 | #define LPFC_SLI4_BAR0 1 | 80 | #define LPFC_SLI4_BAR0 1 |
56 | #define LPFC_SLI4_BAR1 2 | 81 | #define LPFC_SLI4_BAR1 2 |
57 | #define LPFC_SLI4_BAR2 4 | 82 | #define LPFC_SLI4_BAR2 4 |
@@ -1181,6 +1206,32 @@ struct fcf_record { | |||
1181 | #define lpfc_fcf_record_fcf_state_MASK 0x0000FFFF | 1206 | #define lpfc_fcf_record_fcf_state_MASK 0x0000FFFF |
1182 | #define lpfc_fcf_record_fcf_state_WORD word8 | 1207 | #define lpfc_fcf_record_fcf_state_WORD word8 |
1183 | uint8_t vlan_bitmap[512]; | 1208 | uint8_t vlan_bitmap[512]; |
1209 | uint32_t word137; | ||
1210 | #define lpfc_fcf_record_switch_name_0_SHIFT 0 | ||
1211 | #define lpfc_fcf_record_switch_name_0_MASK 0x000000FF | ||
1212 | #define lpfc_fcf_record_switch_name_0_WORD word137 | ||
1213 | #define lpfc_fcf_record_switch_name_1_SHIFT 8 | ||
1214 | #define lpfc_fcf_record_switch_name_1_MASK 0x000000FF | ||
1215 | #define lpfc_fcf_record_switch_name_1_WORD word137 | ||
1216 | #define lpfc_fcf_record_switch_name_2_SHIFT 16 | ||
1217 | #define lpfc_fcf_record_switch_name_2_MASK 0x000000FF | ||
1218 | #define lpfc_fcf_record_switch_name_2_WORD word137 | ||
1219 | #define lpfc_fcf_record_switch_name_3_SHIFT 24 | ||
1220 | #define lpfc_fcf_record_switch_name_3_MASK 0x000000FF | ||
1221 | #define lpfc_fcf_record_switch_name_3_WORD word137 | ||
1222 | uint32_t word138; | ||
1223 | #define lpfc_fcf_record_switch_name_4_SHIFT 0 | ||
1224 | #define lpfc_fcf_record_switch_name_4_MASK 0x000000FF | ||
1225 | #define lpfc_fcf_record_switch_name_4_WORD word138 | ||
1226 | #define lpfc_fcf_record_switch_name_5_SHIFT 8 | ||
1227 | #define lpfc_fcf_record_switch_name_5_MASK 0x000000FF | ||
1228 | #define lpfc_fcf_record_switch_name_5_WORD word138 | ||
1229 | #define lpfc_fcf_record_switch_name_6_SHIFT 16 | ||
1230 | #define lpfc_fcf_record_switch_name_6_MASK 0x000000FF | ||
1231 | #define lpfc_fcf_record_switch_name_6_WORD word138 | ||
1232 | #define lpfc_fcf_record_switch_name_7_SHIFT 24 | ||
1233 | #define lpfc_fcf_record_switch_name_7_MASK 0x000000FF | ||
1234 | #define lpfc_fcf_record_switch_name_7_WORD word138 | ||
1184 | }; | 1235 | }; |
1185 | 1236 | ||
1186 | struct lpfc_mbx_read_fcf_tbl { | 1237 | struct lpfc_mbx_read_fcf_tbl { |
@@ -1385,20 +1436,17 @@ struct lpfc_mbx_unreg_vfi { | |||
1385 | 1436 | ||
1386 | struct lpfc_mbx_resume_rpi { | 1437 | struct lpfc_mbx_resume_rpi { |
1387 | uint32_t word1; | 1438 | uint32_t word1; |
1388 | #define lpfc_resume_rpi_rpi_SHIFT 0 | 1439 | #define lpfc_resume_rpi_index_SHIFT 0 |
1389 | #define lpfc_resume_rpi_rpi_MASK 0x0000FFFF | 1440 | #define lpfc_resume_rpi_index_MASK 0x0000FFFF |
1390 | #define lpfc_resume_rpi_rpi_WORD word1 | 1441 | #define lpfc_resume_rpi_index_WORD word1 |
1442 | #define lpfc_resume_rpi_ii_SHIFT 30 | ||
1443 | #define lpfc_resume_rpi_ii_MASK 0x00000003 | ||
1444 | #define lpfc_resume_rpi_ii_WORD word1 | ||
1445 | #define RESUME_INDEX_RPI 0 | ||
1446 | #define RESUME_INDEX_VPI 1 | ||
1447 | #define RESUME_INDEX_VFI 2 | ||
1448 | #define RESUME_INDEX_FCFI 3 | ||
1391 | uint32_t event_tag; | 1449 | uint32_t event_tag; |
1392 | uint32_t word3_rsvd; | ||
1393 | uint32_t word4_rsvd; | ||
1394 | uint32_t word5_rsvd; | ||
1395 | uint32_t word6; | ||
1396 | #define lpfc_resume_rpi_vpi_SHIFT 0 | ||
1397 | #define lpfc_resume_rpi_vpi_MASK 0x0000FFFF | ||
1398 | #define lpfc_resume_rpi_vpi_WORD word6 | ||
1399 | #define lpfc_resume_rpi_vfi_SHIFT 16 | ||
1400 | #define lpfc_resume_rpi_vfi_MASK 0x0000FFFF | ||
1401 | #define lpfc_resume_rpi_vfi_WORD word6 | ||
1402 | }; | 1450 | }; |
1403 | 1451 | ||
1404 | #define REG_FCF_INVALID_QID 0xFFFF | 1452 | #define REG_FCF_INVALID_QID 0xFFFF |
diff --git a/drivers/scsi/lpfc/lpfc_init.c b/drivers/scsi/lpfc/lpfc_init.c index fc67cc65c63b..562d8cee874b 100644 --- a/drivers/scsi/lpfc/lpfc_init.c +++ b/drivers/scsi/lpfc/lpfc_init.c | |||
@@ -211,7 +211,7 @@ lpfc_config_port_prep(struct lpfc_hba *phba) | |||
211 | goto out_free_mbox; | 211 | goto out_free_mbox; |
212 | 212 | ||
213 | do { | 213 | do { |
214 | lpfc_dump_mem(phba, pmb, offset); | 214 | lpfc_dump_mem(phba, pmb, offset, DMP_REGION_VPD); |
215 | rc = lpfc_sli_issue_mbox(phba, pmb, MBX_POLL); | 215 | rc = lpfc_sli_issue_mbox(phba, pmb, MBX_POLL); |
216 | 216 | ||
217 | if (rc != MBX_SUCCESS) { | 217 | if (rc != MBX_SUCCESS) { |
@@ -425,6 +425,9 @@ lpfc_config_port_post(struct lpfc_hba *phba) | |||
425 | return -EIO; | 425 | return -EIO; |
426 | } | 426 | } |
427 | 427 | ||
428 | /* Check if the port is disabled */ | ||
429 | lpfc_sli_read_link_ste(phba); | ||
430 | |||
428 | /* Reset the DFT_HBA_Q_DEPTH to the max xri */ | 431 | /* Reset the DFT_HBA_Q_DEPTH to the max xri */ |
429 | if (phba->cfg_hba_queue_depth > (mb->un.varRdConfig.max_xri+1)) | 432 | if (phba->cfg_hba_queue_depth > (mb->un.varRdConfig.max_xri+1)) |
430 | phba->cfg_hba_queue_depth = | 433 | phba->cfg_hba_queue_depth = |
@@ -524,27 +527,46 @@ lpfc_config_port_post(struct lpfc_hba *phba) | |||
524 | /* Set up error attention (ERATT) polling timer */ | 527 | /* Set up error attention (ERATT) polling timer */ |
525 | mod_timer(&phba->eratt_poll, jiffies + HZ * LPFC_ERATT_POLL_INTERVAL); | 528 | mod_timer(&phba->eratt_poll, jiffies + HZ * LPFC_ERATT_POLL_INTERVAL); |
526 | 529 | ||
527 | lpfc_init_link(phba, pmb, phba->cfg_topology, phba->cfg_link_speed); | 530 | if (phba->hba_flag & LINK_DISABLED) { |
528 | pmb->mbox_cmpl = lpfc_sli_def_mbox_cmpl; | 531 | lpfc_printf_log(phba, |
529 | lpfc_set_loopback_flag(phba); | 532 | KERN_ERR, LOG_INIT, |
530 | rc = lpfc_sli_issue_mbox(phba, pmb, MBX_NOWAIT); | 533 | "2598 Adapter Link is disabled.\n"); |
531 | if (rc != MBX_SUCCESS) { | 534 | lpfc_down_link(phba, pmb); |
532 | lpfc_printf_log(phba, KERN_ERR, LOG_INIT, | 535 | pmb->mbox_cmpl = lpfc_sli_def_mbox_cmpl; |
536 | rc = lpfc_sli_issue_mbox(phba, pmb, MBX_NOWAIT); | ||
537 | if ((rc != MBX_SUCCESS) && (rc != MBX_BUSY)) { | ||
538 | lpfc_printf_log(phba, | ||
539 | KERN_ERR, LOG_INIT, | ||
540 | "2599 Adapter failed to issue DOWN_LINK" | ||
541 | " mbox command rc 0x%x\n", rc); | ||
542 | |||
543 | mempool_free(pmb, phba->mbox_mem_pool); | ||
544 | return -EIO; | ||
545 | } | ||
546 | } else { | ||
547 | lpfc_init_link(phba, pmb, phba->cfg_topology, | ||
548 | phba->cfg_link_speed); | ||
549 | pmb->mbox_cmpl = lpfc_sli_def_mbox_cmpl; | ||
550 | lpfc_set_loopback_flag(phba); | ||
551 | rc = lpfc_sli_issue_mbox(phba, pmb, MBX_NOWAIT); | ||
552 | if (rc != MBX_SUCCESS) { | ||
553 | lpfc_printf_log(phba, KERN_ERR, LOG_INIT, | ||
533 | "0454 Adapter failed to init, mbxCmd x%x " | 554 | "0454 Adapter failed to init, mbxCmd x%x " |
534 | "INIT_LINK, mbxStatus x%x\n", | 555 | "INIT_LINK, mbxStatus x%x\n", |
535 | mb->mbxCommand, mb->mbxStatus); | 556 | mb->mbxCommand, mb->mbxStatus); |
536 | 557 | ||
537 | /* Clear all interrupt enable conditions */ | 558 | /* Clear all interrupt enable conditions */ |
538 | writel(0, phba->HCregaddr); | 559 | writel(0, phba->HCregaddr); |
539 | readl(phba->HCregaddr); /* flush */ | 560 | readl(phba->HCregaddr); /* flush */ |
540 | /* Clear all pending interrupts */ | 561 | /* Clear all pending interrupts */ |
541 | writel(0xffffffff, phba->HAregaddr); | 562 | writel(0xffffffff, phba->HAregaddr); |
542 | readl(phba->HAregaddr); /* flush */ | 563 | readl(phba->HAregaddr); /* flush */ |
543 | 564 | ||
544 | phba->link_state = LPFC_HBA_ERROR; | 565 | phba->link_state = LPFC_HBA_ERROR; |
545 | if (rc != MBX_BUSY) | 566 | if (rc != MBX_BUSY) |
546 | mempool_free(pmb, phba->mbox_mem_pool); | 567 | mempool_free(pmb, phba->mbox_mem_pool); |
547 | return -EIO; | 568 | return -EIO; |
569 | } | ||
548 | } | 570 | } |
549 | /* MBOX buffer will be freed in mbox compl */ | 571 | /* MBOX buffer will be freed in mbox compl */ |
550 | pmb = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); | 572 | pmb = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); |
@@ -558,7 +580,7 @@ lpfc_config_port_post(struct lpfc_hba *phba) | |||
558 | KERN_ERR, | 580 | KERN_ERR, |
559 | LOG_INIT, | 581 | LOG_INIT, |
560 | "0456 Adapter failed to issue " | 582 | "0456 Adapter failed to issue " |
561 | "ASYNCEVT_ENABLE mbox status x%x \n.", | 583 | "ASYNCEVT_ENABLE mbox status x%x\n", |
562 | rc); | 584 | rc); |
563 | mempool_free(pmb, phba->mbox_mem_pool); | 585 | mempool_free(pmb, phba->mbox_mem_pool); |
564 | } | 586 | } |
@@ -572,7 +594,7 @@ lpfc_config_port_post(struct lpfc_hba *phba) | |||
572 | 594 | ||
573 | if ((rc != MBX_BUSY) && (rc != MBX_SUCCESS)) { | 595 | if ((rc != MBX_BUSY) && (rc != MBX_SUCCESS)) { |
574 | lpfc_printf_log(phba, KERN_ERR, LOG_INIT, "0435 Adapter failed " | 596 | lpfc_printf_log(phba, KERN_ERR, LOG_INIT, "0435 Adapter failed " |
575 | "to get Option ROM version status x%x\n.", rc); | 597 | "to get Option ROM version status x%x\n", rc); |
576 | mempool_free(pmb, phba->mbox_mem_pool); | 598 | mempool_free(pmb, phba->mbox_mem_pool); |
577 | } | 599 | } |
578 | 600 | ||
@@ -2133,6 +2155,8 @@ lpfc_online(struct lpfc_hba *phba) | |||
2133 | vports[i]->fc_flag &= ~FC_OFFLINE_MODE; | 2155 | vports[i]->fc_flag &= ~FC_OFFLINE_MODE; |
2134 | if (phba->sli3_options & LPFC_SLI3_NPIV_ENABLED) | 2156 | if (phba->sli3_options & LPFC_SLI3_NPIV_ENABLED) |
2135 | vports[i]->fc_flag |= FC_VPORT_NEEDS_REG_VPI; | 2157 | vports[i]->fc_flag |= FC_VPORT_NEEDS_REG_VPI; |
2158 | if (phba->sli_rev == LPFC_SLI_REV4) | ||
2159 | vports[i]->fc_flag |= FC_VPORT_NEEDS_INIT_VPI; | ||
2136 | spin_unlock_irq(shost->host_lock); | 2160 | spin_unlock_irq(shost->host_lock); |
2137 | } | 2161 | } |
2138 | lpfc_destroy_vport_work_array(phba, vports); | 2162 | lpfc_destroy_vport_work_array(phba, vports); |
@@ -2807,6 +2831,7 @@ lpfc_sli4_async_link_evt(struct lpfc_hba *phba, | |||
2807 | att_type = lpfc_sli4_parse_latt_type(phba, acqe_link); | 2831 | att_type = lpfc_sli4_parse_latt_type(phba, acqe_link); |
2808 | if (att_type != AT_LINK_DOWN && att_type != AT_LINK_UP) | 2832 | if (att_type != AT_LINK_DOWN && att_type != AT_LINK_UP) |
2809 | return; | 2833 | return; |
2834 | phba->fcoe_eventtag = acqe_link->event_tag; | ||
2810 | pmb = (LPFC_MBOXQ_t *)mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); | 2835 | pmb = (LPFC_MBOXQ_t *)mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); |
2811 | if (!pmb) { | 2836 | if (!pmb) { |
2812 | lpfc_printf_log(phba, KERN_ERR, LOG_SLI, | 2837 | lpfc_printf_log(phba, KERN_ERR, LOG_SLI, |
@@ -2894,18 +2919,20 @@ lpfc_sli4_async_fcoe_evt(struct lpfc_hba *phba, | |||
2894 | uint8_t event_type = bf_get(lpfc_acqe_fcoe_event_type, acqe_fcoe); | 2919 | uint8_t event_type = bf_get(lpfc_acqe_fcoe_event_type, acqe_fcoe); |
2895 | int rc; | 2920 | int rc; |
2896 | 2921 | ||
2922 | phba->fcoe_eventtag = acqe_fcoe->event_tag; | ||
2897 | switch (event_type) { | 2923 | switch (event_type) { |
2898 | case LPFC_FCOE_EVENT_TYPE_NEW_FCF: | 2924 | case LPFC_FCOE_EVENT_TYPE_NEW_FCF: |
2899 | lpfc_printf_log(phba, KERN_ERR, LOG_DISCOVERY, | 2925 | lpfc_printf_log(phba, KERN_ERR, LOG_DISCOVERY, |
2900 | "2546 New FCF found index 0x%x tag 0x%x \n", | 2926 | "2546 New FCF found index 0x%x tag 0x%x\n", |
2901 | acqe_fcoe->fcf_index, | 2927 | acqe_fcoe->fcf_index, |
2902 | acqe_fcoe->event_tag); | 2928 | acqe_fcoe->event_tag); |
2903 | /* | 2929 | /* |
2904 | * If the current FCF is in discovered state, | 2930 | * If the current FCF is in discovered state, or |
2905 | * do nothing. | 2931 | * FCF discovery is in progress do nothing. |
2906 | */ | 2932 | */ |
2907 | spin_lock_irq(&phba->hbalock); | 2933 | spin_lock_irq(&phba->hbalock); |
2908 | if (phba->fcf.fcf_flag & FCF_DISCOVERED) { | 2934 | if ((phba->fcf.fcf_flag & FCF_DISCOVERED) || |
2935 | (phba->hba_flag & FCF_DISC_INPROGRESS)) { | ||
2909 | spin_unlock_irq(&phba->hbalock); | 2936 | spin_unlock_irq(&phba->hbalock); |
2910 | break; | 2937 | break; |
2911 | } | 2938 | } |
@@ -2922,7 +2949,7 @@ lpfc_sli4_async_fcoe_evt(struct lpfc_hba *phba, | |||
2922 | 2949 | ||
2923 | case LPFC_FCOE_EVENT_TYPE_FCF_TABLE_FULL: | 2950 | case LPFC_FCOE_EVENT_TYPE_FCF_TABLE_FULL: |
2924 | lpfc_printf_log(phba, KERN_ERR, LOG_SLI, | 2951 | lpfc_printf_log(phba, KERN_ERR, LOG_SLI, |
2925 | "2548 FCF Table full count 0x%x tag 0x%x \n", | 2952 | "2548 FCF Table full count 0x%x tag 0x%x\n", |
2926 | bf_get(lpfc_acqe_fcoe_fcf_count, acqe_fcoe), | 2953 | bf_get(lpfc_acqe_fcoe_fcf_count, acqe_fcoe), |
2927 | acqe_fcoe->event_tag); | 2954 | acqe_fcoe->event_tag); |
2928 | break; | 2955 | break; |
@@ -2930,7 +2957,7 @@ lpfc_sli4_async_fcoe_evt(struct lpfc_hba *phba, | |||
2930 | case LPFC_FCOE_EVENT_TYPE_FCF_DEAD: | 2957 | case LPFC_FCOE_EVENT_TYPE_FCF_DEAD: |
2931 | lpfc_printf_log(phba, KERN_ERR, LOG_DISCOVERY, | 2958 | lpfc_printf_log(phba, KERN_ERR, LOG_DISCOVERY, |
2932 | "2549 FCF disconnected fron network index 0x%x" | 2959 | "2549 FCF disconnected fron network index 0x%x" |
2933 | " tag 0x%x \n", acqe_fcoe->fcf_index, | 2960 | " tag 0x%x\n", acqe_fcoe->fcf_index, |
2934 | acqe_fcoe->event_tag); | 2961 | acqe_fcoe->event_tag); |
2935 | /* If the event is not for currently used fcf do nothing */ | 2962 | /* If the event is not for currently used fcf do nothing */ |
2936 | if (phba->fcf.fcf_indx != acqe_fcoe->fcf_index) | 2963 | if (phba->fcf.fcf_indx != acqe_fcoe->fcf_index) |
@@ -4130,8 +4157,7 @@ lpfc_hba_alloc(struct pci_dev *pdev) | |||
4130 | /* Allocate memory for HBA structure */ | 4157 | /* Allocate memory for HBA structure */ |
4131 | phba = kzalloc(sizeof(struct lpfc_hba), GFP_KERNEL); | 4158 | phba = kzalloc(sizeof(struct lpfc_hba), GFP_KERNEL); |
4132 | if (!phba) { | 4159 | if (!phba) { |
4133 | lpfc_printf_log(phba, KERN_ERR, LOG_INIT, | 4160 | dev_err(&pdev->dev, "failed to allocate hba struct\n"); |
4134 | "1417 Failed to allocate hba struct.\n"); | ||
4135 | return NULL; | 4161 | return NULL; |
4136 | } | 4162 | } |
4137 | 4163 | ||
@@ -4145,6 +4171,9 @@ lpfc_hba_alloc(struct pci_dev *pdev) | |||
4145 | return NULL; | 4171 | return NULL; |
4146 | } | 4172 | } |
4147 | 4173 | ||
4174 | mutex_init(&phba->ct_event_mutex); | ||
4175 | INIT_LIST_HEAD(&phba->ct_ev_waiters); | ||
4176 | |||
4148 | return phba; | 4177 | return phba; |
4149 | } | 4178 | } |
4150 | 4179 | ||
@@ -4489,23 +4518,6 @@ lpfc_sli4_post_status_check(struct lpfc_hba *phba) | |||
4489 | if (!phba->sli4_hba.STAregaddr) | 4518 | if (!phba->sli4_hba.STAregaddr) |
4490 | return -ENODEV; | 4519 | return -ENODEV; |
4491 | 4520 | ||
4492 | /* With uncoverable error, log the error message and return error */ | ||
4493 | onlnreg0 = readl(phba->sli4_hba.ONLINE0regaddr); | ||
4494 | onlnreg1 = readl(phba->sli4_hba.ONLINE1regaddr); | ||
4495 | if ((onlnreg0 != LPFC_ONLINE_NERR) || (onlnreg1 != LPFC_ONLINE_NERR)) { | ||
4496 | uerrlo_reg.word0 = readl(phba->sli4_hba.UERRLOregaddr); | ||
4497 | uerrhi_reg.word0 = readl(phba->sli4_hba.UERRHIregaddr); | ||
4498 | if (uerrlo_reg.word0 || uerrhi_reg.word0) { | ||
4499 | lpfc_printf_log(phba, KERN_ERR, LOG_INIT, | ||
4500 | "1422 HBA Unrecoverable error: " | ||
4501 | "uerr_lo_reg=0x%x, uerr_hi_reg=0x%x, " | ||
4502 | "online0_reg=0x%x, online1_reg=0x%x\n", | ||
4503 | uerrlo_reg.word0, uerrhi_reg.word0, | ||
4504 | onlnreg0, onlnreg1); | ||
4505 | } | ||
4506 | return -ENODEV; | ||
4507 | } | ||
4508 | |||
4509 | /* Wait up to 30 seconds for the SLI Port POST done and ready */ | 4521 | /* Wait up to 30 seconds for the SLI Port POST done and ready */ |
4510 | for (i = 0; i < 3000; i++) { | 4522 | for (i = 0; i < 3000; i++) { |
4511 | sta_reg.word0 = readl(phba->sli4_hba.STAregaddr); | 4523 | sta_reg.word0 = readl(phba->sli4_hba.STAregaddr); |
@@ -4545,6 +4557,23 @@ lpfc_sli4_post_status_check(struct lpfc_hba *phba) | |||
4545 | bf_get(lpfc_scratchpad_featurelevel1, &scratchpad), | 4557 | bf_get(lpfc_scratchpad_featurelevel1, &scratchpad), |
4546 | bf_get(lpfc_scratchpad_featurelevel2, &scratchpad)); | 4558 | bf_get(lpfc_scratchpad_featurelevel2, &scratchpad)); |
4547 | 4559 | ||
4560 | /* With uncoverable error, log the error message and return error */ | ||
4561 | onlnreg0 = readl(phba->sli4_hba.ONLINE0regaddr); | ||
4562 | onlnreg1 = readl(phba->sli4_hba.ONLINE1regaddr); | ||
4563 | if ((onlnreg0 != LPFC_ONLINE_NERR) || (onlnreg1 != LPFC_ONLINE_NERR)) { | ||
4564 | uerrlo_reg.word0 = readl(phba->sli4_hba.UERRLOregaddr); | ||
4565 | uerrhi_reg.word0 = readl(phba->sli4_hba.UERRHIregaddr); | ||
4566 | if (uerrlo_reg.word0 || uerrhi_reg.word0) { | ||
4567 | lpfc_printf_log(phba, KERN_ERR, LOG_INIT, | ||
4568 | "1422 HBA Unrecoverable error: " | ||
4569 | "uerr_lo_reg=0x%x, uerr_hi_reg=0x%x, " | ||
4570 | "online0_reg=0x%x, online1_reg=0x%x\n", | ||
4571 | uerrlo_reg.word0, uerrhi_reg.word0, | ||
4572 | onlnreg0, onlnreg1); | ||
4573 | } | ||
4574 | return -ENODEV; | ||
4575 | } | ||
4576 | |||
4548 | return port_error; | 4577 | return port_error; |
4549 | } | 4578 | } |
4550 | 4579 | ||
@@ -7347,6 +7376,9 @@ lpfc_pci_probe_one_s4(struct pci_dev *pdev, const struct pci_device_id *pid) | |||
7347 | /* Perform post initialization setup */ | 7376 | /* Perform post initialization setup */ |
7348 | lpfc_post_init_setup(phba); | 7377 | lpfc_post_init_setup(phba); |
7349 | 7378 | ||
7379 | /* Check if there are static vports to be created. */ | ||
7380 | lpfc_create_static_vport(phba); | ||
7381 | |||
7350 | return 0; | 7382 | return 0; |
7351 | 7383 | ||
7352 | out_disable_intr: | 7384 | out_disable_intr: |
@@ -7636,19 +7668,17 @@ static int __devinit | |||
7636 | lpfc_pci_probe_one(struct pci_dev *pdev, const struct pci_device_id *pid) | 7668 | lpfc_pci_probe_one(struct pci_dev *pdev, const struct pci_device_id *pid) |
7637 | { | 7669 | { |
7638 | int rc; | 7670 | int rc; |
7639 | uint16_t dev_id; | 7671 | struct lpfc_sli_intf intf; |
7640 | 7672 | ||
7641 | if (pci_read_config_word(pdev, PCI_DEVICE_ID, &dev_id)) | 7673 | if (pci_read_config_dword(pdev, LPFC_SLIREV_CONF_WORD, &intf.word0)) |
7642 | return -ENODEV; | 7674 | return -ENODEV; |
7643 | 7675 | ||
7644 | switch (dev_id) { | 7676 | if ((bf_get(lpfc_sli_intf_valid, &intf) == LPFC_SLI_INTF_VALID) && |
7645 | case PCI_DEVICE_ID_TIGERSHARK: | 7677 | (bf_get(lpfc_sli_intf_rev, &intf) == LPFC_SLIREV_CONF_SLI4)) |
7646 | rc = lpfc_pci_probe_one_s4(pdev, pid); | 7678 | rc = lpfc_pci_probe_one_s4(pdev, pid); |
7647 | break; | 7679 | else |
7648 | default: | ||
7649 | rc = lpfc_pci_probe_one_s3(pdev, pid); | 7680 | rc = lpfc_pci_probe_one_s3(pdev, pid); |
7650 | break; | 7681 | |
7651 | } | ||
7652 | return rc; | 7682 | return rc; |
7653 | } | 7683 | } |
7654 | 7684 | ||
diff --git a/drivers/scsi/lpfc/lpfc_mbox.c b/drivers/scsi/lpfc/lpfc_mbox.c index 3423571dd1b3..1ab405902a18 100644 --- a/drivers/scsi/lpfc/lpfc_mbox.c +++ b/drivers/scsi/lpfc/lpfc_mbox.c | |||
@@ -52,48 +52,85 @@ | |||
52 | * This routine prepares the mailbox command for dumping list of static | 52 | * This routine prepares the mailbox command for dumping list of static |
53 | * vports to be created. | 53 | * vports to be created. |
54 | **/ | 54 | **/ |
55 | void | 55 | int |
56 | lpfc_dump_static_vport(struct lpfc_hba *phba, LPFC_MBOXQ_t *pmb, | 56 | lpfc_dump_static_vport(struct lpfc_hba *phba, LPFC_MBOXQ_t *pmb, |
57 | uint16_t offset) | 57 | uint16_t offset) |
58 | { | 58 | { |
59 | MAILBOX_t *mb; | 59 | MAILBOX_t *mb; |
60 | void *ctx; | 60 | struct lpfc_dmabuf *mp; |
61 | 61 | ||
62 | mb = &pmb->u.mb; | 62 | mb = &pmb->u.mb; |
63 | ctx = pmb->context2; | ||
64 | 63 | ||
65 | /* Setup to dump vport info region */ | 64 | /* Setup to dump vport info region */ |
66 | memset(pmb, 0, sizeof(LPFC_MBOXQ_t)); | 65 | memset(pmb, 0, sizeof(LPFC_MBOXQ_t)); |
67 | mb->mbxCommand = MBX_DUMP_MEMORY; | 66 | mb->mbxCommand = MBX_DUMP_MEMORY; |
68 | mb->un.varDmp.cv = 1; | ||
69 | mb->un.varDmp.type = DMP_NV_PARAMS; | 67 | mb->un.varDmp.type = DMP_NV_PARAMS; |
70 | mb->un.varDmp.entry_index = offset; | 68 | mb->un.varDmp.entry_index = offset; |
71 | mb->un.varDmp.region_id = DMP_REGION_VPORT; | 69 | mb->un.varDmp.region_id = DMP_REGION_VPORT; |
72 | mb->un.varDmp.word_cnt = DMP_RSP_SIZE/sizeof(uint32_t); | ||
73 | mb->un.varDmp.co = 0; | ||
74 | mb->un.varDmp.resp_offset = 0; | ||
75 | pmb->context2 = ctx; | ||
76 | mb->mbxOwner = OWN_HOST; | 70 | mb->mbxOwner = OWN_HOST; |
77 | 71 | ||
78 | return; | 72 | /* For SLI3 HBAs data is embedded in mailbox */ |
73 | if (phba->sli_rev != LPFC_SLI_REV4) { | ||
74 | mb->un.varDmp.cv = 1; | ||
75 | mb->un.varDmp.word_cnt = DMP_RSP_SIZE/sizeof(uint32_t); | ||
76 | return 0; | ||
77 | } | ||
78 | |||
79 | /* For SLI4 HBAs driver need to allocate memory */ | ||
80 | mp = kmalloc(sizeof(struct lpfc_dmabuf), GFP_KERNEL); | ||
81 | if (mp) | ||
82 | mp->virt = lpfc_mbuf_alloc(phba, 0, &mp->phys); | ||
83 | |||
84 | if (!mp || !mp->virt) { | ||
85 | kfree(mp); | ||
86 | lpfc_printf_log(phba, KERN_ERR, LOG_MBOX, | ||
87 | "2605 lpfc_dump_static_vport: memory" | ||
88 | " allocation failed\n"); | ||
89 | return 1; | ||
90 | } | ||
91 | memset(mp->virt, 0, LPFC_BPL_SIZE); | ||
92 | INIT_LIST_HEAD(&mp->list); | ||
93 | /* save address for completion */ | ||
94 | pmb->context2 = (uint8_t *) mp; | ||
95 | mb->un.varWords[3] = putPaddrLow(mp->phys); | ||
96 | mb->un.varWords[4] = putPaddrHigh(mp->phys); | ||
97 | mb->un.varDmp.sli4_length = sizeof(struct static_vport_info); | ||
98 | |||
99 | return 0; | ||
100 | } | ||
101 | |||
102 | /** | ||
103 | * lpfc_down_link - Bring down HBAs link. | ||
104 | * @phba: pointer to lpfc hba data structure. | ||
105 | * @pmb: pointer to the driver internal queue element for mailbox command. | ||
106 | * | ||
107 | * This routine prepares a mailbox command to bring down HBA link. | ||
108 | **/ | ||
109 | void | ||
110 | lpfc_down_link(struct lpfc_hba *phba, LPFC_MBOXQ_t *pmb) | ||
111 | { | ||
112 | MAILBOX_t *mb; | ||
113 | memset(pmb, 0, sizeof(LPFC_MBOXQ_t)); | ||
114 | mb = &pmb->u.mb; | ||
115 | mb->mbxCommand = MBX_DOWN_LINK; | ||
116 | mb->mbxOwner = OWN_HOST; | ||
79 | } | 117 | } |
80 | 118 | ||
81 | /** | 119 | /** |
82 | * lpfc_dump_mem - Prepare a mailbox command for retrieving HBA's VPD memory | 120 | * lpfc_dump_mem - Prepare a mailbox command for reading a region. |
83 | * @phba: pointer to lpfc hba data structure. | 121 | * @phba: pointer to lpfc hba data structure. |
84 | * @pmb: pointer to the driver internal queue element for mailbox command. | 122 | * @pmb: pointer to the driver internal queue element for mailbox command. |
85 | * @offset: offset for dumping VPD memory mailbox command. | 123 | * @offset: offset into the region. |
124 | * @region_id: config region id. | ||
86 | * | 125 | * |
87 | * The dump mailbox command provides a method for the device driver to obtain | 126 | * The dump mailbox command provides a method for the device driver to obtain |
88 | * various types of information from the HBA device. | 127 | * various types of information from the HBA device. |
89 | * | 128 | * |
90 | * This routine prepares the mailbox command for dumping HBA Vital Product | 129 | * This routine prepares the mailbox command for dumping HBA's config region. |
91 | * Data (VPD) memory. This mailbox command is to be used for retrieving a | ||
92 | * portion (DMP_RSP_SIZE bytes) of a HBA's VPD from the HBA at an address | ||
93 | * offset specified by the offset parameter. | ||
94 | **/ | 130 | **/ |
95 | void | 131 | void |
96 | lpfc_dump_mem(struct lpfc_hba * phba, LPFC_MBOXQ_t * pmb, uint16_t offset) | 132 | lpfc_dump_mem(struct lpfc_hba *phba, LPFC_MBOXQ_t *pmb, uint16_t offset, |
133 | uint16_t region_id) | ||
97 | { | 134 | { |
98 | MAILBOX_t *mb; | 135 | MAILBOX_t *mb; |
99 | void *ctx; | 136 | void *ctx; |
@@ -107,7 +144,7 @@ lpfc_dump_mem(struct lpfc_hba * phba, LPFC_MBOXQ_t * pmb, uint16_t offset) | |||
107 | mb->un.varDmp.cv = 1; | 144 | mb->un.varDmp.cv = 1; |
108 | mb->un.varDmp.type = DMP_NV_PARAMS; | 145 | mb->un.varDmp.type = DMP_NV_PARAMS; |
109 | mb->un.varDmp.entry_index = offset; | 146 | mb->un.varDmp.entry_index = offset; |
110 | mb->un.varDmp.region_id = DMP_REGION_VPD; | 147 | mb->un.varDmp.region_id = region_id; |
111 | mb->un.varDmp.word_cnt = (DMP_RSP_SIZE / sizeof (uint32_t)); | 148 | mb->un.varDmp.word_cnt = (DMP_RSP_SIZE / sizeof (uint32_t)); |
112 | mb->un.varDmp.co = 0; | 149 | mb->un.varDmp.co = 0; |
113 | mb->un.varDmp.resp_offset = 0; | 150 | mb->un.varDmp.resp_offset = 0; |
@@ -1789,6 +1826,7 @@ lpfc_reg_vfi(struct lpfcMboxq *mbox, struct lpfc_vport *vport, dma_addr_t phys) | |||
1789 | 1826 | ||
1790 | /** | 1827 | /** |
1791 | * lpfc_init_vpi - Initialize the INIT_VPI mailbox command | 1828 | * lpfc_init_vpi - Initialize the INIT_VPI mailbox command |
1829 | * @phba: pointer to the hba structure to init the VPI for. | ||
1792 | * @mbox: pointer to lpfc mbox command to initialize. | 1830 | * @mbox: pointer to lpfc mbox command to initialize. |
1793 | * @vpi: VPI to be initialized. | 1831 | * @vpi: VPI to be initialized. |
1794 | * | 1832 | * |
@@ -1799,11 +1837,14 @@ lpfc_reg_vfi(struct lpfcMboxq *mbox, struct lpfc_vport *vport, dma_addr_t phys) | |||
1799 | * successful virtual NPort login. | 1837 | * successful virtual NPort login. |
1800 | **/ | 1838 | **/ |
1801 | void | 1839 | void |
1802 | lpfc_init_vpi(struct lpfcMboxq *mbox, uint16_t vpi) | 1840 | lpfc_init_vpi(struct lpfc_hba *phba, struct lpfcMboxq *mbox, uint16_t vpi) |
1803 | { | 1841 | { |
1804 | memset(mbox, 0, sizeof(*mbox)); | 1842 | memset(mbox, 0, sizeof(*mbox)); |
1805 | bf_set(lpfc_mqe_command, &mbox->u.mqe, MBX_INIT_VPI); | 1843 | bf_set(lpfc_mqe_command, &mbox->u.mqe, MBX_INIT_VPI); |
1806 | bf_set(lpfc_init_vpi_vpi, &mbox->u.mqe.un.init_vpi, vpi); | 1844 | bf_set(lpfc_init_vpi_vpi, &mbox->u.mqe.un.init_vpi, |
1845 | vpi + phba->vpi_base); | ||
1846 | bf_set(lpfc_init_vpi_vfi, &mbox->u.mqe.un.init_vpi, | ||
1847 | phba->pport->vfi + phba->vfi_base); | ||
1807 | } | 1848 | } |
1808 | 1849 | ||
1809 | /** | 1850 | /** |
@@ -1852,7 +1893,7 @@ lpfc_dump_fcoe_param(struct lpfc_hba *phba, | |||
1852 | /* dump_fcoe_param failed to allocate memory */ | 1893 | /* dump_fcoe_param failed to allocate memory */ |
1853 | lpfc_printf_log(phba, KERN_WARNING, LOG_MBOX, | 1894 | lpfc_printf_log(phba, KERN_WARNING, LOG_MBOX, |
1854 | "2569 lpfc_dump_fcoe_param: memory" | 1895 | "2569 lpfc_dump_fcoe_param: memory" |
1855 | " allocation failed \n"); | 1896 | " allocation failed\n"); |
1856 | return 1; | 1897 | return 1; |
1857 | } | 1898 | } |
1858 | 1899 | ||
@@ -1864,8 +1905,8 @@ lpfc_dump_fcoe_param(struct lpfc_hba *phba, | |||
1864 | 1905 | ||
1865 | mb->mbxCommand = MBX_DUMP_MEMORY; | 1906 | mb->mbxCommand = MBX_DUMP_MEMORY; |
1866 | mb->un.varDmp.type = DMP_NV_PARAMS; | 1907 | mb->un.varDmp.type = DMP_NV_PARAMS; |
1867 | mb->un.varDmp.region_id = DMP_REGION_FCOEPARAM; | 1908 | mb->un.varDmp.region_id = DMP_REGION_23; |
1868 | mb->un.varDmp.sli4_length = DMP_FCOEPARAM_RGN_SIZE; | 1909 | mb->un.varDmp.sli4_length = DMP_RGN23_SIZE; |
1869 | mb->un.varWords[3] = putPaddrLow(mp->phys); | 1910 | mb->un.varWords[3] = putPaddrLow(mp->phys); |
1870 | mb->un.varWords[4] = putPaddrHigh(mp->phys); | 1911 | mb->un.varWords[4] = putPaddrHigh(mp->phys); |
1871 | return 0; | 1912 | return 0; |
@@ -1938,9 +1979,7 @@ lpfc_resume_rpi(struct lpfcMboxq *mbox, struct lpfc_nodelist *ndlp) | |||
1938 | memset(mbox, 0, sizeof(*mbox)); | 1979 | memset(mbox, 0, sizeof(*mbox)); |
1939 | resume_rpi = &mbox->u.mqe.un.resume_rpi; | 1980 | resume_rpi = &mbox->u.mqe.un.resume_rpi; |
1940 | bf_set(lpfc_mqe_command, &mbox->u.mqe, MBX_RESUME_RPI); | 1981 | bf_set(lpfc_mqe_command, &mbox->u.mqe, MBX_RESUME_RPI); |
1941 | bf_set(lpfc_resume_rpi_rpi, resume_rpi, ndlp->nlp_rpi); | 1982 | bf_set(lpfc_resume_rpi_index, resume_rpi, ndlp->nlp_rpi); |
1942 | bf_set(lpfc_resume_rpi_vpi, resume_rpi, | 1983 | bf_set(lpfc_resume_rpi_ii, resume_rpi, RESUME_INDEX_RPI); |
1943 | ndlp->vport->vpi + ndlp->vport->phba->vpi_base); | 1984 | resume_rpi->event_tag = ndlp->phba->fc_eventTag; |
1944 | bf_set(lpfc_resume_rpi_vfi, resume_rpi, | ||
1945 | ndlp->vport->vfi + ndlp->vport->phba->vfi_base); | ||
1946 | } | 1985 | } |
diff --git a/drivers/scsi/lpfc/lpfc_mem.c b/drivers/scsi/lpfc/lpfc_mem.c index e198c917c13e..a1b6db6016da 100644 --- a/drivers/scsi/lpfc/lpfc_mem.c +++ b/drivers/scsi/lpfc/lpfc_mem.c | |||
@@ -110,17 +110,28 @@ lpfc_mem_alloc(struct lpfc_hba *phba, int align) | |||
110 | sizeof(struct lpfc_nodelist)); | 110 | sizeof(struct lpfc_nodelist)); |
111 | if (!phba->nlp_mem_pool) | 111 | if (!phba->nlp_mem_pool) |
112 | goto fail_free_mbox_pool; | 112 | goto fail_free_mbox_pool; |
113 | phba->lpfc_hrb_pool = pci_pool_create("lpfc_hrb_pool", | 113 | |
114 | if (phba->sli_rev == LPFC_SLI_REV4) { | ||
115 | phba->lpfc_hrb_pool = pci_pool_create("lpfc_hrb_pool", | ||
114 | phba->pcidev, | 116 | phba->pcidev, |
115 | LPFC_HDR_BUF_SIZE, align, 0); | 117 | LPFC_HDR_BUF_SIZE, align, 0); |
116 | if (!phba->lpfc_hrb_pool) | 118 | if (!phba->lpfc_hrb_pool) |
117 | goto fail_free_nlp_mem_pool; | 119 | goto fail_free_nlp_mem_pool; |
118 | phba->lpfc_drb_pool = pci_pool_create("lpfc_drb_pool", | 120 | |
121 | phba->lpfc_drb_pool = pci_pool_create("lpfc_drb_pool", | ||
119 | phba->pcidev, | 122 | phba->pcidev, |
120 | LPFC_DATA_BUF_SIZE, align, 0); | 123 | LPFC_DATA_BUF_SIZE, align, 0); |
121 | if (!phba->lpfc_drb_pool) | 124 | if (!phba->lpfc_drb_pool) |
122 | goto fail_free_hbq_pool; | 125 | goto fail_free_hrb_pool; |
123 | 126 | phba->lpfc_hbq_pool = NULL; | |
127 | } else { | ||
128 | phba->lpfc_hbq_pool = pci_pool_create("lpfc_hbq_pool", | ||
129 | phba->pcidev, LPFC_BPL_SIZE, align, 0); | ||
130 | if (!phba->lpfc_hbq_pool) | ||
131 | goto fail_free_nlp_mem_pool; | ||
132 | phba->lpfc_hrb_pool = NULL; | ||
133 | phba->lpfc_drb_pool = NULL; | ||
134 | } | ||
124 | /* vpi zero is reserved for the physical port so add 1 to max */ | 135 | /* vpi zero is reserved for the physical port so add 1 to max */ |
125 | longs = ((phba->max_vpi + 1) + BITS_PER_LONG - 1) / BITS_PER_LONG; | 136 | longs = ((phba->max_vpi + 1) + BITS_PER_LONG - 1) / BITS_PER_LONG; |
126 | phba->vpi_bmask = kzalloc(longs * sizeof(unsigned long), GFP_KERNEL); | 137 | phba->vpi_bmask = kzalloc(longs * sizeof(unsigned long), GFP_KERNEL); |
@@ -132,7 +143,7 @@ lpfc_mem_alloc(struct lpfc_hba *phba, int align) | |||
132 | fail_free_dbq_pool: | 143 | fail_free_dbq_pool: |
133 | pci_pool_destroy(phba->lpfc_drb_pool); | 144 | pci_pool_destroy(phba->lpfc_drb_pool); |
134 | phba->lpfc_drb_pool = NULL; | 145 | phba->lpfc_drb_pool = NULL; |
135 | fail_free_hbq_pool: | 146 | fail_free_hrb_pool: |
136 | pci_pool_destroy(phba->lpfc_hrb_pool); | 147 | pci_pool_destroy(phba->lpfc_hrb_pool); |
137 | phba->lpfc_hrb_pool = NULL; | 148 | phba->lpfc_hrb_pool = NULL; |
138 | fail_free_nlp_mem_pool: | 149 | fail_free_nlp_mem_pool: |
@@ -176,11 +187,17 @@ lpfc_mem_free(struct lpfc_hba *phba) | |||
176 | 187 | ||
177 | /* Free HBQ pools */ | 188 | /* Free HBQ pools */ |
178 | lpfc_sli_hbqbuf_free_all(phba); | 189 | lpfc_sli_hbqbuf_free_all(phba); |
179 | pci_pool_destroy(phba->lpfc_drb_pool); | 190 | if (phba->lpfc_drb_pool) |
191 | pci_pool_destroy(phba->lpfc_drb_pool); | ||
180 | phba->lpfc_drb_pool = NULL; | 192 | phba->lpfc_drb_pool = NULL; |
181 | pci_pool_destroy(phba->lpfc_hrb_pool); | 193 | if (phba->lpfc_hrb_pool) |
194 | pci_pool_destroy(phba->lpfc_hrb_pool); | ||
182 | phba->lpfc_hrb_pool = NULL; | 195 | phba->lpfc_hrb_pool = NULL; |
183 | 196 | ||
197 | if (phba->lpfc_hbq_pool) | ||
198 | pci_pool_destroy(phba->lpfc_hbq_pool); | ||
199 | phba->lpfc_hbq_pool = NULL; | ||
200 | |||
184 | /* Free NLP memory pool */ | 201 | /* Free NLP memory pool */ |
185 | mempool_destroy(phba->nlp_mem_pool); | 202 | mempool_destroy(phba->nlp_mem_pool); |
186 | phba->nlp_mem_pool = NULL; | 203 | phba->nlp_mem_pool = NULL; |
@@ -380,7 +397,7 @@ lpfc_els_hbq_alloc(struct lpfc_hba *phba) | |||
380 | if (!hbqbp) | 397 | if (!hbqbp) |
381 | return NULL; | 398 | return NULL; |
382 | 399 | ||
383 | hbqbp->dbuf.virt = pci_pool_alloc(phba->lpfc_hrb_pool, GFP_KERNEL, | 400 | hbqbp->dbuf.virt = pci_pool_alloc(phba->lpfc_hbq_pool, GFP_KERNEL, |
384 | &hbqbp->dbuf.phys); | 401 | &hbqbp->dbuf.phys); |
385 | if (!hbqbp->dbuf.virt) { | 402 | if (!hbqbp->dbuf.virt) { |
386 | kfree(hbqbp); | 403 | kfree(hbqbp); |
@@ -405,7 +422,7 @@ lpfc_els_hbq_alloc(struct lpfc_hba *phba) | |||
405 | void | 422 | void |
406 | lpfc_els_hbq_free(struct lpfc_hba *phba, struct hbq_dmabuf *hbqbp) | 423 | lpfc_els_hbq_free(struct lpfc_hba *phba, struct hbq_dmabuf *hbqbp) |
407 | { | 424 | { |
408 | pci_pool_free(phba->lpfc_hrb_pool, hbqbp->dbuf.virt, hbqbp->dbuf.phys); | 425 | pci_pool_free(phba->lpfc_hbq_pool, hbqbp->dbuf.virt, hbqbp->dbuf.phys); |
409 | kfree(hbqbp); | 426 | kfree(hbqbp); |
410 | return; | 427 | return; |
411 | } | 428 | } |
diff --git a/drivers/scsi/lpfc/lpfc_nl.h b/drivers/scsi/lpfc/lpfc_nl.h index 27d1a88a98fe..d655ed3eebef 100644 --- a/drivers/scsi/lpfc/lpfc_nl.h +++ b/drivers/scsi/lpfc/lpfc_nl.h | |||
@@ -177,3 +177,23 @@ struct temp_event { | |||
177 | uint32_t data; | 177 | uint32_t data; |
178 | }; | 178 | }; |
179 | 179 | ||
180 | /* bsg definitions */ | ||
181 | #define LPFC_BSG_VENDOR_SET_CT_EVENT 1 | ||
182 | #define LPFC_BSG_VENDOR_GET_CT_EVENT 2 | ||
183 | |||
184 | struct set_ct_event { | ||
185 | uint32_t command; | ||
186 | uint32_t ev_req_id; | ||
187 | uint32_t ev_reg_id; | ||
188 | }; | ||
189 | |||
190 | struct get_ct_event { | ||
191 | uint32_t command; | ||
192 | uint32_t ev_reg_id; | ||
193 | uint32_t ev_req_id; | ||
194 | }; | ||
195 | |||
196 | struct get_ct_event_reply { | ||
197 | uint32_t immed_data; | ||
198 | uint32_t type; | ||
199 | }; | ||
diff --git a/drivers/scsi/lpfc/lpfc_scsi.c b/drivers/scsi/lpfc/lpfc_scsi.c index da59c4f0168f..61d089703806 100644 --- a/drivers/scsi/lpfc/lpfc_scsi.c +++ b/drivers/scsi/lpfc/lpfc_scsi.c | |||
@@ -2142,7 +2142,7 @@ lpfc_handle_fcp_err(struct lpfc_vport *vport, struct lpfc_scsi_buf *lpfc_cmd, | |||
2142 | } else if (resp_info & RESID_OVER) { | 2142 | } else if (resp_info & RESID_OVER) { |
2143 | lpfc_printf_vlog(vport, KERN_WARNING, LOG_FCP, | 2143 | lpfc_printf_vlog(vport, KERN_WARNING, LOG_FCP, |
2144 | "9028 FCP command x%x residual overrun error. " | 2144 | "9028 FCP command x%x residual overrun error. " |
2145 | "Data: x%x x%x \n", cmnd->cmnd[0], | 2145 | "Data: x%x x%x\n", cmnd->cmnd[0], |
2146 | scsi_bufflen(cmnd), scsi_get_resid(cmnd)); | 2146 | scsi_bufflen(cmnd), scsi_get_resid(cmnd)); |
2147 | host_status = DID_ERROR; | 2147 | host_status = DID_ERROR; |
2148 | 2148 | ||
@@ -2843,7 +2843,7 @@ lpfc_queuecommand(struct scsi_cmnd *cmnd, void (*done) (struct scsi_cmnd *)) | |||
2843 | dif_op_str[scsi_get_prot_op(cmnd)]); | 2843 | dif_op_str[scsi_get_prot_op(cmnd)]); |
2844 | lpfc_printf_vlog(vport, KERN_WARNING, LOG_BG, | 2844 | lpfc_printf_vlog(vport, KERN_WARNING, LOG_BG, |
2845 | "9034 BLKGRD: CDB: %02x %02x %02x %02x %02x " | 2845 | "9034 BLKGRD: CDB: %02x %02x %02x %02x %02x " |
2846 | "%02x %02x %02x %02x %02x \n", | 2846 | "%02x %02x %02x %02x %02x\n", |
2847 | cmnd->cmnd[0], cmnd->cmnd[1], cmnd->cmnd[2], | 2847 | cmnd->cmnd[0], cmnd->cmnd[1], cmnd->cmnd[2], |
2848 | cmnd->cmnd[3], cmnd->cmnd[4], cmnd->cmnd[5], | 2848 | cmnd->cmnd[3], cmnd->cmnd[4], cmnd->cmnd[5], |
2849 | cmnd->cmnd[6], cmnd->cmnd[7], cmnd->cmnd[8], | 2849 | cmnd->cmnd[6], cmnd->cmnd[7], cmnd->cmnd[8], |
@@ -2871,7 +2871,7 @@ lpfc_queuecommand(struct scsi_cmnd *cmnd, void (*done) (struct scsi_cmnd *)) | |||
2871 | dif_op_str[scsi_get_prot_op(cmnd)]); | 2871 | dif_op_str[scsi_get_prot_op(cmnd)]); |
2872 | lpfc_printf_vlog(vport, KERN_WARNING, LOG_BG, | 2872 | lpfc_printf_vlog(vport, KERN_WARNING, LOG_BG, |
2873 | "9039 BLKGRD: CDB: %02x %02x %02x %02x %02x " | 2873 | "9039 BLKGRD: CDB: %02x %02x %02x %02x %02x " |
2874 | "%02x %02x %02x %02x %02x \n", | 2874 | "%02x %02x %02x %02x %02x\n", |
2875 | cmnd->cmnd[0], cmnd->cmnd[1], cmnd->cmnd[2], | 2875 | cmnd->cmnd[0], cmnd->cmnd[1], cmnd->cmnd[2], |
2876 | cmnd->cmnd[3], cmnd->cmnd[4], cmnd->cmnd[5], | 2876 | cmnd->cmnd[3], cmnd->cmnd[4], cmnd->cmnd[5], |
2877 | cmnd->cmnd[6], cmnd->cmnd[7], cmnd->cmnd[8], | 2877 | cmnd->cmnd[6], cmnd->cmnd[7], cmnd->cmnd[8], |
@@ -3584,6 +3584,7 @@ struct scsi_host_template lpfc_template = { | |||
3584 | .use_clustering = ENABLE_CLUSTERING, | 3584 | .use_clustering = ENABLE_CLUSTERING, |
3585 | .shost_attrs = lpfc_hba_attrs, | 3585 | .shost_attrs = lpfc_hba_attrs, |
3586 | .max_sectors = 0xFFFF, | 3586 | .max_sectors = 0xFFFF, |
3587 | .vendor_id = LPFC_NL_VENDOR_ID, | ||
3587 | }; | 3588 | }; |
3588 | 3589 | ||
3589 | struct scsi_host_template lpfc_vport_template = { | 3590 | struct scsi_host_template lpfc_vport_template = { |
diff --git a/drivers/scsi/lpfc/lpfc_sli.c b/drivers/scsi/lpfc/lpfc_sli.c index acc43b061ba1..43cbe336f1f8 100644 --- a/drivers/scsi/lpfc/lpfc_sli.c +++ b/drivers/scsi/lpfc/lpfc_sli.c | |||
@@ -4139,7 +4139,7 @@ lpfc_sli4_read_fcoe_params(struct lpfc_hba *phba, | |||
4139 | return -EIO; | 4139 | return -EIO; |
4140 | } | 4140 | } |
4141 | data_length = mqe->un.mb_words[5]; | 4141 | data_length = mqe->un.mb_words[5]; |
4142 | if (data_length > DMP_FCOEPARAM_RGN_SIZE) { | 4142 | if (data_length > DMP_RGN23_SIZE) { |
4143 | lpfc_mbuf_free(phba, mp->virt, mp->phys); | 4143 | lpfc_mbuf_free(phba, mp->virt, mp->phys); |
4144 | kfree(mp); | 4144 | kfree(mp); |
4145 | return -EIO; | 4145 | return -EIO; |
@@ -4304,7 +4304,7 @@ lpfc_sli4_hba_setup(struct lpfc_hba *phba) | |||
4304 | */ | 4304 | */ |
4305 | if (lpfc_sli4_read_fcoe_params(phba, mboxq)) | 4305 | if (lpfc_sli4_read_fcoe_params(phba, mboxq)) |
4306 | lpfc_printf_log(phba, KERN_ERR, LOG_MBOX | LOG_INIT, | 4306 | lpfc_printf_log(phba, KERN_ERR, LOG_MBOX | LOG_INIT, |
4307 | "2570 Failed to read FCoE parameters \n"); | 4307 | "2570 Failed to read FCoE parameters\n"); |
4308 | 4308 | ||
4309 | /* Issue READ_REV to collect vpd and FW information. */ | 4309 | /* Issue READ_REV to collect vpd and FW information. */ |
4310 | vpd_size = PAGE_SIZE; | 4310 | vpd_size = PAGE_SIZE; |
@@ -4522,12 +4522,8 @@ lpfc_sli4_hba_setup(struct lpfc_hba *phba) | |||
4522 | lpfc_sli4_rb_setup(phba); | 4522 | lpfc_sli4_rb_setup(phba); |
4523 | 4523 | ||
4524 | /* Start the ELS watchdog timer */ | 4524 | /* Start the ELS watchdog timer */ |
4525 | /* | 4525 | mod_timer(&vport->els_tmofunc, |
4526 | * The driver for SLI4 is not yet ready to process timeouts | 4526 | jiffies + HZ * (phba->fc_ratov * 2)); |
4527 | * or interrupts. Once it is, the comment bars can be removed. | ||
4528 | */ | ||
4529 | /* mod_timer(&vport->els_tmofunc, | ||
4530 | * jiffies + HZ * (phba->fc_ratov*2)); */ | ||
4531 | 4527 | ||
4532 | /* Start heart beat timer */ | 4528 | /* Start heart beat timer */ |
4533 | mod_timer(&phba->hb_tmofunc, | 4529 | mod_timer(&phba->hb_tmofunc, |
@@ -4706,13 +4702,13 @@ lpfc_sli_issue_mbox_s3(struct lpfc_hba *phba, LPFC_MBOXQ_t *pmbox, | |||
4706 | 4702 | ||
4707 | spin_lock_irqsave(&phba->hbalock, drvr_flag); | 4703 | spin_lock_irqsave(&phba->hbalock, drvr_flag); |
4708 | if (!pmbox) { | 4704 | if (!pmbox) { |
4705 | phba->sli.sli_flag &= ~LPFC_SLI_MBOX_ACTIVE; | ||
4709 | /* processing mbox queue from intr_handler */ | 4706 | /* processing mbox queue from intr_handler */ |
4710 | if (unlikely(psli->sli_flag & LPFC_SLI_ASYNC_MBX_BLK)) { | 4707 | if (unlikely(psli->sli_flag & LPFC_SLI_ASYNC_MBX_BLK)) { |
4711 | spin_unlock_irqrestore(&phba->hbalock, drvr_flag); | 4708 | spin_unlock_irqrestore(&phba->hbalock, drvr_flag); |
4712 | return MBX_SUCCESS; | 4709 | return MBX_SUCCESS; |
4713 | } | 4710 | } |
4714 | processing_queue = 1; | 4711 | processing_queue = 1; |
4715 | phba->sli.sli_flag &= ~LPFC_SLI_MBOX_ACTIVE; | ||
4716 | pmbox = lpfc_mbox_get(phba); | 4712 | pmbox = lpfc_mbox_get(phba); |
4717 | if (!pmbox) { | 4713 | if (!pmbox) { |
4718 | spin_unlock_irqrestore(&phba->hbalock, drvr_flag); | 4714 | spin_unlock_irqrestore(&phba->hbalock, drvr_flag); |
@@ -5279,6 +5275,18 @@ lpfc_sli_issue_mbox_s4(struct lpfc_hba *phba, LPFC_MBOXQ_t *mboxq, | |||
5279 | unsigned long iflags; | 5275 | unsigned long iflags; |
5280 | int rc; | 5276 | int rc; |
5281 | 5277 | ||
5278 | rc = lpfc_mbox_dev_check(phba); | ||
5279 | if (unlikely(rc)) { | ||
5280 | lpfc_printf_log(phba, KERN_ERR, LOG_MBOX | LOG_SLI, | ||
5281 | "(%d):2544 Mailbox command x%x (x%x) " | ||
5282 | "cannot issue Data: x%x x%x\n", | ||
5283 | mboxq->vport ? mboxq->vport->vpi : 0, | ||
5284 | mboxq->u.mb.mbxCommand, | ||
5285 | lpfc_sli4_mbox_opcode_get(phba, mboxq), | ||
5286 | psli->sli_flag, flag); | ||
5287 | goto out_not_finished; | ||
5288 | } | ||
5289 | |||
5282 | /* Detect polling mode and jump to a handler */ | 5290 | /* Detect polling mode and jump to a handler */ |
5283 | if (!phba->sli4_hba.intr_enable) { | 5291 | if (!phba->sli4_hba.intr_enable) { |
5284 | if (flag == MBX_POLL) | 5292 | if (flag == MBX_POLL) |
@@ -5338,17 +5346,6 @@ lpfc_sli_issue_mbox_s4(struct lpfc_hba *phba, LPFC_MBOXQ_t *mboxq, | |||
5338 | psli->sli_flag, flag); | 5346 | psli->sli_flag, flag); |
5339 | goto out_not_finished; | 5347 | goto out_not_finished; |
5340 | } | 5348 | } |
5341 | rc = lpfc_mbox_dev_check(phba); | ||
5342 | if (unlikely(rc)) { | ||
5343 | lpfc_printf_log(phba, KERN_ERR, LOG_MBOX | LOG_SLI, | ||
5344 | "(%d):2544 Mailbox command x%x (x%x) " | ||
5345 | "cannot issue Data: x%x x%x\n", | ||
5346 | mboxq->vport ? mboxq->vport->vpi : 0, | ||
5347 | mboxq->u.mb.mbxCommand, | ||
5348 | lpfc_sli4_mbox_opcode_get(phba, mboxq), | ||
5349 | psli->sli_flag, flag); | ||
5350 | goto out_not_finished; | ||
5351 | } | ||
5352 | 5349 | ||
5353 | /* Put the mailbox command to the driver internal FIFO */ | 5350 | /* Put the mailbox command to the driver internal FIFO */ |
5354 | psli->slistat.mbox_busy++; | 5351 | psli->slistat.mbox_busy++; |
@@ -5817,19 +5814,21 @@ lpfc_sli4_bpl2sgl(struct lpfc_hba *phba, struct lpfc_iocbq *piocbq, | |||
5817 | /** | 5814 | /** |
5818 | * lpfc_sli4_scmd_to_wqidx_distr - scsi command to SLI4 WQ index distribution | 5815 | * lpfc_sli4_scmd_to_wqidx_distr - scsi command to SLI4 WQ index distribution |
5819 | * @phba: Pointer to HBA context object. | 5816 | * @phba: Pointer to HBA context object. |
5820 | * @piocb: Pointer to command iocb. | ||
5821 | * | 5817 | * |
5822 | * This routine performs a round robin SCSI command to SLI4 FCP WQ index | 5818 | * This routine performs a round robin SCSI command to SLI4 FCP WQ index |
5823 | * distribution. | 5819 | * distribution. This is called by __lpfc_sli_issue_iocb_s4() with the hbalock |
5820 | * held. | ||
5824 | * | 5821 | * |
5825 | * Return: index into SLI4 fast-path FCP queue index. | 5822 | * Return: index into SLI4 fast-path FCP queue index. |
5826 | **/ | 5823 | **/ |
5827 | static uint32_t | 5824 | static uint32_t |
5828 | lpfc_sli4_scmd_to_wqidx_distr(struct lpfc_hba *phba, struct lpfc_iocbq *piocb) | 5825 | lpfc_sli4_scmd_to_wqidx_distr(struct lpfc_hba *phba) |
5829 | { | 5826 | { |
5830 | static uint32_t fcp_qidx; | 5827 | ++phba->fcp_qidx; |
5828 | if (phba->fcp_qidx >= phba->cfg_fcp_wq_count) | ||
5829 | phba->fcp_qidx = 0; | ||
5831 | 5830 | ||
5832 | return fcp_qidx++ % phba->cfg_fcp_wq_count; | 5831 | return phba->fcp_qidx; |
5833 | } | 5832 | } |
5834 | 5833 | ||
5835 | /** | 5834 | /** |
@@ -6156,7 +6155,7 @@ __lpfc_sli_issue_iocb_s4(struct lpfc_hba *phba, uint32_t ring_number, | |||
6156 | return IOCB_ERROR; | 6155 | return IOCB_ERROR; |
6157 | 6156 | ||
6158 | if (piocb->iocb_flag & LPFC_IO_FCP) { | 6157 | if (piocb->iocb_flag & LPFC_IO_FCP) { |
6159 | fcp_wqidx = lpfc_sli4_scmd_to_wqidx_distr(phba, piocb); | 6158 | fcp_wqidx = lpfc_sli4_scmd_to_wqidx_distr(phba); |
6160 | if (lpfc_sli4_wq_put(phba->sli4_hba.fcp_wq[fcp_wqidx], &wqe)) | 6159 | if (lpfc_sli4_wq_put(phba->sli4_hba.fcp_wq[fcp_wqidx], &wqe)) |
6161 | return IOCB_ERROR; | 6160 | return IOCB_ERROR; |
6162 | } else { | 6161 | } else { |
@@ -6327,7 +6326,7 @@ lpfc_sli_async_event_handler(struct lpfc_hba * phba, | |||
6327 | KERN_ERR, | 6326 | KERN_ERR, |
6328 | LOG_SLI, | 6327 | LOG_SLI, |
6329 | "0346 Ring %d handler: unexpected ASYNC_STATUS" | 6328 | "0346 Ring %d handler: unexpected ASYNC_STATUS" |
6330 | " evt_code 0x%x \n" | 6329 | " evt_code 0x%x\n" |
6331 | "W0 0x%08x W1 0x%08x W2 0x%08x W3 0x%08x\n" | 6330 | "W0 0x%08x W1 0x%08x W2 0x%08x W3 0x%08x\n" |
6332 | "W4 0x%08x W5 0x%08x W6 0x%08x W7 0x%08x\n" | 6331 | "W4 0x%08x W5 0x%08x W6 0x%08x W7 0x%08x\n" |
6333 | "W8 0x%08x W9 0x%08x W10 0x%08x W11 0x%08x\n" | 6332 | "W8 0x%08x W9 0x%08x W10 0x%08x W11 0x%08x\n" |
@@ -6790,6 +6789,33 @@ lpfc_sli_pcimem_bcopy(void *srcp, void *destp, uint32_t cnt) | |||
6790 | 6789 | ||
6791 | 6790 | ||
6792 | /** | 6791 | /** |
6792 | * lpfc_sli_bemem_bcopy - SLI memory copy function | ||
6793 | * @srcp: Source memory pointer. | ||
6794 | * @destp: Destination memory pointer. | ||
6795 | * @cnt: Number of words required to be copied. | ||
6796 | * | ||
6797 | * This function is used for copying data between a data structure | ||
6798 | * with big endian representation to local endianness. | ||
6799 | * This function can be called with or without lock. | ||
6800 | **/ | ||
6801 | void | ||
6802 | lpfc_sli_bemem_bcopy(void *srcp, void *destp, uint32_t cnt) | ||
6803 | { | ||
6804 | uint32_t *src = srcp; | ||
6805 | uint32_t *dest = destp; | ||
6806 | uint32_t ldata; | ||
6807 | int i; | ||
6808 | |||
6809 | for (i = 0; i < (int)cnt; i += sizeof(uint32_t)) { | ||
6810 | ldata = *src; | ||
6811 | ldata = be32_to_cpu(ldata); | ||
6812 | *dest = ldata; | ||
6813 | src++; | ||
6814 | dest++; | ||
6815 | } | ||
6816 | } | ||
6817 | |||
6818 | /** | ||
6793 | * lpfc_sli_ringpostbuf_put - Function to add a buffer to postbufq | 6819 | * lpfc_sli_ringpostbuf_put - Function to add a buffer to postbufq |
6794 | * @phba: Pointer to HBA context object. | 6820 | * @phba: Pointer to HBA context object. |
6795 | * @pring: Pointer to driver SLI ring object. | 6821 | * @pring: Pointer to driver SLI ring object. |
@@ -7678,12 +7704,6 @@ lpfc_sli4_eratt_read(struct lpfc_hba *phba) | |||
7678 | "online0_reg=0x%x, online1_reg=0x%x\n", | 7704 | "online0_reg=0x%x, online1_reg=0x%x\n", |
7679 | uerr_sta_lo, uerr_sta_hi, | 7705 | uerr_sta_lo, uerr_sta_hi, |
7680 | onlnreg0, onlnreg1); | 7706 | onlnreg0, onlnreg1); |
7681 | /* TEMP: as the driver error recover logic is not | ||
7682 | * fully developed, we just log the error message | ||
7683 | * and the device error attention action is now | ||
7684 | * temporarily disabled. | ||
7685 | */ | ||
7686 | return 0; | ||
7687 | phba->work_status[0] = uerr_sta_lo; | 7707 | phba->work_status[0] = uerr_sta_lo; |
7688 | phba->work_status[1] = uerr_sta_hi; | 7708 | phba->work_status[1] = uerr_sta_hi; |
7689 | /* Set the driver HA work bitmap */ | 7709 | /* Set the driver HA work bitmap */ |
@@ -9499,8 +9519,7 @@ lpfc_eq_create(struct lpfc_hba *phba, struct lpfc_queue *eq, uint16_t imax) | |||
9499 | eq->host_index = 0; | 9519 | eq->host_index = 0; |
9500 | eq->hba_index = 0; | 9520 | eq->hba_index = 0; |
9501 | 9521 | ||
9502 | if (rc != MBX_TIMEOUT) | 9522 | mempool_free(mbox, phba->mbox_mem_pool); |
9503 | mempool_free(mbox, phba->mbox_mem_pool); | ||
9504 | return status; | 9523 | return status; |
9505 | } | 9524 | } |
9506 | 9525 | ||
@@ -9604,10 +9623,9 @@ lpfc_cq_create(struct lpfc_hba *phba, struct lpfc_queue *cq, | |||
9604 | cq->queue_id = bf_get(lpfc_mbx_cq_create_q_id, &cq_create->u.response); | 9623 | cq->queue_id = bf_get(lpfc_mbx_cq_create_q_id, &cq_create->u.response); |
9605 | cq->host_index = 0; | 9624 | cq->host_index = 0; |
9606 | cq->hba_index = 0; | 9625 | cq->hba_index = 0; |
9607 | out: | ||
9608 | 9626 | ||
9609 | if (rc != MBX_TIMEOUT) | 9627 | out: |
9610 | mempool_free(mbox, phba->mbox_mem_pool); | 9628 | mempool_free(mbox, phba->mbox_mem_pool); |
9611 | return status; | 9629 | return status; |
9612 | } | 9630 | } |
9613 | 9631 | ||
@@ -9712,8 +9730,7 @@ lpfc_mq_create(struct lpfc_hba *phba, struct lpfc_queue *mq, | |||
9712 | /* link the mq onto the parent cq child list */ | 9730 | /* link the mq onto the parent cq child list */ |
9713 | list_add_tail(&mq->list, &cq->child_list); | 9731 | list_add_tail(&mq->list, &cq->child_list); |
9714 | out: | 9732 | out: |
9715 | if (rc != MBX_TIMEOUT) | 9733 | mempool_free(mbox, phba->mbox_mem_pool); |
9716 | mempool_free(mbox, phba->mbox_mem_pool); | ||
9717 | return status; | 9734 | return status; |
9718 | } | 9735 | } |
9719 | 9736 | ||
@@ -9795,8 +9812,7 @@ lpfc_wq_create(struct lpfc_hba *phba, struct lpfc_queue *wq, | |||
9795 | /* link the wq onto the parent cq child list */ | 9812 | /* link the wq onto the parent cq child list */ |
9796 | list_add_tail(&wq->list, &cq->child_list); | 9813 | list_add_tail(&wq->list, &cq->child_list); |
9797 | out: | 9814 | out: |
9798 | if (rc != MBX_TIMEOUT) | 9815 | mempool_free(mbox, phba->mbox_mem_pool); |
9799 | mempool_free(mbox, phba->mbox_mem_pool); | ||
9800 | return status; | 9816 | return status; |
9801 | } | 9817 | } |
9802 | 9818 | ||
@@ -9970,8 +9986,7 @@ lpfc_rq_create(struct lpfc_hba *phba, struct lpfc_queue *hrq, | |||
9970 | list_add_tail(&drq->list, &cq->child_list); | 9986 | list_add_tail(&drq->list, &cq->child_list); |
9971 | 9987 | ||
9972 | out: | 9988 | out: |
9973 | if (rc != MBX_TIMEOUT) | 9989 | mempool_free(mbox, phba->mbox_mem_pool); |
9974 | mempool_free(mbox, phba->mbox_mem_pool); | ||
9975 | return status; | 9990 | return status; |
9976 | } | 9991 | } |
9977 | 9992 | ||
@@ -10026,8 +10041,7 @@ lpfc_eq_destroy(struct lpfc_hba *phba, struct lpfc_queue *eq) | |||
10026 | 10041 | ||
10027 | /* Remove eq from any list */ | 10042 | /* Remove eq from any list */ |
10028 | list_del_init(&eq->list); | 10043 | list_del_init(&eq->list); |
10029 | if (rc != MBX_TIMEOUT) | 10044 | mempool_free(mbox, eq->phba->mbox_mem_pool); |
10030 | mempool_free(mbox, eq->phba->mbox_mem_pool); | ||
10031 | return status; | 10045 | return status; |
10032 | } | 10046 | } |
10033 | 10047 | ||
@@ -10080,8 +10094,7 @@ lpfc_cq_destroy(struct lpfc_hba *phba, struct lpfc_queue *cq) | |||
10080 | } | 10094 | } |
10081 | /* Remove cq from any list */ | 10095 | /* Remove cq from any list */ |
10082 | list_del_init(&cq->list); | 10096 | list_del_init(&cq->list); |
10083 | if (rc != MBX_TIMEOUT) | 10097 | mempool_free(mbox, cq->phba->mbox_mem_pool); |
10084 | mempool_free(mbox, cq->phba->mbox_mem_pool); | ||
10085 | return status; | 10098 | return status; |
10086 | } | 10099 | } |
10087 | 10100 | ||
@@ -10134,8 +10147,7 @@ lpfc_mq_destroy(struct lpfc_hba *phba, struct lpfc_queue *mq) | |||
10134 | } | 10147 | } |
10135 | /* Remove mq from any list */ | 10148 | /* Remove mq from any list */ |
10136 | list_del_init(&mq->list); | 10149 | list_del_init(&mq->list); |
10137 | if (rc != MBX_TIMEOUT) | 10150 | mempool_free(mbox, mq->phba->mbox_mem_pool); |
10138 | mempool_free(mbox, mq->phba->mbox_mem_pool); | ||
10139 | return status; | 10151 | return status; |
10140 | } | 10152 | } |
10141 | 10153 | ||
@@ -10187,8 +10199,7 @@ lpfc_wq_destroy(struct lpfc_hba *phba, struct lpfc_queue *wq) | |||
10187 | } | 10199 | } |
10188 | /* Remove wq from any list */ | 10200 | /* Remove wq from any list */ |
10189 | list_del_init(&wq->list); | 10201 | list_del_init(&wq->list); |
10190 | if (rc != MBX_TIMEOUT) | 10202 | mempool_free(mbox, wq->phba->mbox_mem_pool); |
10191 | mempool_free(mbox, wq->phba->mbox_mem_pool); | ||
10192 | return status; | 10203 | return status; |
10193 | } | 10204 | } |
10194 | 10205 | ||
@@ -10258,8 +10269,7 @@ lpfc_rq_destroy(struct lpfc_hba *phba, struct lpfc_queue *hrq, | |||
10258 | } | 10269 | } |
10259 | list_del_init(&hrq->list); | 10270 | list_del_init(&hrq->list); |
10260 | list_del_init(&drq->list); | 10271 | list_del_init(&drq->list); |
10261 | if (rc != MBX_TIMEOUT) | 10272 | mempool_free(mbox, hrq->phba->mbox_mem_pool); |
10262 | mempool_free(mbox, hrq->phba->mbox_mem_pool); | ||
10263 | return status; | 10273 | return status; |
10264 | } | 10274 | } |
10265 | 10275 | ||
@@ -10933,6 +10943,7 @@ lpfc_prep_seq(struct lpfc_vport *vport, struct hbq_dmabuf *seq_dmabuf) | |||
10933 | first_iocbq = lpfc_sli_get_iocbq(vport->phba); | 10943 | first_iocbq = lpfc_sli_get_iocbq(vport->phba); |
10934 | if (first_iocbq) { | 10944 | if (first_iocbq) { |
10935 | /* Initialize the first IOCB. */ | 10945 | /* Initialize the first IOCB. */ |
10946 | first_iocbq->iocb.unsli3.rcvsli3.acc_len = 0; | ||
10936 | first_iocbq->iocb.ulpStatus = IOSTAT_SUCCESS; | 10947 | first_iocbq->iocb.ulpStatus = IOSTAT_SUCCESS; |
10937 | first_iocbq->iocb.ulpCommand = CMD_IOCB_RCV_SEQ64_CX; | 10948 | first_iocbq->iocb.ulpCommand = CMD_IOCB_RCV_SEQ64_CX; |
10938 | first_iocbq->iocb.ulpContext = be16_to_cpu(fc_hdr->fh_ox_id); | 10949 | first_iocbq->iocb.ulpContext = be16_to_cpu(fc_hdr->fh_ox_id); |
@@ -10945,6 +10956,8 @@ lpfc_prep_seq(struct lpfc_vport *vport, struct hbq_dmabuf *seq_dmabuf) | |||
10945 | first_iocbq->iocb.un.cont64[0].tus.f.bdeSize = | 10956 | first_iocbq->iocb.un.cont64[0].tus.f.bdeSize = |
10946 | LPFC_DATA_BUF_SIZE; | 10957 | LPFC_DATA_BUF_SIZE; |
10947 | first_iocbq->iocb.un.rcvels.remoteID = sid; | 10958 | first_iocbq->iocb.un.rcvels.remoteID = sid; |
10959 | first_iocbq->iocb.unsli3.rcvsli3.acc_len += | ||
10960 | bf_get(lpfc_rcqe_length, &seq_dmabuf->rcqe); | ||
10948 | } | 10961 | } |
10949 | iocbq = first_iocbq; | 10962 | iocbq = first_iocbq; |
10950 | /* | 10963 | /* |
@@ -10961,6 +10974,8 @@ lpfc_prep_seq(struct lpfc_vport *vport, struct hbq_dmabuf *seq_dmabuf) | |||
10961 | iocbq->iocb.ulpBdeCount++; | 10974 | iocbq->iocb.ulpBdeCount++; |
10962 | iocbq->iocb.unsli3.rcvsli3.bde2.tus.f.bdeSize = | 10975 | iocbq->iocb.unsli3.rcvsli3.bde2.tus.f.bdeSize = |
10963 | LPFC_DATA_BUF_SIZE; | 10976 | LPFC_DATA_BUF_SIZE; |
10977 | first_iocbq->iocb.unsli3.rcvsli3.acc_len += | ||
10978 | bf_get(lpfc_rcqe_length, &seq_dmabuf->rcqe); | ||
10964 | } else { | 10979 | } else { |
10965 | iocbq = lpfc_sli_get_iocbq(vport->phba); | 10980 | iocbq = lpfc_sli_get_iocbq(vport->phba); |
10966 | if (!iocbq) { | 10981 | if (!iocbq) { |
@@ -10978,6 +10993,8 @@ lpfc_prep_seq(struct lpfc_vport *vport, struct hbq_dmabuf *seq_dmabuf) | |||
10978 | iocbq->iocb.ulpBdeCount = 1; | 10993 | iocbq->iocb.ulpBdeCount = 1; |
10979 | iocbq->iocb.un.cont64[0].tus.f.bdeSize = | 10994 | iocbq->iocb.un.cont64[0].tus.f.bdeSize = |
10980 | LPFC_DATA_BUF_SIZE; | 10995 | LPFC_DATA_BUF_SIZE; |
10996 | first_iocbq->iocb.unsli3.rcvsli3.acc_len += | ||
10997 | bf_get(lpfc_rcqe_length, &seq_dmabuf->rcqe); | ||
10981 | iocbq->iocb.un.rcvels.remoteID = sid; | 10998 | iocbq->iocb.un.rcvels.remoteID = sid; |
10982 | list_add_tail(&iocbq->list, &first_iocbq->list); | 10999 | list_add_tail(&iocbq->list, &first_iocbq->list); |
10983 | } | 11000 | } |
@@ -11324,7 +11341,7 @@ lpfc_sli4_init_vpi(struct lpfc_hba *phba, uint16_t vpi) | |||
11324 | mboxq = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); | 11341 | mboxq = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); |
11325 | if (!mboxq) | 11342 | if (!mboxq) |
11326 | return -ENOMEM; | 11343 | return -ENOMEM; |
11327 | lpfc_init_vpi(mboxq, vpi); | 11344 | lpfc_init_vpi(phba, mboxq, vpi); |
11328 | mbox_tmo = lpfc_mbox_tmo_val(phba, MBX_INIT_VPI); | 11345 | mbox_tmo = lpfc_mbox_tmo_val(phba, MBX_INIT_VPI); |
11329 | rc = lpfc_sli_issue_mbox_wait(phba, mboxq, mbox_tmo); | 11346 | rc = lpfc_sli_issue_mbox_wait(phba, mboxq, mbox_tmo); |
11330 | if (rc != MBX_TIMEOUT) | 11347 | if (rc != MBX_TIMEOUT) |
@@ -11519,6 +11536,7 @@ lpfc_sli4_read_fcf_record(struct lpfc_hba *phba, uint16_t fcf_index) | |||
11519 | uint32_t alloc_len, req_len; | 11536 | uint32_t alloc_len, req_len; |
11520 | struct lpfc_mbx_read_fcf_tbl *read_fcf; | 11537 | struct lpfc_mbx_read_fcf_tbl *read_fcf; |
11521 | 11538 | ||
11539 | phba->fcoe_eventtag_at_fcf_scan = phba->fcoe_eventtag; | ||
11522 | mboxq = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); | 11540 | mboxq = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); |
11523 | if (!mboxq) { | 11541 | if (!mboxq) { |
11524 | lpfc_printf_log(phba, KERN_ERR, LOG_INIT, | 11542 | lpfc_printf_log(phba, KERN_ERR, LOG_INIT, |
@@ -11570,7 +11588,140 @@ lpfc_sli4_read_fcf_record(struct lpfc_hba *phba, uint16_t fcf_index) | |||
11570 | if (rc == MBX_NOT_FINISHED) { | 11588 | if (rc == MBX_NOT_FINISHED) { |
11571 | lpfc_sli4_mbox_cmd_free(phba, mboxq); | 11589 | lpfc_sli4_mbox_cmd_free(phba, mboxq); |
11572 | error = -EIO; | 11590 | error = -EIO; |
11573 | } else | 11591 | } else { |
11592 | spin_lock_irq(&phba->hbalock); | ||
11593 | phba->hba_flag |= FCF_DISC_INPROGRESS; | ||
11594 | spin_unlock_irq(&phba->hbalock); | ||
11574 | error = 0; | 11595 | error = 0; |
11596 | } | ||
11575 | return error; | 11597 | return error; |
11576 | } | 11598 | } |
11599 | |||
11600 | /** | ||
11601 | * lpfc_sli_read_link_ste - Read region 23 to decide if link is disabled. | ||
11602 | * @phba: pointer to lpfc hba data structure. | ||
11603 | * | ||
11604 | * This function read region 23 and parse TLV for port status to | ||
11605 | * decide if the user disaled the port. If the TLV indicates the | ||
11606 | * port is disabled, the hba_flag is set accordingly. | ||
11607 | **/ | ||
11608 | void | ||
11609 | lpfc_sli_read_link_ste(struct lpfc_hba *phba) | ||
11610 | { | ||
11611 | LPFC_MBOXQ_t *pmb = NULL; | ||
11612 | MAILBOX_t *mb; | ||
11613 | uint8_t *rgn23_data = NULL; | ||
11614 | uint32_t offset = 0, data_size, sub_tlv_len, tlv_offset; | ||
11615 | int rc; | ||
11616 | |||
11617 | pmb = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); | ||
11618 | if (!pmb) { | ||
11619 | lpfc_printf_log(phba, KERN_ERR, LOG_INIT, | ||
11620 | "2600 lpfc_sli_read_serdes_param failed to" | ||
11621 | " allocate mailbox memory\n"); | ||
11622 | goto out; | ||
11623 | } | ||
11624 | mb = &pmb->u.mb; | ||
11625 | |||
11626 | /* Get adapter Region 23 data */ | ||
11627 | rgn23_data = kzalloc(DMP_RGN23_SIZE, GFP_KERNEL); | ||
11628 | if (!rgn23_data) | ||
11629 | goto out; | ||
11630 | |||
11631 | do { | ||
11632 | lpfc_dump_mem(phba, pmb, offset, DMP_REGION_23); | ||
11633 | rc = lpfc_sli_issue_mbox(phba, pmb, MBX_POLL); | ||
11634 | |||
11635 | if (rc != MBX_SUCCESS) { | ||
11636 | lpfc_printf_log(phba, KERN_INFO, LOG_INIT, | ||
11637 | "2601 lpfc_sli_read_link_ste failed to" | ||
11638 | " read config region 23 rc 0x%x Status 0x%x\n", | ||
11639 | rc, mb->mbxStatus); | ||
11640 | mb->un.varDmp.word_cnt = 0; | ||
11641 | } | ||
11642 | /* | ||
11643 | * dump mem may return a zero when finished or we got a | ||
11644 | * mailbox error, either way we are done. | ||
11645 | */ | ||
11646 | if (mb->un.varDmp.word_cnt == 0) | ||
11647 | break; | ||
11648 | if (mb->un.varDmp.word_cnt > DMP_RGN23_SIZE - offset) | ||
11649 | mb->un.varDmp.word_cnt = DMP_RGN23_SIZE - offset; | ||
11650 | |||
11651 | lpfc_sli_pcimem_bcopy(((uint8_t *)mb) + DMP_RSP_OFFSET, | ||
11652 | rgn23_data + offset, | ||
11653 | mb->un.varDmp.word_cnt); | ||
11654 | offset += mb->un.varDmp.word_cnt; | ||
11655 | } while (mb->un.varDmp.word_cnt && offset < DMP_RGN23_SIZE); | ||
11656 | |||
11657 | data_size = offset; | ||
11658 | offset = 0; | ||
11659 | |||
11660 | if (!data_size) | ||
11661 | goto out; | ||
11662 | |||
11663 | /* Check the region signature first */ | ||
11664 | if (memcmp(&rgn23_data[offset], LPFC_REGION23_SIGNATURE, 4)) { | ||
11665 | lpfc_printf_log(phba, KERN_ERR, LOG_INIT, | ||
11666 | "2619 Config region 23 has bad signature\n"); | ||
11667 | goto out; | ||
11668 | } | ||
11669 | offset += 4; | ||
11670 | |||
11671 | /* Check the data structure version */ | ||
11672 | if (rgn23_data[offset] != LPFC_REGION23_VERSION) { | ||
11673 | lpfc_printf_log(phba, KERN_ERR, LOG_INIT, | ||
11674 | "2620 Config region 23 has bad version\n"); | ||
11675 | goto out; | ||
11676 | } | ||
11677 | offset += 4; | ||
11678 | |||
11679 | /* Parse TLV entries in the region */ | ||
11680 | while (offset < data_size) { | ||
11681 | if (rgn23_data[offset] == LPFC_REGION23_LAST_REC) | ||
11682 | break; | ||
11683 | /* | ||
11684 | * If the TLV is not driver specific TLV or driver id is | ||
11685 | * not linux driver id, skip the record. | ||
11686 | */ | ||
11687 | if ((rgn23_data[offset] != DRIVER_SPECIFIC_TYPE) || | ||
11688 | (rgn23_data[offset + 2] != LINUX_DRIVER_ID) || | ||
11689 | (rgn23_data[offset + 3] != 0)) { | ||
11690 | offset += rgn23_data[offset + 1] * 4 + 4; | ||
11691 | continue; | ||
11692 | } | ||
11693 | |||
11694 | /* Driver found a driver specific TLV in the config region */ | ||
11695 | sub_tlv_len = rgn23_data[offset + 1] * 4; | ||
11696 | offset += 4; | ||
11697 | tlv_offset = 0; | ||
11698 | |||
11699 | /* | ||
11700 | * Search for configured port state sub-TLV. | ||
11701 | */ | ||
11702 | while ((offset < data_size) && | ||
11703 | (tlv_offset < sub_tlv_len)) { | ||
11704 | if (rgn23_data[offset] == LPFC_REGION23_LAST_REC) { | ||
11705 | offset += 4; | ||
11706 | tlv_offset += 4; | ||
11707 | break; | ||
11708 | } | ||
11709 | if (rgn23_data[offset] != PORT_STE_TYPE) { | ||
11710 | offset += rgn23_data[offset + 1] * 4 + 4; | ||
11711 | tlv_offset += rgn23_data[offset + 1] * 4 + 4; | ||
11712 | continue; | ||
11713 | } | ||
11714 | |||
11715 | /* This HBA contains PORT_STE configured */ | ||
11716 | if (!rgn23_data[offset + 2]) | ||
11717 | phba->hba_flag |= LINK_DISABLED; | ||
11718 | |||
11719 | goto out; | ||
11720 | } | ||
11721 | } | ||
11722 | out: | ||
11723 | if (pmb) | ||
11724 | mempool_free(pmb, phba->mbox_mem_pool); | ||
11725 | kfree(rgn23_data); | ||
11726 | return; | ||
11727 | } | ||
diff --git a/drivers/scsi/lpfc/lpfc_sli4.h b/drivers/scsi/lpfc/lpfc_sli4.h index 3b276b47d18f..b5f4ba1a5c27 100644 --- a/drivers/scsi/lpfc/lpfc_sli4.h +++ b/drivers/scsi/lpfc/lpfc_sli4.h | |||
@@ -132,6 +132,7 @@ struct lpfc_sli4_link { | |||
132 | 132 | ||
133 | struct lpfc_fcf { | 133 | struct lpfc_fcf { |
134 | uint8_t fabric_name[8]; | 134 | uint8_t fabric_name[8]; |
135 | uint8_t switch_name[8]; | ||
135 | uint8_t mac_addr[6]; | 136 | uint8_t mac_addr[6]; |
136 | uint16_t fcf_indx; | 137 | uint16_t fcf_indx; |
137 | uint16_t fcfi; | 138 | uint16_t fcfi; |
@@ -150,6 +151,10 @@ struct lpfc_fcf { | |||
150 | #define LPFC_REGION23_SIGNATURE "RG23" | 151 | #define LPFC_REGION23_SIGNATURE "RG23" |
151 | #define LPFC_REGION23_VERSION 1 | 152 | #define LPFC_REGION23_VERSION 1 |
152 | #define LPFC_REGION23_LAST_REC 0xff | 153 | #define LPFC_REGION23_LAST_REC 0xff |
154 | #define DRIVER_SPECIFIC_TYPE 0xA2 | ||
155 | #define LINUX_DRIVER_ID 0x20 | ||
156 | #define PORT_STE_TYPE 0x1 | ||
157 | |||
153 | struct lpfc_fip_param_hdr { | 158 | struct lpfc_fip_param_hdr { |
154 | uint8_t type; | 159 | uint8_t type; |
155 | #define FCOE_PARAM_TYPE 0xA0 | 160 | #define FCOE_PARAM_TYPE 0xA0 |
diff --git a/drivers/scsi/lpfc/lpfc_version.h b/drivers/scsi/lpfc/lpfc_version.h index 41094e02304b..9ae20af4bdb7 100644 --- a/drivers/scsi/lpfc/lpfc_version.h +++ b/drivers/scsi/lpfc/lpfc_version.h | |||
@@ -18,7 +18,7 @@ | |||
18 | * included with this package. * | 18 | * included with this package. * |
19 | *******************************************************************/ | 19 | *******************************************************************/ |
20 | 20 | ||
21 | #define LPFC_DRIVER_VERSION "8.3.3" | 21 | #define LPFC_DRIVER_VERSION "8.3.4" |
22 | 22 | ||
23 | #define LPFC_DRIVER_NAME "lpfc" | 23 | #define LPFC_DRIVER_NAME "lpfc" |
24 | #define LPFC_SP_DRIVER_HANDLER_NAME "lpfc:sp" | 24 | #define LPFC_SP_DRIVER_HANDLER_NAME "lpfc:sp" |
diff --git a/drivers/scsi/lpfc/lpfc_vport.c b/drivers/scsi/lpfc/lpfc_vport.c index e0b49922193e..606efa767548 100644 --- a/drivers/scsi/lpfc/lpfc_vport.c +++ b/drivers/scsi/lpfc/lpfc_vport.c | |||
@@ -313,22 +313,6 @@ lpfc_vport_create(struct fc_vport *fc_vport, bool disable) | |||
313 | goto error_out; | 313 | goto error_out; |
314 | } | 314 | } |
315 | 315 | ||
316 | /* | ||
317 | * In SLI4, the vpi must be activated before it can be used | ||
318 | * by the port. | ||
319 | */ | ||
320 | if (phba->sli_rev == LPFC_SLI_REV4) { | ||
321 | rc = lpfc_sli4_init_vpi(phba, vpi); | ||
322 | if (rc) { | ||
323 | lpfc_printf_log(phba, KERN_ERR, LOG_VPORT, | ||
324 | "1838 Failed to INIT_VPI on vpi %d " | ||
325 | "status %d\n", vpi, rc); | ||
326 | rc = VPORT_NORESOURCES; | ||
327 | lpfc_free_vpi(phba, vpi); | ||
328 | goto error_out; | ||
329 | } | ||
330 | } | ||
331 | |||
332 | /* Assign an unused board number */ | 316 | /* Assign an unused board number */ |
333 | if ((instance = lpfc_get_instance()) < 0) { | 317 | if ((instance = lpfc_get_instance()) < 0) { |
334 | lpfc_printf_log(phba, KERN_ERR, LOG_VPORT, | 318 | lpfc_printf_log(phba, KERN_ERR, LOG_VPORT, |
@@ -367,12 +351,8 @@ lpfc_vport_create(struct fc_vport *fc_vport, bool disable) | |||
367 | goto error_out; | 351 | goto error_out; |
368 | } | 352 | } |
369 | 353 | ||
370 | memcpy(vport->fc_portname.u.wwn, vport->fc_sparam.portName.u.wwn, 8); | 354 | u64_to_wwn(fc_vport->node_name, vport->fc_nodename.u.wwn); |
371 | memcpy(vport->fc_nodename.u.wwn, vport->fc_sparam.nodeName.u.wwn, 8); | 355 | u64_to_wwn(fc_vport->port_name, vport->fc_portname.u.wwn); |
372 | if (fc_vport->node_name != 0) | ||
373 | u64_to_wwn(fc_vport->node_name, vport->fc_nodename.u.wwn); | ||
374 | if (fc_vport->port_name != 0) | ||
375 | u64_to_wwn(fc_vport->port_name, vport->fc_portname.u.wwn); | ||
376 | 356 | ||
377 | memcpy(&vport->fc_sparam.portName, vport->fc_portname.u.wwn, 8); | 357 | memcpy(&vport->fc_sparam.portName, vport->fc_portname.u.wwn, 8); |
378 | memcpy(&vport->fc_sparam.nodeName, vport->fc_nodename.u.wwn, 8); | 358 | memcpy(&vport->fc_sparam.nodeName, vport->fc_nodename.u.wwn, 8); |
@@ -404,7 +384,34 @@ lpfc_vport_create(struct fc_vport *fc_vport, bool disable) | |||
404 | *(struct lpfc_vport **)fc_vport->dd_data = vport; | 384 | *(struct lpfc_vport **)fc_vport->dd_data = vport; |
405 | vport->fc_vport = fc_vport; | 385 | vport->fc_vport = fc_vport; |
406 | 386 | ||
387 | /* | ||
388 | * In SLI4, the vpi must be activated before it can be used | ||
389 | * by the port. | ||
390 | */ | ||
391 | if ((phba->sli_rev == LPFC_SLI_REV4) && | ||
392 | (pport->vfi_state & LPFC_VFI_REGISTERED)) { | ||
393 | rc = lpfc_sli4_init_vpi(phba, vpi); | ||
394 | if (rc) { | ||
395 | lpfc_printf_log(phba, KERN_ERR, LOG_VPORT, | ||
396 | "1838 Failed to INIT_VPI on vpi %d " | ||
397 | "status %d\n", vpi, rc); | ||
398 | rc = VPORT_NORESOURCES; | ||
399 | lpfc_free_vpi(phba, vpi); | ||
400 | goto error_out; | ||
401 | } | ||
402 | } else if (phba->sli_rev == LPFC_SLI_REV4) { | ||
403 | /* | ||
404 | * Driver cannot INIT_VPI now. Set the flags to | ||
405 | * init_vpi when reg_vfi complete. | ||
406 | */ | ||
407 | vport->fc_flag |= FC_VPORT_NEEDS_INIT_VPI; | ||
408 | lpfc_vport_set_state(vport, FC_VPORT_LINKDOWN); | ||
409 | rc = VPORT_OK; | ||
410 | goto out; | ||
411 | } | ||
412 | |||
407 | if ((phba->link_state < LPFC_LINK_UP) || | 413 | if ((phba->link_state < LPFC_LINK_UP) || |
414 | (pport->port_state < LPFC_FABRIC_CFG_LINK) || | ||
408 | (phba->fc_topology == TOPOLOGY_LOOP)) { | 415 | (phba->fc_topology == TOPOLOGY_LOOP)) { |
409 | lpfc_vport_set_state(vport, FC_VPORT_LINKDOWN); | 416 | lpfc_vport_set_state(vport, FC_VPORT_LINKDOWN); |
410 | rc = VPORT_OK; | 417 | rc = VPORT_OK; |
@@ -661,7 +668,7 @@ lpfc_vport_delete(struct fc_vport *fc_vport) | |||
661 | lpfc_printf_log(vport->phba, KERN_WARNING, | 668 | lpfc_printf_log(vport->phba, KERN_WARNING, |
662 | LOG_VPORT, | 669 | LOG_VPORT, |
663 | "1829 CT command failed to " | 670 | "1829 CT command failed to " |
664 | "delete objects on fabric. \n"); | 671 | "delete objects on fabric\n"); |
665 | } | 672 | } |
666 | /* First look for the Fabric ndlp */ | 673 | /* First look for the Fabric ndlp */ |
667 | ndlp = lpfc_findnode_did(vport, Fabric_DID); | 674 | ndlp = lpfc_findnode_did(vport, Fabric_DID); |
diff --git a/drivers/scsi/mpt2sas/mpt2sas_base.c b/drivers/scsi/mpt2sas/mpt2sas_base.c index 35a13867495e..d95d2f274cb3 100644 --- a/drivers/scsi/mpt2sas/mpt2sas_base.c +++ b/drivers/scsi/mpt2sas/mpt2sas_base.c | |||
@@ -94,7 +94,7 @@ _base_fault_reset_work(struct work_struct *work) | |||
94 | int rc; | 94 | int rc; |
95 | 95 | ||
96 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); | 96 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); |
97 | if (ioc->ioc_reset_in_progress) | 97 | if (ioc->shost_recovery) |
98 | goto rearm_timer; | 98 | goto rearm_timer; |
99 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | 99 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); |
100 | 100 | ||
@@ -687,6 +687,14 @@ _base_unmask_interrupts(struct MPT2SAS_ADAPTER *ioc) | |||
687 | ioc->mask_interrupts = 0; | 687 | ioc->mask_interrupts = 0; |
688 | } | 688 | } |
689 | 689 | ||
690 | union reply_descriptor { | ||
691 | u64 word; | ||
692 | struct { | ||
693 | u32 low; | ||
694 | u32 high; | ||
695 | } u; | ||
696 | }; | ||
697 | |||
690 | /** | 698 | /** |
691 | * _base_interrupt - MPT adapter (IOC) specific interrupt handler. | 699 | * _base_interrupt - MPT adapter (IOC) specific interrupt handler. |
692 | * @irq: irq number (not used) | 700 | * @irq: irq number (not used) |
@@ -698,47 +706,38 @@ _base_unmask_interrupts(struct MPT2SAS_ADAPTER *ioc) | |||
698 | static irqreturn_t | 706 | static irqreturn_t |
699 | _base_interrupt(int irq, void *bus_id) | 707 | _base_interrupt(int irq, void *bus_id) |
700 | { | 708 | { |
701 | union reply_descriptor { | ||
702 | u64 word; | ||
703 | struct { | ||
704 | u32 low; | ||
705 | u32 high; | ||
706 | } u; | ||
707 | }; | ||
708 | union reply_descriptor rd; | 709 | union reply_descriptor rd; |
709 | u32 post_index, post_index_next, completed_cmds; | 710 | u32 completed_cmds; |
710 | u8 request_desript_type; | 711 | u8 request_desript_type; |
711 | u16 smid; | 712 | u16 smid; |
712 | u8 cb_idx; | 713 | u8 cb_idx; |
713 | u32 reply; | 714 | u32 reply; |
714 | u8 VF_ID; | 715 | u8 VF_ID; |
715 | int i; | ||
716 | struct MPT2SAS_ADAPTER *ioc = bus_id; | 716 | struct MPT2SAS_ADAPTER *ioc = bus_id; |
717 | Mpi2ReplyDescriptorsUnion_t *rpf; | ||
717 | 718 | ||
718 | if (ioc->mask_interrupts) | 719 | if (ioc->mask_interrupts) |
719 | return IRQ_NONE; | 720 | return IRQ_NONE; |
720 | 721 | ||
721 | post_index = ioc->reply_post_host_index; | 722 | rpf = &ioc->reply_post_free[ioc->reply_post_host_index]; |
722 | request_desript_type = ioc->reply_post_free[post_index]. | 723 | request_desript_type = rpf->Default.ReplyFlags |
723 | Default.ReplyFlags & MPI2_RPY_DESCRIPT_FLAGS_TYPE_MASK; | 724 | & MPI2_RPY_DESCRIPT_FLAGS_TYPE_MASK; |
724 | if (request_desript_type == MPI2_RPY_DESCRIPT_FLAGS_UNUSED) | 725 | if (request_desript_type == MPI2_RPY_DESCRIPT_FLAGS_UNUSED) |
725 | return IRQ_NONE; | 726 | return IRQ_NONE; |
726 | 727 | ||
727 | completed_cmds = 0; | 728 | completed_cmds = 0; |
728 | do { | 729 | do { |
729 | rd.word = ioc->reply_post_free[post_index].Words; | 730 | rd.word = rpf->Words; |
730 | if (rd.u.low == UINT_MAX || rd.u.high == UINT_MAX) | 731 | if (rd.u.low == UINT_MAX || rd.u.high == UINT_MAX) |
731 | goto out; | 732 | goto out; |
732 | reply = 0; | 733 | reply = 0; |
733 | cb_idx = 0xFF; | 734 | cb_idx = 0xFF; |
734 | smid = le16_to_cpu(ioc->reply_post_free[post_index]. | 735 | smid = le16_to_cpu(rpf->Default.DescriptorTypeDependent1); |
735 | Default.DescriptorTypeDependent1); | 736 | VF_ID = rpf->Default.VF_ID; |
736 | VF_ID = ioc->reply_post_free[post_index]. | ||
737 | Default.VF_ID; | ||
738 | if (request_desript_type == | 737 | if (request_desript_type == |
739 | MPI2_RPY_DESCRIPT_FLAGS_ADDRESS_REPLY) { | 738 | MPI2_RPY_DESCRIPT_FLAGS_ADDRESS_REPLY) { |
740 | reply = le32_to_cpu(ioc->reply_post_free[post_index]. | 739 | reply = le32_to_cpu |
741 | AddressReply.ReplyFrameAddress); | 740 | (rpf->AddressReply.ReplyFrameAddress); |
742 | } else if (request_desript_type == | 741 | } else if (request_desript_type == |
743 | MPI2_RPY_DESCRIPT_FLAGS_TARGET_COMMAND_BUFFER) | 742 | MPI2_RPY_DESCRIPT_FLAGS_TARGET_COMMAND_BUFFER) |
744 | goto next; | 743 | goto next; |
@@ -765,21 +764,27 @@ _base_interrupt(int irq, void *bus_id) | |||
765 | 0 : ioc->reply_free_host_index + 1; | 764 | 0 : ioc->reply_free_host_index + 1; |
766 | ioc->reply_free[ioc->reply_free_host_index] = | 765 | ioc->reply_free[ioc->reply_free_host_index] = |
767 | cpu_to_le32(reply); | 766 | cpu_to_le32(reply); |
767 | wmb(); | ||
768 | writel(ioc->reply_free_host_index, | 768 | writel(ioc->reply_free_host_index, |
769 | &ioc->chip->ReplyFreeHostIndex); | 769 | &ioc->chip->ReplyFreeHostIndex); |
770 | wmb(); | ||
771 | } | 770 | } |
772 | 771 | ||
773 | next: | 772 | next: |
774 | post_index_next = (post_index == (ioc->reply_post_queue_depth - | 773 | |
775 | 1)) ? 0 : post_index + 1; | 774 | rpf->Words = ULLONG_MAX; |
775 | ioc->reply_post_host_index = (ioc->reply_post_host_index == | ||
776 | (ioc->reply_post_queue_depth - 1)) ? 0 : | ||
777 | ioc->reply_post_host_index + 1; | ||
776 | request_desript_type = | 778 | request_desript_type = |
777 | ioc->reply_post_free[post_index_next].Default.ReplyFlags | 779 | ioc->reply_post_free[ioc->reply_post_host_index].Default. |
778 | & MPI2_RPY_DESCRIPT_FLAGS_TYPE_MASK; | 780 | ReplyFlags & MPI2_RPY_DESCRIPT_FLAGS_TYPE_MASK; |
779 | completed_cmds++; | 781 | completed_cmds++; |
780 | if (request_desript_type == MPI2_RPY_DESCRIPT_FLAGS_UNUSED) | 782 | if (request_desript_type == MPI2_RPY_DESCRIPT_FLAGS_UNUSED) |
781 | goto out; | 783 | goto out; |
782 | post_index = post_index_next; | 784 | if (!ioc->reply_post_host_index) |
785 | rpf = ioc->reply_post_free; | ||
786 | else | ||
787 | rpf++; | ||
783 | } while (1); | 788 | } while (1); |
784 | 789 | ||
785 | out: | 790 | out: |
@@ -787,19 +792,8 @@ _base_interrupt(int irq, void *bus_id) | |||
787 | if (!completed_cmds) | 792 | if (!completed_cmds) |
788 | return IRQ_NONE; | 793 | return IRQ_NONE; |
789 | 794 | ||
790 | /* reply post descriptor handling */ | ||
791 | post_index_next = ioc->reply_post_host_index; | ||
792 | for (i = 0 ; i < completed_cmds; i++) { | ||
793 | post_index = post_index_next; | ||
794 | /* poison the reply post descriptor */ | ||
795 | ioc->reply_post_free[post_index_next].Words = ULLONG_MAX; | ||
796 | post_index_next = (post_index == | ||
797 | (ioc->reply_post_queue_depth - 1)) | ||
798 | ? 0 : post_index + 1; | ||
799 | } | ||
800 | ioc->reply_post_host_index = post_index_next; | ||
801 | writel(post_index_next, &ioc->chip->ReplyPostHostIndex); | ||
802 | wmb(); | 795 | wmb(); |
796 | writel(ioc->reply_post_host_index, &ioc->chip->ReplyPostHostIndex); | ||
803 | return IRQ_HANDLED; | 797 | return IRQ_HANDLED; |
804 | } | 798 | } |
805 | 799 | ||
@@ -1542,6 +1536,8 @@ _base_display_ioc_capabilities(struct MPT2SAS_ADAPTER *ioc) | |||
1542 | (ioc->bios_pg3.BiosVersion & 0x0000FF00) >> 8, | 1536 | (ioc->bios_pg3.BiosVersion & 0x0000FF00) >> 8, |
1543 | ioc->bios_pg3.BiosVersion & 0x000000FF); | 1537 | ioc->bios_pg3.BiosVersion & 0x000000FF); |
1544 | 1538 | ||
1539 | _base_display_dell_branding(ioc); | ||
1540 | |||
1545 | printk(MPT2SAS_INFO_FMT "Protocol=(", ioc->name); | 1541 | printk(MPT2SAS_INFO_FMT "Protocol=(", ioc->name); |
1546 | 1542 | ||
1547 | if (ioc->facts.ProtocolFlags & MPI2_IOCFACTS_PROTOCOL_SCSI_INITIATOR) { | 1543 | if (ioc->facts.ProtocolFlags & MPI2_IOCFACTS_PROTOCOL_SCSI_INITIATOR) { |
@@ -1554,8 +1550,6 @@ _base_display_ioc_capabilities(struct MPT2SAS_ADAPTER *ioc) | |||
1554 | i++; | 1550 | i++; |
1555 | } | 1551 | } |
1556 | 1552 | ||
1557 | _base_display_dell_branding(ioc); | ||
1558 | |||
1559 | i = 0; | 1553 | i = 0; |
1560 | printk("), "); | 1554 | printk("), "); |
1561 | printk("Capabilities=("); | 1555 | printk("Capabilities=("); |
@@ -1627,6 +1621,9 @@ _base_static_config_pages(struct MPT2SAS_ADAPTER *ioc) | |||
1627 | u32 iounit_pg1_flags; | 1621 | u32 iounit_pg1_flags; |
1628 | 1622 | ||
1629 | mpt2sas_config_get_manufacturing_pg0(ioc, &mpi_reply, &ioc->manu_pg0); | 1623 | mpt2sas_config_get_manufacturing_pg0(ioc, &mpi_reply, &ioc->manu_pg0); |
1624 | if (ioc->ir_firmware) | ||
1625 | mpt2sas_config_get_manufacturing_pg10(ioc, &mpi_reply, | ||
1626 | &ioc->manu_pg10); | ||
1630 | mpt2sas_config_get_bios_pg2(ioc, &mpi_reply, &ioc->bios_pg2); | 1627 | mpt2sas_config_get_bios_pg2(ioc, &mpi_reply, &ioc->bios_pg2); |
1631 | mpt2sas_config_get_bios_pg3(ioc, &mpi_reply, &ioc->bios_pg3); | 1628 | mpt2sas_config_get_bios_pg3(ioc, &mpi_reply, &ioc->bios_pg3); |
1632 | mpt2sas_config_get_ioc_pg8(ioc, &mpi_reply, &ioc->ioc_pg8); | 1629 | mpt2sas_config_get_ioc_pg8(ioc, &mpi_reply, &ioc->ioc_pg8); |
@@ -1647,7 +1644,7 @@ _base_static_config_pages(struct MPT2SAS_ADAPTER *ioc) | |||
1647 | iounit_pg1_flags |= | 1644 | iounit_pg1_flags |= |
1648 | MPI2_IOUNITPAGE1_DISABLE_TASK_SET_FULL_HANDLING; | 1645 | MPI2_IOUNITPAGE1_DISABLE_TASK_SET_FULL_HANDLING; |
1649 | ioc->iounit_pg1.Flags = cpu_to_le32(iounit_pg1_flags); | 1646 | ioc->iounit_pg1.Flags = cpu_to_le32(iounit_pg1_flags); |
1650 | mpt2sas_config_set_iounit_pg1(ioc, &mpi_reply, ioc->iounit_pg1); | 1647 | mpt2sas_config_set_iounit_pg1(ioc, &mpi_reply, &ioc->iounit_pg1); |
1651 | } | 1648 | } |
1652 | 1649 | ||
1653 | /** | 1650 | /** |
@@ -3303,13 +3300,11 @@ mpt2sas_base_attach(struct MPT2SAS_ADAPTER *ioc) | |||
3303 | ioc->tm_cmds.reply = kzalloc(ioc->reply_sz, GFP_KERNEL); | 3300 | ioc->tm_cmds.reply = kzalloc(ioc->reply_sz, GFP_KERNEL); |
3304 | ioc->tm_cmds.status = MPT2_CMD_NOT_USED; | 3301 | ioc->tm_cmds.status = MPT2_CMD_NOT_USED; |
3305 | mutex_init(&ioc->tm_cmds.mutex); | 3302 | mutex_init(&ioc->tm_cmds.mutex); |
3306 | init_completion(&ioc->tm_cmds.done); | ||
3307 | 3303 | ||
3308 | /* config page internal command bits */ | 3304 | /* config page internal command bits */ |
3309 | ioc->config_cmds.reply = kzalloc(ioc->reply_sz, GFP_KERNEL); | 3305 | ioc->config_cmds.reply = kzalloc(ioc->reply_sz, GFP_KERNEL); |
3310 | ioc->config_cmds.status = MPT2_CMD_NOT_USED; | 3306 | ioc->config_cmds.status = MPT2_CMD_NOT_USED; |
3311 | mutex_init(&ioc->config_cmds.mutex); | 3307 | mutex_init(&ioc->config_cmds.mutex); |
3312 | init_completion(&ioc->config_cmds.done); | ||
3313 | 3308 | ||
3314 | /* ctl module internal command bits */ | 3309 | /* ctl module internal command bits */ |
3315 | ioc->ctl_cmds.reply = kzalloc(ioc->reply_sz, GFP_KERNEL); | 3310 | ioc->ctl_cmds.reply = kzalloc(ioc->reply_sz, GFP_KERNEL); |
@@ -3433,6 +3428,7 @@ _base_reset_handler(struct MPT2SAS_ADAPTER *ioc, int reset_phase) | |||
3433 | if (ioc->config_cmds.status & MPT2_CMD_PENDING) { | 3428 | if (ioc->config_cmds.status & MPT2_CMD_PENDING) { |
3434 | ioc->config_cmds.status |= MPT2_CMD_RESET; | 3429 | ioc->config_cmds.status |= MPT2_CMD_RESET; |
3435 | mpt2sas_base_free_smid(ioc, ioc->config_cmds.smid); | 3430 | mpt2sas_base_free_smid(ioc, ioc->config_cmds.smid); |
3431 | ioc->config_cmds.smid = USHORT_MAX; | ||
3436 | complete(&ioc->config_cmds.done); | 3432 | complete(&ioc->config_cmds.done); |
3437 | } | 3433 | } |
3438 | break; | 3434 | break; |
@@ -3501,20 +3497,13 @@ mpt2sas_base_hard_reset_handler(struct MPT2SAS_ADAPTER *ioc, int sleep_flag, | |||
3501 | __func__)); | 3497 | __func__)); |
3502 | 3498 | ||
3503 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); | 3499 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); |
3504 | if (ioc->ioc_reset_in_progress) { | 3500 | if (ioc->shost_recovery) { |
3505 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | 3501 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); |
3506 | printk(MPT2SAS_ERR_FMT "%s: busy\n", | 3502 | printk(MPT2SAS_ERR_FMT "%s: busy\n", |
3507 | ioc->name, __func__); | 3503 | ioc->name, __func__); |
3508 | return -EBUSY; | 3504 | return -EBUSY; |
3509 | } | 3505 | } |
3510 | ioc->ioc_reset_in_progress = 1; | ||
3511 | ioc->shost_recovery = 1; | 3506 | ioc->shost_recovery = 1; |
3512 | if (ioc->shost->shost_state == SHOST_RUNNING) { | ||
3513 | /* set back to SHOST_RUNNING in mpt2sas_scsih.c */ | ||
3514 | scsi_host_set_state(ioc->shost, SHOST_RECOVERY); | ||
3515 | printk(MPT2SAS_INFO_FMT "putting controller into " | ||
3516 | "SHOST_RECOVERY\n", ioc->name); | ||
3517 | } | ||
3518 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | 3507 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); |
3519 | 3508 | ||
3520 | _base_reset_handler(ioc, MPT2_IOC_PRE_RESET); | 3509 | _base_reset_handler(ioc, MPT2_IOC_PRE_RESET); |
@@ -3534,7 +3523,10 @@ mpt2sas_base_hard_reset_handler(struct MPT2SAS_ADAPTER *ioc, int sleep_flag, | |||
3534 | ioc->name, __func__, ((r == 0) ? "SUCCESS" : "FAILED"))); | 3523 | ioc->name, __func__, ((r == 0) ? "SUCCESS" : "FAILED"))); |
3535 | 3524 | ||
3536 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); | 3525 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); |
3537 | ioc->ioc_reset_in_progress = 0; | 3526 | ioc->shost_recovery = 0; |
3538 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | 3527 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); |
3528 | |||
3529 | if (!r) | ||
3530 | _base_reset_handler(ioc, MPT2_IOC_RUNNING); | ||
3539 | return r; | 3531 | return r; |
3540 | } | 3532 | } |
diff --git a/drivers/scsi/mpt2sas/mpt2sas_base.h b/drivers/scsi/mpt2sas/mpt2sas_base.h index acdcff150a35..2faab1e690e9 100644 --- a/drivers/scsi/mpt2sas/mpt2sas_base.h +++ b/drivers/scsi/mpt2sas/mpt2sas_base.h | |||
@@ -69,10 +69,10 @@ | |||
69 | #define MPT2SAS_DRIVER_NAME "mpt2sas" | 69 | #define MPT2SAS_DRIVER_NAME "mpt2sas" |
70 | #define MPT2SAS_AUTHOR "LSI Corporation <DL-MPTFusionLinux@lsi.com>" | 70 | #define MPT2SAS_AUTHOR "LSI Corporation <DL-MPTFusionLinux@lsi.com>" |
71 | #define MPT2SAS_DESCRIPTION "LSI MPT Fusion SAS 2.0 Device Driver" | 71 | #define MPT2SAS_DESCRIPTION "LSI MPT Fusion SAS 2.0 Device Driver" |
72 | #define MPT2SAS_DRIVER_VERSION "01.100.04.00" | 72 | #define MPT2SAS_DRIVER_VERSION "01.100.06.00" |
73 | #define MPT2SAS_MAJOR_VERSION 01 | 73 | #define MPT2SAS_MAJOR_VERSION 01 |
74 | #define MPT2SAS_MINOR_VERSION 100 | 74 | #define MPT2SAS_MINOR_VERSION 100 |
75 | #define MPT2SAS_BUILD_VERSION 04 | 75 | #define MPT2SAS_BUILD_VERSION 06 |
76 | #define MPT2SAS_RELEASE_VERSION 00 | 76 | #define MPT2SAS_RELEASE_VERSION 00 |
77 | 77 | ||
78 | /* | 78 | /* |
@@ -119,6 +119,7 @@ | |||
119 | #define MPT2_IOC_PRE_RESET 1 /* prior to host reset */ | 119 | #define MPT2_IOC_PRE_RESET 1 /* prior to host reset */ |
120 | #define MPT2_IOC_AFTER_RESET 2 /* just after host reset */ | 120 | #define MPT2_IOC_AFTER_RESET 2 /* just after host reset */ |
121 | #define MPT2_IOC_DONE_RESET 3 /* links re-initialized */ | 121 | #define MPT2_IOC_DONE_RESET 3 /* links re-initialized */ |
122 | #define MPT2_IOC_RUNNING 4 /* shost running */ | ||
122 | 123 | ||
123 | /* | 124 | /* |
124 | * logging format | 125 | * logging format |
@@ -196,6 +197,38 @@ struct MPT2SAS_TARGET { | |||
196 | * @block: device is in SDEV_BLOCK state | 197 | * @block: device is in SDEV_BLOCK state |
197 | * @tlr_snoop_check: flag used in determining whether to disable TLR | 198 | * @tlr_snoop_check: flag used in determining whether to disable TLR |
198 | */ | 199 | */ |
200 | |||
201 | /* OEM Identifiers */ | ||
202 | #define MFG10_OEM_ID_INVALID (0x00000000) | ||
203 | #define MFG10_OEM_ID_DELL (0x00000001) | ||
204 | #define MFG10_OEM_ID_FSC (0x00000002) | ||
205 | #define MFG10_OEM_ID_SUN (0x00000003) | ||
206 | #define MFG10_OEM_ID_IBM (0x00000004) | ||
207 | |||
208 | /* GENERIC Flags 0*/ | ||
209 | #define MFG10_GF0_OCE_DISABLED (0x00000001) | ||
210 | #define MFG10_GF0_R1E_DRIVE_COUNT (0x00000002) | ||
211 | #define MFG10_GF0_R10_DISPLAY (0x00000004) | ||
212 | #define MFG10_GF0_SSD_DATA_SCRUB_DISABLE (0x00000008) | ||
213 | #define MFG10_GF0_SINGLE_DRIVE_R0 (0x00000010) | ||
214 | |||
215 | /* OEM Specific Flags will come from OEM specific header files */ | ||
216 | typedef struct _MPI2_CONFIG_PAGE_MAN_10 { | ||
217 | MPI2_CONFIG_PAGE_HEADER Header; /* 00h */ | ||
218 | U8 OEMIdentifier; /* 04h */ | ||
219 | U8 Reserved1; /* 05h */ | ||
220 | U16 Reserved2; /* 08h */ | ||
221 | U32 Reserved3; /* 0Ch */ | ||
222 | U32 GenericFlags0; /* 10h */ | ||
223 | U32 GenericFlags1; /* 14h */ | ||
224 | U32 Reserved4; /* 18h */ | ||
225 | U32 OEMSpecificFlags0; /* 1Ch */ | ||
226 | U32 OEMSpecificFlags1; /* 20h */ | ||
227 | U32 Reserved5[18]; /* 24h-60h*/ | ||
228 | } MPI2_CONFIG_PAGE_MAN_10, | ||
229 | MPI2_POINTER PTR_MPI2_CONFIG_PAGE_MAN_10, | ||
230 | Mpi2ManufacturingPage10_t, MPI2_POINTER pMpi2ManufacturingPage10_t; | ||
231 | |||
199 | struct MPT2SAS_DEVICE { | 232 | struct MPT2SAS_DEVICE { |
200 | struct MPT2SAS_TARGET *sas_target; | 233 | struct MPT2SAS_TARGET *sas_target; |
201 | unsigned int lun; | 234 | unsigned int lun; |
@@ -431,7 +464,7 @@ typedef void (*MPT_ADD_SGE)(void *paddr, u32 flags_length, dma_addr_t dma_addr); | |||
431 | * @fw_event_list: list of fw events | 464 | * @fw_event_list: list of fw events |
432 | * @aen_event_read_flag: event log was read | 465 | * @aen_event_read_flag: event log was read |
433 | * @broadcast_aen_busy: broadcast aen waiting to be serviced | 466 | * @broadcast_aen_busy: broadcast aen waiting to be serviced |
434 | * @ioc_reset_in_progress: host reset in progress | 467 | * @shost_recovery: host reset in progress |
435 | * @ioc_reset_in_progress_lock: | 468 | * @ioc_reset_in_progress_lock: |
436 | * @ioc_link_reset_in_progress: phy/hard reset in progress | 469 | * @ioc_link_reset_in_progress: phy/hard reset in progress |
437 | * @ignore_loginfos: ignore loginfos during task managment | 470 | * @ignore_loginfos: ignore loginfos during task managment |
@@ -460,6 +493,7 @@ typedef void (*MPT_ADD_SGE)(void *paddr, u32 flags_length, dma_addr_t dma_addr); | |||
460 | * @facts: static facts data | 493 | * @facts: static facts data |
461 | * @pfacts: static port facts data | 494 | * @pfacts: static port facts data |
462 | * @manu_pg0: static manufacturing page 0 | 495 | * @manu_pg0: static manufacturing page 0 |
496 | * @manu_pg10: static manufacturing page 10 | ||
463 | * @bios_pg2: static bios page 2 | 497 | * @bios_pg2: static bios page 2 |
464 | * @bios_pg3: static bios page 3 | 498 | * @bios_pg3: static bios page 3 |
465 | * @ioc_pg8: static ioc page 8 | 499 | * @ioc_pg8: static ioc page 8 |
@@ -544,7 +578,6 @@ struct MPT2SAS_ADAPTER { | |||
544 | /* misc flags */ | 578 | /* misc flags */ |
545 | int aen_event_read_flag; | 579 | int aen_event_read_flag; |
546 | u8 broadcast_aen_busy; | 580 | u8 broadcast_aen_busy; |
547 | u8 ioc_reset_in_progress; | ||
548 | u8 shost_recovery; | 581 | u8 shost_recovery; |
549 | spinlock_t ioc_reset_in_progress_lock; | 582 | spinlock_t ioc_reset_in_progress_lock; |
550 | u8 ioc_link_reset_in_progress; | 583 | u8 ioc_link_reset_in_progress; |
@@ -663,6 +696,7 @@ struct MPT2SAS_ADAPTER { | |||
663 | dma_addr_t diag_buffer_dma[MPI2_DIAG_BUF_TYPE_COUNT]; | 696 | dma_addr_t diag_buffer_dma[MPI2_DIAG_BUF_TYPE_COUNT]; |
664 | u8 diag_buffer_status[MPI2_DIAG_BUF_TYPE_COUNT]; | 697 | u8 diag_buffer_status[MPI2_DIAG_BUF_TYPE_COUNT]; |
665 | u32 unique_id[MPI2_DIAG_BUF_TYPE_COUNT]; | 698 | u32 unique_id[MPI2_DIAG_BUF_TYPE_COUNT]; |
699 | Mpi2ManufacturingPage10_t manu_pg10; | ||
666 | u32 product_specific[MPI2_DIAG_BUF_TYPE_COUNT][23]; | 700 | u32 product_specific[MPI2_DIAG_BUF_TYPE_COUNT][23]; |
667 | u32 diagnostic_flags[MPI2_DIAG_BUF_TYPE_COUNT]; | 701 | u32 diagnostic_flags[MPI2_DIAG_BUF_TYPE_COUNT]; |
668 | }; | 702 | }; |
@@ -734,6 +768,8 @@ void mpt2sas_config_done(struct MPT2SAS_ADAPTER *ioc, u16 smid, u8 VF_ID, u32 re | |||
734 | int mpt2sas_config_get_number_hba_phys(struct MPT2SAS_ADAPTER *ioc, u8 *num_phys); | 768 | int mpt2sas_config_get_number_hba_phys(struct MPT2SAS_ADAPTER *ioc, u8 *num_phys); |
735 | int mpt2sas_config_get_manufacturing_pg0(struct MPT2SAS_ADAPTER *ioc, | 769 | int mpt2sas_config_get_manufacturing_pg0(struct MPT2SAS_ADAPTER *ioc, |
736 | Mpi2ConfigReply_t *mpi_reply, Mpi2ManufacturingPage0_t *config_page); | 770 | Mpi2ConfigReply_t *mpi_reply, Mpi2ManufacturingPage0_t *config_page); |
771 | int mpt2sas_config_get_manufacturing_pg10(struct MPT2SAS_ADAPTER *ioc, | ||
772 | Mpi2ConfigReply_t *mpi_reply, Mpi2ManufacturingPage10_t *config_page); | ||
737 | int mpt2sas_config_get_bios_pg2(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | 773 | int mpt2sas_config_get_bios_pg2(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t |
738 | *mpi_reply, Mpi2BiosPage2_t *config_page); | 774 | *mpi_reply, Mpi2BiosPage2_t *config_page); |
739 | int mpt2sas_config_get_bios_pg3(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | 775 | int mpt2sas_config_get_bios_pg3(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t |
@@ -749,7 +785,7 @@ int mpt2sas_config_get_sas_iounit_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigRep | |||
749 | int mpt2sas_config_get_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | 785 | int mpt2sas_config_get_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t |
750 | *mpi_reply, Mpi2IOUnitPage1_t *config_page); | 786 | *mpi_reply, Mpi2IOUnitPage1_t *config_page); |
751 | int mpt2sas_config_set_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | 787 | int mpt2sas_config_set_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t |
752 | *mpi_reply, Mpi2IOUnitPage1_t config_page); | 788 | *mpi_reply, Mpi2IOUnitPage1_t *config_page); |
753 | int mpt2sas_config_get_sas_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | 789 | int mpt2sas_config_get_sas_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t |
754 | *mpi_reply, Mpi2SasIOUnitPage1_t *config_page, u16 sz); | 790 | *mpi_reply, Mpi2SasIOUnitPage1_t *config_page, u16 sz); |
755 | int mpt2sas_config_get_ioc_pg8(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | 791 | int mpt2sas_config_get_ioc_pg8(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t |
@@ -776,7 +812,6 @@ int mpt2sas_config_get_volume_handle(struct MPT2SAS_ADAPTER *ioc, u16 pd_handle, | |||
776 | u16 *volume_handle); | 812 | u16 *volume_handle); |
777 | int mpt2sas_config_get_volume_wwid(struct MPT2SAS_ADAPTER *ioc, u16 volume_handle, | 813 | int mpt2sas_config_get_volume_wwid(struct MPT2SAS_ADAPTER *ioc, u16 volume_handle, |
778 | u64 *wwid); | 814 | u64 *wwid); |
779 | |||
780 | /* ctl shared API */ | 815 | /* ctl shared API */ |
781 | extern struct device_attribute *mpt2sas_host_attrs[]; | 816 | extern struct device_attribute *mpt2sas_host_attrs[]; |
782 | extern struct device_attribute *mpt2sas_dev_attrs[]; | 817 | extern struct device_attribute *mpt2sas_dev_attrs[]; |
@@ -798,9 +833,11 @@ int mpt2sas_transport_add_host_phy(struct MPT2SAS_ADAPTER *ioc, struct _sas_phy | |||
798 | *mpt2sas_phy, Mpi2SasPhyPage0_t phy_pg0, struct device *parent_dev); | 833 | *mpt2sas_phy, Mpi2SasPhyPage0_t phy_pg0, struct device *parent_dev); |
799 | int mpt2sas_transport_add_expander_phy(struct MPT2SAS_ADAPTER *ioc, struct _sas_phy | 834 | int mpt2sas_transport_add_expander_phy(struct MPT2SAS_ADAPTER *ioc, struct _sas_phy |
800 | *mpt2sas_phy, Mpi2ExpanderPage1_t expander_pg1, struct device *parent_dev); | 835 | *mpt2sas_phy, Mpi2ExpanderPage1_t expander_pg1, struct device *parent_dev); |
801 | void mpt2sas_transport_update_phy_link_change(struct MPT2SAS_ADAPTER *ioc, u16 handle, | 836 | void mpt2sas_transport_update_links(struct MPT2SAS_ADAPTER *ioc, u16 handle, |
802 | u16 attached_handle, u8 phy_number, u8 link_rate); | 837 | u16 attached_handle, u8 phy_number, u8 link_rate); |
803 | extern struct sas_function_template mpt2sas_transport_functions; | 838 | extern struct sas_function_template mpt2sas_transport_functions; |
804 | extern struct scsi_transport_template *mpt2sas_transport_template; | 839 | extern struct scsi_transport_template *mpt2sas_transport_template; |
840 | extern int scsi_internal_device_block(struct scsi_device *sdev); | ||
841 | extern int scsi_internal_device_unblock(struct scsi_device *sdev); | ||
805 | 842 | ||
806 | #endif /* MPT2SAS_BASE_H_INCLUDED */ | 843 | #endif /* MPT2SAS_BASE_H_INCLUDED */ |
diff --git a/drivers/scsi/mpt2sas/mpt2sas_config.c b/drivers/scsi/mpt2sas/mpt2sas_config.c index 6ddee161beb3..ab8c560865d8 100644 --- a/drivers/scsi/mpt2sas/mpt2sas_config.c +++ b/drivers/scsi/mpt2sas/mpt2sas_config.c | |||
@@ -72,15 +72,15 @@ | |||
72 | 72 | ||
73 | /** | 73 | /** |
74 | * struct config_request - obtain dma memory via routine | 74 | * struct config_request - obtain dma memory via routine |
75 | * @config_page_sz: size | 75 | * @sz: size |
76 | * @config_page: virt pointer | 76 | * @page: virt pointer |
77 | * @config_page_dma: phys pointer | 77 | * @page_dma: phys pointer |
78 | * | 78 | * |
79 | */ | 79 | */ |
80 | struct config_request{ | 80 | struct config_request{ |
81 | u16 config_page_sz; | 81 | u16 sz; |
82 | void *config_page; | 82 | void *page; |
83 | dma_addr_t config_page_dma; | 83 | dma_addr_t page_dma; |
84 | }; | 84 | }; |
85 | 85 | ||
86 | #ifdef CONFIG_SCSI_MPT2SAS_LOGGING | 86 | #ifdef CONFIG_SCSI_MPT2SAS_LOGGING |
@@ -175,6 +175,55 @@ _config_display_some_debug(struct MPT2SAS_ADAPTER *ioc, u16 smid, | |||
175 | #endif | 175 | #endif |
176 | 176 | ||
177 | /** | 177 | /** |
178 | * _config_alloc_config_dma_memory - obtain physical memory | ||
179 | * @ioc: per adapter object | ||
180 | * @mem: struct config_request | ||
181 | * | ||
182 | * A wrapper for obtaining dma-able memory for config page request. | ||
183 | * | ||
184 | * Returns 0 for success, non-zero for failure. | ||
185 | */ | ||
186 | static int | ||
187 | _config_alloc_config_dma_memory(struct MPT2SAS_ADAPTER *ioc, | ||
188 | struct config_request *mem) | ||
189 | { | ||
190 | int r = 0; | ||
191 | |||
192 | if (mem->sz > ioc->config_page_sz) { | ||
193 | mem->page = dma_alloc_coherent(&ioc->pdev->dev, mem->sz, | ||
194 | &mem->page_dma, GFP_KERNEL); | ||
195 | if (!mem->page) { | ||
196 | printk(MPT2SAS_ERR_FMT "%s: dma_alloc_coherent" | ||
197 | " failed asking for (%d) bytes!!\n", | ||
198 | ioc->name, __func__, mem->sz); | ||
199 | r = -ENOMEM; | ||
200 | } | ||
201 | } else { /* use tmp buffer if less than 512 bytes */ | ||
202 | mem->page = ioc->config_page; | ||
203 | mem->page_dma = ioc->config_page_dma; | ||
204 | } | ||
205 | return r; | ||
206 | } | ||
207 | |||
208 | /** | ||
209 | * _config_free_config_dma_memory - wrapper to free the memory | ||
210 | * @ioc: per adapter object | ||
211 | * @mem: struct config_request | ||
212 | * | ||
213 | * A wrapper to free dma-able memory when using _config_alloc_config_dma_memory. | ||
214 | * | ||
215 | * Returns 0 for success, non-zero for failure. | ||
216 | */ | ||
217 | static void | ||
218 | _config_free_config_dma_memory(struct MPT2SAS_ADAPTER *ioc, | ||
219 | struct config_request *mem) | ||
220 | { | ||
221 | if (mem->sz > ioc->config_page_sz) | ||
222 | dma_free_coherent(&ioc->pdev->dev, mem->sz, mem->page, | ||
223 | mem->page_dma); | ||
224 | } | ||
225 | |||
226 | /** | ||
178 | * mpt2sas_config_done - config page completion routine | 227 | * mpt2sas_config_done - config page completion routine |
179 | * @ioc: per adapter object | 228 | * @ioc: per adapter object |
180 | * @smid: system request message index | 229 | * @smid: system request message index |
@@ -206,6 +255,7 @@ mpt2sas_config_done(struct MPT2SAS_ADAPTER *ioc, u16 smid, u8 VF_ID, u32 reply) | |||
206 | #ifdef CONFIG_SCSI_MPT2SAS_LOGGING | 255 | #ifdef CONFIG_SCSI_MPT2SAS_LOGGING |
207 | _config_display_some_debug(ioc, smid, "config_done", mpi_reply); | 256 | _config_display_some_debug(ioc, smid, "config_done", mpi_reply); |
208 | #endif | 257 | #endif |
258 | ioc->config_cmds.smid = USHORT_MAX; | ||
209 | complete(&ioc->config_cmds.done); | 259 | complete(&ioc->config_cmds.done); |
210 | } | 260 | } |
211 | 261 | ||
@@ -215,7 +265,9 @@ mpt2sas_config_done(struct MPT2SAS_ADAPTER *ioc, u16 smid, u8 VF_ID, u32 reply) | |||
215 | * @mpi_request: request message frame | 265 | * @mpi_request: request message frame |
216 | * @mpi_reply: reply mf payload returned from firmware | 266 | * @mpi_reply: reply mf payload returned from firmware |
217 | * @timeout: timeout in seconds | 267 | * @timeout: timeout in seconds |
218 | * Context: sleep, the calling function needs to acquire the config_cmds.mutex | 268 | * @config_page: contents of the config page |
269 | * @config_page_sz: size of config page | ||
270 | * Context: sleep | ||
219 | * | 271 | * |
220 | * A generic API for config page requests to firmware. | 272 | * A generic API for config page requests to firmware. |
221 | * | 273 | * |
@@ -228,16 +280,17 @@ mpt2sas_config_done(struct MPT2SAS_ADAPTER *ioc, u16 smid, u8 VF_ID, u32 reply) | |||
228 | */ | 280 | */ |
229 | static int | 281 | static int |
230 | _config_request(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigRequest_t | 282 | _config_request(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigRequest_t |
231 | *mpi_request, Mpi2ConfigReply_t *mpi_reply, int timeout) | 283 | *mpi_request, Mpi2ConfigReply_t *mpi_reply, int timeout, |
284 | void *config_page, u16 config_page_sz) | ||
232 | { | 285 | { |
233 | u16 smid; | 286 | u16 smid; |
234 | u32 ioc_state; | 287 | u32 ioc_state; |
235 | unsigned long timeleft; | 288 | unsigned long timeleft; |
236 | Mpi2ConfigRequest_t *config_request; | 289 | Mpi2ConfigRequest_t *config_request; |
237 | int r; | 290 | int r; |
238 | u8 retry_count; | 291 | u8 retry_count, issue_host_reset = 0; |
239 | u8 issue_host_reset = 0; | ||
240 | u16 wait_state_count; | 292 | u16 wait_state_count; |
293 | struct config_request mem; | ||
241 | 294 | ||
242 | mutex_lock(&ioc->config_cmds.mutex); | 295 | mutex_lock(&ioc->config_cmds.mutex); |
243 | if (ioc->config_cmds.status != MPT2_CMD_NOT_USED) { | 296 | if (ioc->config_cmds.status != MPT2_CMD_NOT_USED) { |
@@ -246,12 +299,44 @@ _config_request(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigRequest_t | |||
246 | mutex_unlock(&ioc->config_cmds.mutex); | 299 | mutex_unlock(&ioc->config_cmds.mutex); |
247 | return -EAGAIN; | 300 | return -EAGAIN; |
248 | } | 301 | } |
302 | |||
249 | retry_count = 0; | 303 | retry_count = 0; |
304 | memset(&mem, 0, sizeof(struct config_request)); | ||
305 | |||
306 | if (config_page) { | ||
307 | mpi_request->Header.PageVersion = mpi_reply->Header.PageVersion; | ||
308 | mpi_request->Header.PageNumber = mpi_reply->Header.PageNumber; | ||
309 | mpi_request->Header.PageType = mpi_reply->Header.PageType; | ||
310 | mpi_request->Header.PageLength = mpi_reply->Header.PageLength; | ||
311 | mpi_request->ExtPageLength = mpi_reply->ExtPageLength; | ||
312 | mpi_request->ExtPageType = mpi_reply->ExtPageType; | ||
313 | if (mpi_request->Header.PageLength) | ||
314 | mem.sz = mpi_request->Header.PageLength * 4; | ||
315 | else | ||
316 | mem.sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4; | ||
317 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
318 | if (r != 0) | ||
319 | goto out; | ||
320 | if (mpi_request->Action == | ||
321 | MPI2_CONFIG_ACTION_PAGE_WRITE_CURRENT) { | ||
322 | ioc->base_add_sg_single(&mpi_request->PageBufferSGE, | ||
323 | MPT2_CONFIG_COMMON_WRITE_SGLFLAGS | mem.sz, | ||
324 | mem.page_dma); | ||
325 | memcpy(mem.page, config_page, min_t(u16, mem.sz, | ||
326 | config_page_sz)); | ||
327 | } else { | ||
328 | memset(config_page, 0, config_page_sz); | ||
329 | ioc->base_add_sg_single(&mpi_request->PageBufferSGE, | ||
330 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.sz, mem.page_dma); | ||
331 | } | ||
332 | } | ||
250 | 333 | ||
251 | retry_config: | 334 | retry_config: |
252 | if (retry_count) { | 335 | if (retry_count) { |
253 | if (retry_count > 2) /* attempt only 2 retries */ | 336 | if (retry_count > 2) { /* attempt only 2 retries */ |
254 | return -EFAULT; | 337 | r = -EFAULT; |
338 | goto free_mem; | ||
339 | } | ||
255 | printk(MPT2SAS_INFO_FMT "%s: attempting retry (%d)\n", | 340 | printk(MPT2SAS_INFO_FMT "%s: attempting retry (%d)\n", |
256 | ioc->name, __func__, retry_count); | 341 | ioc->name, __func__, retry_count); |
257 | } | 342 | } |
@@ -262,8 +347,9 @@ _config_request(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigRequest_t | |||
262 | printk(MPT2SAS_ERR_FMT | 347 | printk(MPT2SAS_ERR_FMT |
263 | "%s: failed due to ioc not operational\n", | 348 | "%s: failed due to ioc not operational\n", |
264 | ioc->name, __func__); | 349 | ioc->name, __func__); |
350 | ioc->config_cmds.status = MPT2_CMD_NOT_USED; | ||
265 | r = -EFAULT; | 351 | r = -EFAULT; |
266 | goto out; | 352 | goto free_mem; |
267 | } | 353 | } |
268 | ssleep(1); | 354 | ssleep(1); |
269 | ioc_state = mpt2sas_base_get_iocstate(ioc, 1); | 355 | ioc_state = mpt2sas_base_get_iocstate(ioc, 1); |
@@ -279,8 +365,9 @@ _config_request(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigRequest_t | |||
279 | if (!smid) { | 365 | if (!smid) { |
280 | printk(MPT2SAS_ERR_FMT "%s: failed obtaining a smid\n", | 366 | printk(MPT2SAS_ERR_FMT "%s: failed obtaining a smid\n", |
281 | ioc->name, __func__); | 367 | ioc->name, __func__); |
368 | ioc->config_cmds.status = MPT2_CMD_NOT_USED; | ||
282 | r = -EAGAIN; | 369 | r = -EAGAIN; |
283 | goto out; | 370 | goto free_mem; |
284 | } | 371 | } |
285 | 372 | ||
286 | r = 0; | 373 | r = 0; |
@@ -292,6 +379,7 @@ _config_request(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigRequest_t | |||
292 | #ifdef CONFIG_SCSI_MPT2SAS_LOGGING | 379 | #ifdef CONFIG_SCSI_MPT2SAS_LOGGING |
293 | _config_display_some_debug(ioc, smid, "config_request", NULL); | 380 | _config_display_some_debug(ioc, smid, "config_request", NULL); |
294 | #endif | 381 | #endif |
382 | init_completion(&ioc->config_cmds.done); | ||
295 | mpt2sas_base_put_smid_default(ioc, smid, config_request->VF_ID); | 383 | mpt2sas_base_put_smid_default(ioc, smid, config_request->VF_ID); |
296 | timeleft = wait_for_completion_timeout(&ioc->config_cmds.done, | 384 | timeleft = wait_for_completion_timeout(&ioc->config_cmds.done, |
297 | timeout*HZ); | 385 | timeout*HZ); |
@@ -303,22 +391,31 @@ _config_request(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigRequest_t | |||
303 | retry_count++; | 391 | retry_count++; |
304 | if (ioc->config_cmds.smid == smid) | 392 | if (ioc->config_cmds.smid == smid) |
305 | mpt2sas_base_free_smid(ioc, smid); | 393 | mpt2sas_base_free_smid(ioc, smid); |
306 | if ((ioc->shost_recovery) || | 394 | if ((ioc->shost_recovery) || (ioc->config_cmds.status & |
307 | (ioc->config_cmds.status & MPT2_CMD_RESET)) | 395 | MPT2_CMD_RESET)) |
308 | goto retry_config; | 396 | goto retry_config; |
309 | issue_host_reset = 1; | 397 | issue_host_reset = 1; |
310 | r = -EFAULT; | 398 | r = -EFAULT; |
311 | goto out; | 399 | goto free_mem; |
312 | } | 400 | } |
401 | |||
313 | if (ioc->config_cmds.status & MPT2_CMD_REPLY_VALID) | 402 | if (ioc->config_cmds.status & MPT2_CMD_REPLY_VALID) |
314 | memcpy(mpi_reply, ioc->config_cmds.reply, | 403 | memcpy(mpi_reply, ioc->config_cmds.reply, |
315 | sizeof(Mpi2ConfigReply_t)); | 404 | sizeof(Mpi2ConfigReply_t)); |
316 | if (retry_count) | 405 | if (retry_count) |
317 | printk(MPT2SAS_INFO_FMT "%s: retry completed!!\n", | 406 | printk(MPT2SAS_INFO_FMT "%s: retry (%d) completed!!\n", |
318 | ioc->name, __func__); | 407 | ioc->name, __func__, retry_count); |
319 | out: | 408 | if (config_page && mpi_request->Action == |
409 | MPI2_CONFIG_ACTION_PAGE_READ_CURRENT) | ||
410 | memcpy(config_page, mem.page, min_t(u16, mem.sz, | ||
411 | config_page_sz)); | ||
412 | free_mem: | ||
413 | if (config_page) | ||
414 | _config_free_config_dma_memory(ioc, &mem); | ||
415 | out: | ||
320 | ioc->config_cmds.status = MPT2_CMD_NOT_USED; | 416 | ioc->config_cmds.status = MPT2_CMD_NOT_USED; |
321 | mutex_unlock(&ioc->config_cmds.mutex); | 417 | mutex_unlock(&ioc->config_cmds.mutex); |
418 | |||
322 | if (issue_host_reset) | 419 | if (issue_host_reset) |
323 | mpt2sas_base_hard_reset_handler(ioc, CAN_SLEEP, | 420 | mpt2sas_base_hard_reset_handler(ioc, CAN_SLEEP, |
324 | FORCE_BIG_HAMMER); | 421 | FORCE_BIG_HAMMER); |
@@ -326,46 +423,43 @@ out: | |||
326 | } | 423 | } |
327 | 424 | ||
328 | /** | 425 | /** |
329 | * _config_alloc_config_dma_memory - obtain physical memory | 426 | * mpt2sas_config_get_manufacturing_pg0 - obtain manufacturing page 0 |
330 | * @ioc: per adapter object | 427 | * @ioc: per adapter object |
331 | * @mem: struct config_request | 428 | * @mpi_reply: reply mf payload returned from firmware |
332 | * | 429 | * @config_page: contents of the config page |
333 | * A wrapper for obtaining dma-able memory for config page request. | 430 | * Context: sleep. |
334 | * | 431 | * |
335 | * Returns 0 for success, non-zero for failure. | 432 | * Returns 0 for success, non-zero for failure. |
336 | */ | 433 | */ |
337 | static int | 434 | int |
338 | _config_alloc_config_dma_memory(struct MPT2SAS_ADAPTER *ioc, | 435 | mpt2sas_config_get_manufacturing_pg0(struct MPT2SAS_ADAPTER *ioc, |
339 | struct config_request *mem) | 436 | Mpi2ConfigReply_t *mpi_reply, Mpi2ManufacturingPage0_t *config_page) |
340 | { | 437 | { |
341 | int r = 0; | 438 | Mpi2ConfigRequest_t mpi_request; |
439 | int r; | ||
342 | 440 | ||
343 | mem->config_page = pci_alloc_consistent(ioc->pdev, mem->config_page_sz, | 441 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
344 | &mem->config_page_dma); | 442 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
345 | if (!mem->config_page) | 443 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
346 | r = -ENOMEM; | 444 | mpi_request.Header.PageType = MPI2_CONFIG_PAGETYPE_MANUFACTURING; |
347 | return r; | 445 | mpi_request.Header.PageNumber = 0; |
348 | } | 446 | mpi_request.Header.PageVersion = MPI2_MANUFACTURING0_PAGEVERSION; |
447 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | ||
448 | r = _config_request(ioc, &mpi_request, mpi_reply, | ||
449 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); | ||
450 | if (r) | ||
451 | goto out; | ||
349 | 452 | ||
350 | /** | 453 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
351 | * _config_free_config_dma_memory - wrapper to free the memory | 454 | r = _config_request(ioc, &mpi_request, mpi_reply, |
352 | * @ioc: per adapter object | 455 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
353 | * @mem: struct config_request | 456 | sizeof(*config_page)); |
354 | * | 457 | out: |
355 | * A wrapper to free dma-able memory when using _config_alloc_config_dma_memory. | 458 | return r; |
356 | * | ||
357 | * Returns 0 for success, non-zero for failure. | ||
358 | */ | ||
359 | static void | ||
360 | _config_free_config_dma_memory(struct MPT2SAS_ADAPTER *ioc, | ||
361 | struct config_request *mem) | ||
362 | { | ||
363 | pci_free_consistent(ioc->pdev, mem->config_page_sz, mem->config_page, | ||
364 | mem->config_page_dma); | ||
365 | } | 459 | } |
366 | 460 | ||
367 | /** | 461 | /** |
368 | * mpt2sas_config_get_manufacturing_pg0 - obtain manufacturing page 0 | 462 | * mpt2sas_config_get_manufacturing_pg10 - obtain manufacturing page 10 |
369 | * @ioc: per adapter object | 463 | * @ioc: per adapter object |
370 | * @mpi_reply: reply mf payload returned from firmware | 464 | * @mpi_reply: reply mf payload returned from firmware |
371 | * @config_page: contents of the config page | 465 | * @config_page: contents of the config page |
@@ -374,53 +468,28 @@ _config_free_config_dma_memory(struct MPT2SAS_ADAPTER *ioc, | |||
374 | * Returns 0 for success, non-zero for failure. | 468 | * Returns 0 for success, non-zero for failure. |
375 | */ | 469 | */ |
376 | int | 470 | int |
377 | mpt2sas_config_get_manufacturing_pg0(struct MPT2SAS_ADAPTER *ioc, | 471 | mpt2sas_config_get_manufacturing_pg10(struct MPT2SAS_ADAPTER *ioc, |
378 | Mpi2ConfigReply_t *mpi_reply, Mpi2ManufacturingPage0_t *config_page) | 472 | Mpi2ConfigReply_t *mpi_reply, Mpi2ManufacturingPage10_t *config_page) |
379 | { | 473 | { |
380 | Mpi2ConfigRequest_t mpi_request; | 474 | Mpi2ConfigRequest_t mpi_request; |
381 | int r; | 475 | int r; |
382 | struct config_request mem; | ||
383 | 476 | ||
384 | memset(config_page, 0, sizeof(Mpi2ManufacturingPage0_t)); | ||
385 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 477 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
386 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 478 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
387 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 479 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
388 | mpi_request.Header.PageType = MPI2_CONFIG_PAGETYPE_MANUFACTURING; | 480 | mpi_request.Header.PageType = MPI2_CONFIG_PAGETYPE_MANUFACTURING; |
389 | mpi_request.Header.PageNumber = 0; | 481 | mpi_request.Header.PageNumber = 10; |
390 | mpi_request.Header.PageVersion = MPI2_MANUFACTURING0_PAGEVERSION; | 482 | mpi_request.Header.PageVersion = MPI2_MANUFACTURING0_PAGEVERSION; |
391 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 483 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
392 | r = _config_request(ioc, &mpi_request, mpi_reply, | 484 | r = _config_request(ioc, &mpi_request, mpi_reply, |
393 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 485 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
394 | if (r) | 486 | if (r) |
395 | goto out; | 487 | goto out; |
396 | 488 | ||
397 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 489 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
398 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
399 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
400 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
401 | mpi_request.Header.PageLength = mpi_reply->Header.PageLength; | ||
402 | mem.config_page_sz = le16_to_cpu(mpi_reply->Header.PageLength) * 4; | ||
403 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
404 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
405 | if (r) | ||
406 | goto out; | ||
407 | } else { | ||
408 | mem.config_page_dma = ioc->config_page_dma; | ||
409 | mem.config_page = ioc->config_page; | ||
410 | } | ||
411 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
412 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
413 | mem.config_page_dma); | ||
414 | r = _config_request(ioc, &mpi_request, mpi_reply, | 490 | r = _config_request(ioc, &mpi_request, mpi_reply, |
415 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 491 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
416 | if (!r) | 492 | sizeof(*config_page)); |
417 | memcpy(config_page, mem.config_page, | ||
418 | min_t(u16, mem.config_page_sz, | ||
419 | sizeof(Mpi2ManufacturingPage0_t))); | ||
420 | |||
421 | if (mem.config_page_sz > ioc->config_page_sz) | ||
422 | _config_free_config_dma_memory(ioc, &mem); | ||
423 | |||
424 | out: | 493 | out: |
425 | return r; | 494 | return r; |
426 | } | 495 | } |
@@ -440,9 +509,7 @@ mpt2sas_config_get_bios_pg2(struct MPT2SAS_ADAPTER *ioc, | |||
440 | { | 509 | { |
441 | Mpi2ConfigRequest_t mpi_request; | 510 | Mpi2ConfigRequest_t mpi_request; |
442 | int r; | 511 | int r; |
443 | struct config_request mem; | ||
444 | 512 | ||
445 | memset(config_page, 0, sizeof(Mpi2BiosPage2_t)); | ||
446 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 513 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
447 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 514 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
448 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 515 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -451,37 +518,14 @@ mpt2sas_config_get_bios_pg2(struct MPT2SAS_ADAPTER *ioc, | |||
451 | mpi_request.Header.PageVersion = MPI2_BIOSPAGE2_PAGEVERSION; | 518 | mpi_request.Header.PageVersion = MPI2_BIOSPAGE2_PAGEVERSION; |
452 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 519 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
453 | r = _config_request(ioc, &mpi_request, mpi_reply, | 520 | r = _config_request(ioc, &mpi_request, mpi_reply, |
454 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 521 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
455 | if (r) | 522 | if (r) |
456 | goto out; | 523 | goto out; |
457 | 524 | ||
458 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 525 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
459 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
460 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
461 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
462 | mpi_request.Header.PageLength = mpi_reply->Header.PageLength; | ||
463 | mem.config_page_sz = le16_to_cpu(mpi_reply->Header.PageLength) * 4; | ||
464 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
465 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
466 | if (r) | ||
467 | goto out; | ||
468 | } else { | ||
469 | mem.config_page_dma = ioc->config_page_dma; | ||
470 | mem.config_page = ioc->config_page; | ||
471 | } | ||
472 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
473 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
474 | mem.config_page_dma); | ||
475 | r = _config_request(ioc, &mpi_request, mpi_reply, | 526 | r = _config_request(ioc, &mpi_request, mpi_reply, |
476 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 527 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
477 | if (!r) | 528 | sizeof(*config_page)); |
478 | memcpy(config_page, mem.config_page, | ||
479 | min_t(u16, mem.config_page_sz, | ||
480 | sizeof(Mpi2BiosPage2_t))); | ||
481 | |||
482 | if (mem.config_page_sz > ioc->config_page_sz) | ||
483 | _config_free_config_dma_memory(ioc, &mem); | ||
484 | |||
485 | out: | 529 | out: |
486 | return r; | 530 | return r; |
487 | } | 531 | } |
@@ -501,9 +545,7 @@ mpt2sas_config_get_bios_pg3(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
501 | { | 545 | { |
502 | Mpi2ConfigRequest_t mpi_request; | 546 | Mpi2ConfigRequest_t mpi_request; |
503 | int r; | 547 | int r; |
504 | struct config_request mem; | ||
505 | 548 | ||
506 | memset(config_page, 0, sizeof(Mpi2BiosPage3_t)); | ||
507 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 549 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
508 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 550 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
509 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 551 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -512,37 +554,14 @@ mpt2sas_config_get_bios_pg3(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
512 | mpi_request.Header.PageVersion = MPI2_BIOSPAGE3_PAGEVERSION; | 554 | mpi_request.Header.PageVersion = MPI2_BIOSPAGE3_PAGEVERSION; |
513 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 555 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
514 | r = _config_request(ioc, &mpi_request, mpi_reply, | 556 | r = _config_request(ioc, &mpi_request, mpi_reply, |
515 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 557 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
516 | if (r) | 558 | if (r) |
517 | goto out; | 559 | goto out; |
518 | 560 | ||
519 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 561 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
520 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
521 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
522 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
523 | mpi_request.Header.PageLength = mpi_reply->Header.PageLength; | ||
524 | mem.config_page_sz = le16_to_cpu(mpi_reply->Header.PageLength) * 4; | ||
525 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
526 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
527 | if (r) | ||
528 | goto out; | ||
529 | } else { | ||
530 | mem.config_page_dma = ioc->config_page_dma; | ||
531 | mem.config_page = ioc->config_page; | ||
532 | } | ||
533 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
534 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
535 | mem.config_page_dma); | ||
536 | r = _config_request(ioc, &mpi_request, mpi_reply, | 562 | r = _config_request(ioc, &mpi_request, mpi_reply, |
537 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 563 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
538 | if (!r) | 564 | sizeof(*config_page)); |
539 | memcpy(config_page, mem.config_page, | ||
540 | min_t(u16, mem.config_page_sz, | ||
541 | sizeof(Mpi2BiosPage3_t))); | ||
542 | |||
543 | if (mem.config_page_sz > ioc->config_page_sz) | ||
544 | _config_free_config_dma_memory(ioc, &mem); | ||
545 | |||
546 | out: | 565 | out: |
547 | return r; | 566 | return r; |
548 | } | 567 | } |
@@ -562,9 +581,7 @@ mpt2sas_config_get_iounit_pg0(struct MPT2SAS_ADAPTER *ioc, | |||
562 | { | 581 | { |
563 | Mpi2ConfigRequest_t mpi_request; | 582 | Mpi2ConfigRequest_t mpi_request; |
564 | int r; | 583 | int r; |
565 | struct config_request mem; | ||
566 | 584 | ||
567 | memset(config_page, 0, sizeof(Mpi2IOUnitPage0_t)); | ||
568 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 585 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
569 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 586 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
570 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 587 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -573,37 +590,14 @@ mpt2sas_config_get_iounit_pg0(struct MPT2SAS_ADAPTER *ioc, | |||
573 | mpi_request.Header.PageVersion = MPI2_IOUNITPAGE0_PAGEVERSION; | 590 | mpi_request.Header.PageVersion = MPI2_IOUNITPAGE0_PAGEVERSION; |
574 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 591 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
575 | r = _config_request(ioc, &mpi_request, mpi_reply, | 592 | r = _config_request(ioc, &mpi_request, mpi_reply, |
576 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 593 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
577 | if (r) | 594 | if (r) |
578 | goto out; | 595 | goto out; |
579 | 596 | ||
580 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 597 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
581 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
582 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
583 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
584 | mpi_request.Header.PageLength = mpi_reply->Header.PageLength; | ||
585 | mem.config_page_sz = le16_to_cpu(mpi_reply->Header.PageLength) * 4; | ||
586 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
587 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
588 | if (r) | ||
589 | goto out; | ||
590 | } else { | ||
591 | mem.config_page_dma = ioc->config_page_dma; | ||
592 | mem.config_page = ioc->config_page; | ||
593 | } | ||
594 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
595 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
596 | mem.config_page_dma); | ||
597 | r = _config_request(ioc, &mpi_request, mpi_reply, | 598 | r = _config_request(ioc, &mpi_request, mpi_reply, |
598 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 599 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
599 | if (!r) | 600 | sizeof(*config_page)); |
600 | memcpy(config_page, mem.config_page, | ||
601 | min_t(u16, mem.config_page_sz, | ||
602 | sizeof(Mpi2IOUnitPage0_t))); | ||
603 | |||
604 | if (mem.config_page_sz > ioc->config_page_sz) | ||
605 | _config_free_config_dma_memory(ioc, &mem); | ||
606 | |||
607 | out: | 601 | out: |
608 | return r; | 602 | return r; |
609 | } | 603 | } |
@@ -623,9 +617,7 @@ mpt2sas_config_get_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, | |||
623 | { | 617 | { |
624 | Mpi2ConfigRequest_t mpi_request; | 618 | Mpi2ConfigRequest_t mpi_request; |
625 | int r; | 619 | int r; |
626 | struct config_request mem; | ||
627 | 620 | ||
628 | memset(config_page, 0, sizeof(Mpi2IOUnitPage1_t)); | ||
629 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 621 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
630 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 622 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
631 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 623 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -634,37 +626,14 @@ mpt2sas_config_get_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, | |||
634 | mpi_request.Header.PageVersion = MPI2_IOUNITPAGE1_PAGEVERSION; | 626 | mpi_request.Header.PageVersion = MPI2_IOUNITPAGE1_PAGEVERSION; |
635 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 627 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
636 | r = _config_request(ioc, &mpi_request, mpi_reply, | 628 | r = _config_request(ioc, &mpi_request, mpi_reply, |
637 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 629 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
638 | if (r) | 630 | if (r) |
639 | goto out; | 631 | goto out; |
640 | 632 | ||
641 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 633 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
642 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
643 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
644 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
645 | mpi_request.Header.PageLength = mpi_reply->Header.PageLength; | ||
646 | mem.config_page_sz = le16_to_cpu(mpi_reply->Header.PageLength) * 4; | ||
647 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
648 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
649 | if (r) | ||
650 | goto out; | ||
651 | } else { | ||
652 | mem.config_page_dma = ioc->config_page_dma; | ||
653 | mem.config_page = ioc->config_page; | ||
654 | } | ||
655 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
656 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
657 | mem.config_page_dma); | ||
658 | r = _config_request(ioc, &mpi_request, mpi_reply, | 634 | r = _config_request(ioc, &mpi_request, mpi_reply, |
659 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 635 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
660 | if (!r) | 636 | sizeof(*config_page)); |
661 | memcpy(config_page, mem.config_page, | ||
662 | min_t(u16, mem.config_page_sz, | ||
663 | sizeof(Mpi2IOUnitPage1_t))); | ||
664 | |||
665 | if (mem.config_page_sz > ioc->config_page_sz) | ||
666 | _config_free_config_dma_memory(ioc, &mem); | ||
667 | |||
668 | out: | 637 | out: |
669 | return r; | 638 | return r; |
670 | } | 639 | } |
@@ -680,11 +649,10 @@ mpt2sas_config_get_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, | |||
680 | */ | 649 | */ |
681 | int | 650 | int |
682 | mpt2sas_config_set_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, | 651 | mpt2sas_config_set_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, |
683 | Mpi2ConfigReply_t *mpi_reply, Mpi2IOUnitPage1_t config_page) | 652 | Mpi2ConfigReply_t *mpi_reply, Mpi2IOUnitPage1_t *config_page) |
684 | { | 653 | { |
685 | Mpi2ConfigRequest_t mpi_request; | 654 | Mpi2ConfigRequest_t mpi_request; |
686 | int r; | 655 | int r; |
687 | struct config_request mem; | ||
688 | 656 | ||
689 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 657 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
690 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 658 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
@@ -694,38 +662,14 @@ mpt2sas_config_set_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, | |||
694 | mpi_request.Header.PageVersion = MPI2_IOUNITPAGE1_PAGEVERSION; | 662 | mpi_request.Header.PageVersion = MPI2_IOUNITPAGE1_PAGEVERSION; |
695 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 663 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
696 | r = _config_request(ioc, &mpi_request, mpi_reply, | 664 | r = _config_request(ioc, &mpi_request, mpi_reply, |
697 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 665 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
698 | if (r) | 666 | if (r) |
699 | goto out; | 667 | goto out; |
700 | 668 | ||
701 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_WRITE_CURRENT; | 669 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_WRITE_CURRENT; |
702 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
703 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
704 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
705 | mpi_request.Header.PageLength = mpi_reply->Header.PageLength; | ||
706 | mem.config_page_sz = le16_to_cpu(mpi_reply->Header.PageLength) * 4; | ||
707 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
708 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
709 | if (r) | ||
710 | goto out; | ||
711 | } else { | ||
712 | mem.config_page_dma = ioc->config_page_dma; | ||
713 | mem.config_page = ioc->config_page; | ||
714 | } | ||
715 | |||
716 | memset(mem.config_page, 0, mem.config_page_sz); | ||
717 | memcpy(mem.config_page, &config_page, | ||
718 | sizeof(Mpi2IOUnitPage1_t)); | ||
719 | |||
720 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
721 | MPT2_CONFIG_COMMON_WRITE_SGLFLAGS | mem.config_page_sz, | ||
722 | mem.config_page_dma); | ||
723 | r = _config_request(ioc, &mpi_request, mpi_reply, | 670 | r = _config_request(ioc, &mpi_request, mpi_reply, |
724 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 671 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
725 | 672 | sizeof(*config_page)); | |
726 | if (mem.config_page_sz > ioc->config_page_sz) | ||
727 | _config_free_config_dma_memory(ioc, &mem); | ||
728 | |||
729 | out: | 673 | out: |
730 | return r; | 674 | return r; |
731 | } | 675 | } |
@@ -745,9 +689,7 @@ mpt2sas_config_get_ioc_pg8(struct MPT2SAS_ADAPTER *ioc, | |||
745 | { | 689 | { |
746 | Mpi2ConfigRequest_t mpi_request; | 690 | Mpi2ConfigRequest_t mpi_request; |
747 | int r; | 691 | int r; |
748 | struct config_request mem; | ||
749 | 692 | ||
750 | memset(config_page, 0, sizeof(Mpi2IOCPage8_t)); | ||
751 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 693 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
752 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 694 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
753 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 695 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -756,37 +698,14 @@ mpt2sas_config_get_ioc_pg8(struct MPT2SAS_ADAPTER *ioc, | |||
756 | mpi_request.Header.PageVersion = MPI2_IOCPAGE8_PAGEVERSION; | 698 | mpi_request.Header.PageVersion = MPI2_IOCPAGE8_PAGEVERSION; |
757 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 699 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
758 | r = _config_request(ioc, &mpi_request, mpi_reply, | 700 | r = _config_request(ioc, &mpi_request, mpi_reply, |
759 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 701 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
760 | if (r) | 702 | if (r) |
761 | goto out; | 703 | goto out; |
762 | 704 | ||
763 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 705 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
764 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
765 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
766 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
767 | mpi_request.Header.PageLength = mpi_reply->Header.PageLength; | ||
768 | mem.config_page_sz = le16_to_cpu(mpi_reply->Header.PageLength) * 4; | ||
769 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
770 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
771 | if (r) | ||
772 | goto out; | ||
773 | } else { | ||
774 | mem.config_page_dma = ioc->config_page_dma; | ||
775 | mem.config_page = ioc->config_page; | ||
776 | } | ||
777 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
778 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
779 | mem.config_page_dma); | ||
780 | r = _config_request(ioc, &mpi_request, mpi_reply, | 706 | r = _config_request(ioc, &mpi_request, mpi_reply, |
781 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 707 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
782 | if (!r) | 708 | sizeof(*config_page)); |
783 | memcpy(config_page, mem.config_page, | ||
784 | min_t(u16, mem.config_page_sz, | ||
785 | sizeof(Mpi2IOCPage8_t))); | ||
786 | |||
787 | if (mem.config_page_sz > ioc->config_page_sz) | ||
788 | _config_free_config_dma_memory(ioc, &mem); | ||
789 | |||
790 | out: | 709 | out: |
791 | return r; | 710 | return r; |
792 | } | 711 | } |
@@ -808,9 +727,7 @@ mpt2sas_config_get_sas_device_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
808 | { | 727 | { |
809 | Mpi2ConfigRequest_t mpi_request; | 728 | Mpi2ConfigRequest_t mpi_request; |
810 | int r; | 729 | int r; |
811 | struct config_request mem; | ||
812 | 730 | ||
813 | memset(config_page, 0, sizeof(Mpi2SasDevicePage0_t)); | ||
814 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 731 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
815 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 732 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
816 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 733 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -820,39 +737,15 @@ mpt2sas_config_get_sas_device_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
820 | mpi_request.Header.PageNumber = 0; | 737 | mpi_request.Header.PageNumber = 0; |
821 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 738 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
822 | r = _config_request(ioc, &mpi_request, mpi_reply, | 739 | r = _config_request(ioc, &mpi_request, mpi_reply, |
823 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 740 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
824 | if (r) | 741 | if (r) |
825 | goto out; | 742 | goto out; |
826 | 743 | ||
827 | mpi_request.PageAddress = cpu_to_le32(form | handle); | 744 | mpi_request.PageAddress = cpu_to_le32(form | handle); |
828 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 745 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
829 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
830 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
831 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
832 | mpi_request.ExtPageLength = mpi_reply->ExtPageLength; | ||
833 | mpi_request.ExtPageType = mpi_reply->ExtPageType; | ||
834 | mem.config_page_sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4; | ||
835 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
836 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
837 | if (r) | ||
838 | goto out; | ||
839 | } else { | ||
840 | mem.config_page_dma = ioc->config_page_dma; | ||
841 | mem.config_page = ioc->config_page; | ||
842 | } | ||
843 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
844 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
845 | mem.config_page_dma); | ||
846 | r = _config_request(ioc, &mpi_request, mpi_reply, | 746 | r = _config_request(ioc, &mpi_request, mpi_reply, |
847 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 747 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
848 | if (!r) | 748 | sizeof(*config_page)); |
849 | memcpy(config_page, mem.config_page, | ||
850 | min_t(u16, mem.config_page_sz, | ||
851 | sizeof(Mpi2SasDevicePage0_t))); | ||
852 | |||
853 | if (mem.config_page_sz > ioc->config_page_sz) | ||
854 | _config_free_config_dma_memory(ioc, &mem); | ||
855 | |||
856 | out: | 749 | out: |
857 | return r; | 750 | return r; |
858 | } | 751 | } |
@@ -874,9 +767,7 @@ mpt2sas_config_get_sas_device_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
874 | { | 767 | { |
875 | Mpi2ConfigRequest_t mpi_request; | 768 | Mpi2ConfigRequest_t mpi_request; |
876 | int r; | 769 | int r; |
877 | struct config_request mem; | ||
878 | 770 | ||
879 | memset(config_page, 0, sizeof(Mpi2SasDevicePage1_t)); | ||
880 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 771 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
881 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 772 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
882 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 773 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -886,39 +777,15 @@ mpt2sas_config_get_sas_device_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
886 | mpi_request.Header.PageNumber = 1; | 777 | mpi_request.Header.PageNumber = 1; |
887 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 778 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
888 | r = _config_request(ioc, &mpi_request, mpi_reply, | 779 | r = _config_request(ioc, &mpi_request, mpi_reply, |
889 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 780 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
890 | if (r) | 781 | if (r) |
891 | goto out; | 782 | goto out; |
892 | 783 | ||
893 | mpi_request.PageAddress = cpu_to_le32(form | handle); | 784 | mpi_request.PageAddress = cpu_to_le32(form | handle); |
894 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 785 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
895 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
896 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
897 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
898 | mpi_request.ExtPageLength = mpi_reply->ExtPageLength; | ||
899 | mpi_request.ExtPageType = mpi_reply->ExtPageType; | ||
900 | mem.config_page_sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4; | ||
901 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
902 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
903 | if (r) | ||
904 | goto out; | ||
905 | } else { | ||
906 | mem.config_page_dma = ioc->config_page_dma; | ||
907 | mem.config_page = ioc->config_page; | ||
908 | } | ||
909 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
910 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
911 | mem.config_page_dma); | ||
912 | r = _config_request(ioc, &mpi_request, mpi_reply, | 786 | r = _config_request(ioc, &mpi_request, mpi_reply, |
913 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 787 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
914 | if (!r) | 788 | sizeof(*config_page)); |
915 | memcpy(config_page, mem.config_page, | ||
916 | min_t(u16, mem.config_page_sz, | ||
917 | sizeof(Mpi2SasDevicePage1_t))); | ||
918 | |||
919 | if (mem.config_page_sz > ioc->config_page_sz) | ||
920 | _config_free_config_dma_memory(ioc, &mem); | ||
921 | |||
922 | out: | 789 | out: |
923 | return r; | 790 | return r; |
924 | } | 791 | } |
@@ -936,11 +803,11 @@ mpt2sas_config_get_number_hba_phys(struct MPT2SAS_ADAPTER *ioc, u8 *num_phys) | |||
936 | { | 803 | { |
937 | Mpi2ConfigRequest_t mpi_request; | 804 | Mpi2ConfigRequest_t mpi_request; |
938 | int r; | 805 | int r; |
939 | struct config_request mem; | ||
940 | u16 ioc_status; | 806 | u16 ioc_status; |
941 | Mpi2ConfigReply_t mpi_reply; | 807 | Mpi2ConfigReply_t mpi_reply; |
942 | Mpi2SasIOUnitPage0_t config_page; | 808 | Mpi2SasIOUnitPage0_t config_page; |
943 | 809 | ||
810 | *num_phys = 0; | ||
944 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 811 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
945 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 812 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
946 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 813 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -950,44 +817,20 @@ mpt2sas_config_get_number_hba_phys(struct MPT2SAS_ADAPTER *ioc, u8 *num_phys) | |||
950 | mpi_request.Header.PageVersion = MPI2_SASIOUNITPAGE0_PAGEVERSION; | 817 | mpi_request.Header.PageVersion = MPI2_SASIOUNITPAGE0_PAGEVERSION; |
951 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 818 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
952 | r = _config_request(ioc, &mpi_request, &mpi_reply, | 819 | r = _config_request(ioc, &mpi_request, &mpi_reply, |
953 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 820 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
954 | if (r) | 821 | if (r) |
955 | goto out; | 822 | goto out; |
956 | 823 | ||
957 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 824 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
958 | mpi_request.Header.PageVersion = mpi_reply.Header.PageVersion; | ||
959 | mpi_request.Header.PageNumber = mpi_reply.Header.PageNumber; | ||
960 | mpi_request.Header.PageType = mpi_reply.Header.PageType; | ||
961 | mpi_request.ExtPageLength = mpi_reply.ExtPageLength; | ||
962 | mpi_request.ExtPageType = mpi_reply.ExtPageType; | ||
963 | mem.config_page_sz = le16_to_cpu(mpi_reply.ExtPageLength) * 4; | ||
964 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
965 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
966 | if (r) | ||
967 | goto out; | ||
968 | } else { | ||
969 | mem.config_page_dma = ioc->config_page_dma; | ||
970 | mem.config_page = ioc->config_page; | ||
971 | } | ||
972 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
973 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
974 | mem.config_page_dma); | ||
975 | r = _config_request(ioc, &mpi_request, &mpi_reply, | 825 | r = _config_request(ioc, &mpi_request, &mpi_reply, |
976 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 826 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, &config_page, |
827 | sizeof(Mpi2SasIOUnitPage0_t)); | ||
977 | if (!r) { | 828 | if (!r) { |
978 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & | 829 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
979 | MPI2_IOCSTATUS_MASK; | 830 | MPI2_IOCSTATUS_MASK; |
980 | if (ioc_status == MPI2_IOCSTATUS_SUCCESS) { | 831 | if (ioc_status == MPI2_IOCSTATUS_SUCCESS) |
981 | memcpy(&config_page, mem.config_page, | ||
982 | min_t(u16, mem.config_page_sz, | ||
983 | sizeof(Mpi2SasIOUnitPage0_t))); | ||
984 | *num_phys = config_page.NumPhys; | 832 | *num_phys = config_page.NumPhys; |
985 | } | ||
986 | } | 833 | } |
987 | |||
988 | if (mem.config_page_sz > ioc->config_page_sz) | ||
989 | _config_free_config_dma_memory(ioc, &mem); | ||
990 | |||
991 | out: | 834 | out: |
992 | return r; | 835 | return r; |
993 | } | 836 | } |
@@ -1011,8 +854,7 @@ mpt2sas_config_get_sas_iounit_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1011 | { | 854 | { |
1012 | Mpi2ConfigRequest_t mpi_request; | 855 | Mpi2ConfigRequest_t mpi_request; |
1013 | int r; | 856 | int r; |
1014 | struct config_request mem; | 857 | |
1015 | memset(config_page, 0, sz); | ||
1016 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 858 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
1017 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 859 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
1018 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 860 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -1022,37 +864,13 @@ mpt2sas_config_get_sas_iounit_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1022 | mpi_request.Header.PageVersion = MPI2_SASIOUNITPAGE0_PAGEVERSION; | 864 | mpi_request.Header.PageVersion = MPI2_SASIOUNITPAGE0_PAGEVERSION; |
1023 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 865 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
1024 | r = _config_request(ioc, &mpi_request, mpi_reply, | 866 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1025 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 867 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
1026 | if (r) | 868 | if (r) |
1027 | goto out; | 869 | goto out; |
1028 | 870 | ||
1029 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 871 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
1030 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
1031 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
1032 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
1033 | mpi_request.ExtPageLength = mpi_reply->ExtPageLength; | ||
1034 | mpi_request.ExtPageType = mpi_reply->ExtPageType; | ||
1035 | mem.config_page_sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4; | ||
1036 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
1037 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
1038 | if (r) | ||
1039 | goto out; | ||
1040 | } else { | ||
1041 | mem.config_page_dma = ioc->config_page_dma; | ||
1042 | mem.config_page = ioc->config_page; | ||
1043 | } | ||
1044 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
1045 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
1046 | mem.config_page_dma); | ||
1047 | r = _config_request(ioc, &mpi_request, mpi_reply, | 872 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1048 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 873 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, sz); |
1049 | if (!r) | ||
1050 | memcpy(config_page, mem.config_page, | ||
1051 | min_t(u16, sz, mem.config_page_sz)); | ||
1052 | |||
1053 | if (mem.config_page_sz > ioc->config_page_sz) | ||
1054 | _config_free_config_dma_memory(ioc, &mem); | ||
1055 | |||
1056 | out: | 874 | out: |
1057 | return r; | 875 | return r; |
1058 | } | 876 | } |
@@ -1076,9 +894,7 @@ mpt2sas_config_get_sas_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1076 | { | 894 | { |
1077 | Mpi2ConfigRequest_t mpi_request; | 895 | Mpi2ConfigRequest_t mpi_request; |
1078 | int r; | 896 | int r; |
1079 | struct config_request mem; | ||
1080 | 897 | ||
1081 | memset(config_page, 0, sz); | ||
1082 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 898 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
1083 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 899 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
1084 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 900 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -1088,37 +904,13 @@ mpt2sas_config_get_sas_iounit_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1088 | mpi_request.Header.PageVersion = MPI2_SASIOUNITPAGE0_PAGEVERSION; | 904 | mpi_request.Header.PageVersion = MPI2_SASIOUNITPAGE0_PAGEVERSION; |
1089 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 905 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
1090 | r = _config_request(ioc, &mpi_request, mpi_reply, | 906 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1091 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 907 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
1092 | if (r) | 908 | if (r) |
1093 | goto out; | 909 | goto out; |
1094 | 910 | ||
1095 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 911 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
1096 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
1097 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
1098 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
1099 | mpi_request.ExtPageLength = mpi_reply->ExtPageLength; | ||
1100 | mpi_request.ExtPageType = mpi_reply->ExtPageType; | ||
1101 | mem.config_page_sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4; | ||
1102 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
1103 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
1104 | if (r) | ||
1105 | goto out; | ||
1106 | } else { | ||
1107 | mem.config_page_dma = ioc->config_page_dma; | ||
1108 | mem.config_page = ioc->config_page; | ||
1109 | } | ||
1110 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
1111 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
1112 | mem.config_page_dma); | ||
1113 | r = _config_request(ioc, &mpi_request, mpi_reply, | 912 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1114 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 913 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, sz); |
1115 | if (!r) | ||
1116 | memcpy(config_page, mem.config_page, | ||
1117 | min_t(u16, sz, mem.config_page_sz)); | ||
1118 | |||
1119 | if (mem.config_page_sz > ioc->config_page_sz) | ||
1120 | _config_free_config_dma_memory(ioc, &mem); | ||
1121 | |||
1122 | out: | 914 | out: |
1123 | return r; | 915 | return r; |
1124 | } | 916 | } |
@@ -1140,9 +932,7 @@ mpt2sas_config_get_expander_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1140 | { | 932 | { |
1141 | Mpi2ConfigRequest_t mpi_request; | 933 | Mpi2ConfigRequest_t mpi_request; |
1142 | int r; | 934 | int r; |
1143 | struct config_request mem; | ||
1144 | 935 | ||
1145 | memset(config_page, 0, sizeof(Mpi2ExpanderPage0_t)); | ||
1146 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 936 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
1147 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 937 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
1148 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 938 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -1152,39 +942,15 @@ mpt2sas_config_get_expander_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1152 | mpi_request.Header.PageVersion = MPI2_SASEXPANDER0_PAGEVERSION; | 942 | mpi_request.Header.PageVersion = MPI2_SASEXPANDER0_PAGEVERSION; |
1153 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 943 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
1154 | r = _config_request(ioc, &mpi_request, mpi_reply, | 944 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1155 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 945 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
1156 | if (r) | 946 | if (r) |
1157 | goto out; | 947 | goto out; |
1158 | 948 | ||
1159 | mpi_request.PageAddress = cpu_to_le32(form | handle); | 949 | mpi_request.PageAddress = cpu_to_le32(form | handle); |
1160 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 950 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
1161 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
1162 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
1163 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
1164 | mpi_request.ExtPageLength = mpi_reply->ExtPageLength; | ||
1165 | mpi_request.ExtPageType = mpi_reply->ExtPageType; | ||
1166 | mem.config_page_sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4; | ||
1167 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
1168 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
1169 | if (r) | ||
1170 | goto out; | ||
1171 | } else { | ||
1172 | mem.config_page_dma = ioc->config_page_dma; | ||
1173 | mem.config_page = ioc->config_page; | ||
1174 | } | ||
1175 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
1176 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
1177 | mem.config_page_dma); | ||
1178 | r = _config_request(ioc, &mpi_request, mpi_reply, | 951 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1179 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 952 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
1180 | if (!r) | 953 | sizeof(*config_page)); |
1181 | memcpy(config_page, mem.config_page, | ||
1182 | min_t(u16, mem.config_page_sz, | ||
1183 | sizeof(Mpi2ExpanderPage0_t))); | ||
1184 | |||
1185 | if (mem.config_page_sz > ioc->config_page_sz) | ||
1186 | _config_free_config_dma_memory(ioc, &mem); | ||
1187 | |||
1188 | out: | 954 | out: |
1189 | return r; | 955 | return r; |
1190 | } | 956 | } |
@@ -1207,9 +973,7 @@ mpt2sas_config_get_expander_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1207 | { | 973 | { |
1208 | Mpi2ConfigRequest_t mpi_request; | 974 | Mpi2ConfigRequest_t mpi_request; |
1209 | int r; | 975 | int r; |
1210 | struct config_request mem; | ||
1211 | 976 | ||
1212 | memset(config_page, 0, sizeof(Mpi2ExpanderPage1_t)); | ||
1213 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 977 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
1214 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 978 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
1215 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 979 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -1219,7 +983,7 @@ mpt2sas_config_get_expander_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1219 | mpi_request.Header.PageVersion = MPI2_SASEXPANDER1_PAGEVERSION; | 983 | mpi_request.Header.PageVersion = MPI2_SASEXPANDER1_PAGEVERSION; |
1220 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 984 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
1221 | r = _config_request(ioc, &mpi_request, mpi_reply, | 985 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1222 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 986 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
1223 | if (r) | 987 | if (r) |
1224 | goto out; | 988 | goto out; |
1225 | 989 | ||
@@ -1227,33 +991,9 @@ mpt2sas_config_get_expander_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1227 | cpu_to_le32(MPI2_SAS_EXPAND_PGAD_FORM_HNDL_PHY_NUM | | 991 | cpu_to_le32(MPI2_SAS_EXPAND_PGAD_FORM_HNDL_PHY_NUM | |
1228 | (phy_number << MPI2_SAS_EXPAND_PGAD_PHYNUM_SHIFT) | handle); | 992 | (phy_number << MPI2_SAS_EXPAND_PGAD_PHYNUM_SHIFT) | handle); |
1229 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 993 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
1230 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
1231 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
1232 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
1233 | mpi_request.ExtPageLength = mpi_reply->ExtPageLength; | ||
1234 | mpi_request.ExtPageType = mpi_reply->ExtPageType; | ||
1235 | mem.config_page_sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4; | ||
1236 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
1237 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
1238 | if (r) | ||
1239 | goto out; | ||
1240 | } else { | ||
1241 | mem.config_page_dma = ioc->config_page_dma; | ||
1242 | mem.config_page = ioc->config_page; | ||
1243 | } | ||
1244 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
1245 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
1246 | mem.config_page_dma); | ||
1247 | r = _config_request(ioc, &mpi_request, mpi_reply, | 994 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1248 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 995 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
1249 | if (!r) | 996 | sizeof(*config_page)); |
1250 | memcpy(config_page, mem.config_page, | ||
1251 | min_t(u16, mem.config_page_sz, | ||
1252 | sizeof(Mpi2ExpanderPage1_t))); | ||
1253 | |||
1254 | if (mem.config_page_sz > ioc->config_page_sz) | ||
1255 | _config_free_config_dma_memory(ioc, &mem); | ||
1256 | |||
1257 | out: | 997 | out: |
1258 | return r; | 998 | return r; |
1259 | } | 999 | } |
@@ -1275,9 +1015,7 @@ mpt2sas_config_get_enclosure_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1275 | { | 1015 | { |
1276 | Mpi2ConfigRequest_t mpi_request; | 1016 | Mpi2ConfigRequest_t mpi_request; |
1277 | int r; | 1017 | int r; |
1278 | struct config_request mem; | ||
1279 | 1018 | ||
1280 | memset(config_page, 0, sizeof(Mpi2SasEnclosurePage0_t)); | ||
1281 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 1019 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
1282 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 1020 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
1283 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 1021 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -1287,39 +1025,15 @@ mpt2sas_config_get_enclosure_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1287 | mpi_request.Header.PageVersion = MPI2_SASENCLOSURE0_PAGEVERSION; | 1025 | mpi_request.Header.PageVersion = MPI2_SASENCLOSURE0_PAGEVERSION; |
1288 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 1026 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
1289 | r = _config_request(ioc, &mpi_request, mpi_reply, | 1027 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1290 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1028 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
1291 | if (r) | 1029 | if (r) |
1292 | goto out; | 1030 | goto out; |
1293 | 1031 | ||
1294 | mpi_request.PageAddress = cpu_to_le32(form | handle); | 1032 | mpi_request.PageAddress = cpu_to_le32(form | handle); |
1295 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 1033 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
1296 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
1297 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
1298 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
1299 | mpi_request.ExtPageLength = mpi_reply->ExtPageLength; | ||
1300 | mpi_request.ExtPageType = mpi_reply->ExtPageType; | ||
1301 | mem.config_page_sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4; | ||
1302 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
1303 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
1304 | if (r) | ||
1305 | goto out; | ||
1306 | } else { | ||
1307 | mem.config_page_dma = ioc->config_page_dma; | ||
1308 | mem.config_page = ioc->config_page; | ||
1309 | } | ||
1310 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
1311 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
1312 | mem.config_page_dma); | ||
1313 | r = _config_request(ioc, &mpi_request, mpi_reply, | 1034 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1314 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1035 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
1315 | if (!r) | 1036 | sizeof(*config_page)); |
1316 | memcpy(config_page, mem.config_page, | ||
1317 | min_t(u16, mem.config_page_sz, | ||
1318 | sizeof(Mpi2SasEnclosurePage0_t))); | ||
1319 | |||
1320 | if (mem.config_page_sz > ioc->config_page_sz) | ||
1321 | _config_free_config_dma_memory(ioc, &mem); | ||
1322 | |||
1323 | out: | 1037 | out: |
1324 | return r; | 1038 | return r; |
1325 | } | 1039 | } |
@@ -1340,9 +1054,7 @@ mpt2sas_config_get_phy_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1340 | { | 1054 | { |
1341 | Mpi2ConfigRequest_t mpi_request; | 1055 | Mpi2ConfigRequest_t mpi_request; |
1342 | int r; | 1056 | int r; |
1343 | struct config_request mem; | ||
1344 | 1057 | ||
1345 | memset(config_page, 0, sizeof(Mpi2SasPhyPage0_t)); | ||
1346 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 1058 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
1347 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 1059 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
1348 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 1060 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -1352,40 +1064,16 @@ mpt2sas_config_get_phy_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1352 | mpi_request.Header.PageVersion = MPI2_SASPHY0_PAGEVERSION; | 1064 | mpi_request.Header.PageVersion = MPI2_SASPHY0_PAGEVERSION; |
1353 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 1065 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
1354 | r = _config_request(ioc, &mpi_request, mpi_reply, | 1066 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1355 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1067 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
1356 | if (r) | 1068 | if (r) |
1357 | goto out; | 1069 | goto out; |
1358 | 1070 | ||
1359 | mpi_request.PageAddress = | 1071 | mpi_request.PageAddress = |
1360 | cpu_to_le32(MPI2_SAS_PHY_PGAD_FORM_PHY_NUMBER | phy_number); | 1072 | cpu_to_le32(MPI2_SAS_PHY_PGAD_FORM_PHY_NUMBER | phy_number); |
1361 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 1073 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
1362 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
1363 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
1364 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
1365 | mpi_request.ExtPageLength = mpi_reply->ExtPageLength; | ||
1366 | mpi_request.ExtPageType = mpi_reply->ExtPageType; | ||
1367 | mem.config_page_sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4; | ||
1368 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
1369 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
1370 | if (r) | ||
1371 | goto out; | ||
1372 | } else { | ||
1373 | mem.config_page_dma = ioc->config_page_dma; | ||
1374 | mem.config_page = ioc->config_page; | ||
1375 | } | ||
1376 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
1377 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
1378 | mem.config_page_dma); | ||
1379 | r = _config_request(ioc, &mpi_request, mpi_reply, | 1074 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1380 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1075 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
1381 | if (!r) | 1076 | sizeof(*config_page)); |
1382 | memcpy(config_page, mem.config_page, | ||
1383 | min_t(u16, mem.config_page_sz, | ||
1384 | sizeof(Mpi2SasPhyPage0_t))); | ||
1385 | |||
1386 | if (mem.config_page_sz > ioc->config_page_sz) | ||
1387 | _config_free_config_dma_memory(ioc, &mem); | ||
1388 | |||
1389 | out: | 1077 | out: |
1390 | return r; | 1078 | return r; |
1391 | } | 1079 | } |
@@ -1406,9 +1094,7 @@ mpt2sas_config_get_phy_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1406 | { | 1094 | { |
1407 | Mpi2ConfigRequest_t mpi_request; | 1095 | Mpi2ConfigRequest_t mpi_request; |
1408 | int r; | 1096 | int r; |
1409 | struct config_request mem; | ||
1410 | 1097 | ||
1411 | memset(config_page, 0, sizeof(Mpi2SasPhyPage1_t)); | ||
1412 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 1098 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
1413 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 1099 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
1414 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 1100 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -1418,40 +1104,16 @@ mpt2sas_config_get_phy_pg1(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1418 | mpi_request.Header.PageVersion = MPI2_SASPHY1_PAGEVERSION; | 1104 | mpi_request.Header.PageVersion = MPI2_SASPHY1_PAGEVERSION; |
1419 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 1105 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
1420 | r = _config_request(ioc, &mpi_request, mpi_reply, | 1106 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1421 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1107 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
1422 | if (r) | 1108 | if (r) |
1423 | goto out; | 1109 | goto out; |
1424 | 1110 | ||
1425 | mpi_request.PageAddress = | 1111 | mpi_request.PageAddress = |
1426 | cpu_to_le32(MPI2_SAS_PHY_PGAD_FORM_PHY_NUMBER | phy_number); | 1112 | cpu_to_le32(MPI2_SAS_PHY_PGAD_FORM_PHY_NUMBER | phy_number); |
1427 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 1113 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
1428 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
1429 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
1430 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
1431 | mpi_request.ExtPageLength = mpi_reply->ExtPageLength; | ||
1432 | mpi_request.ExtPageType = mpi_reply->ExtPageType; | ||
1433 | mem.config_page_sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4; | ||
1434 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
1435 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
1436 | if (r) | ||
1437 | goto out; | ||
1438 | } else { | ||
1439 | mem.config_page_dma = ioc->config_page_dma; | ||
1440 | mem.config_page = ioc->config_page; | ||
1441 | } | ||
1442 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
1443 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
1444 | mem.config_page_dma); | ||
1445 | r = _config_request(ioc, &mpi_request, mpi_reply, | 1114 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1446 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1115 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
1447 | if (!r) | 1116 | sizeof(*config_page)); |
1448 | memcpy(config_page, mem.config_page, | ||
1449 | min_t(u16, mem.config_page_sz, | ||
1450 | sizeof(Mpi2SasPhyPage1_t))); | ||
1451 | |||
1452 | if (mem.config_page_sz > ioc->config_page_sz) | ||
1453 | _config_free_config_dma_memory(ioc, &mem); | ||
1454 | |||
1455 | out: | 1117 | out: |
1456 | return r; | 1118 | return r; |
1457 | } | 1119 | } |
@@ -1474,9 +1136,7 @@ mpt2sas_config_get_raid_volume_pg1(struct MPT2SAS_ADAPTER *ioc, | |||
1474 | { | 1136 | { |
1475 | Mpi2ConfigRequest_t mpi_request; | 1137 | Mpi2ConfigRequest_t mpi_request; |
1476 | int r; | 1138 | int r; |
1477 | struct config_request mem; | ||
1478 | 1139 | ||
1479 | memset(config_page, 0, sizeof(Mpi2RaidVolPage1_t)); | ||
1480 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 1140 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
1481 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 1141 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
1482 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 1142 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
@@ -1485,38 +1145,15 @@ mpt2sas_config_get_raid_volume_pg1(struct MPT2SAS_ADAPTER *ioc, | |||
1485 | mpi_request.Header.PageVersion = MPI2_RAIDVOLPAGE1_PAGEVERSION; | 1145 | mpi_request.Header.PageVersion = MPI2_RAIDVOLPAGE1_PAGEVERSION; |
1486 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 1146 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
1487 | r = _config_request(ioc, &mpi_request, mpi_reply, | 1147 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1488 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1148 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
1489 | if (r) | 1149 | if (r) |
1490 | goto out; | 1150 | goto out; |
1491 | 1151 | ||
1492 | mpi_request.PageAddress = cpu_to_le32(form | handle); | 1152 | mpi_request.PageAddress = cpu_to_le32(form | handle); |
1493 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 1153 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
1494 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
1495 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
1496 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
1497 | mpi_request.Header.PageLength = mpi_reply->Header.PageLength; | ||
1498 | mem.config_page_sz = le16_to_cpu(mpi_reply->Header.PageLength) * 4; | ||
1499 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
1500 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
1501 | if (r) | ||
1502 | goto out; | ||
1503 | } else { | ||
1504 | mem.config_page_dma = ioc->config_page_dma; | ||
1505 | mem.config_page = ioc->config_page; | ||
1506 | } | ||
1507 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
1508 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
1509 | mem.config_page_dma); | ||
1510 | r = _config_request(ioc, &mpi_request, mpi_reply, | 1154 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1511 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1155 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
1512 | if (!r) | 1156 | sizeof(*config_page)); |
1513 | memcpy(config_page, mem.config_page, | ||
1514 | min_t(u16, mem.config_page_sz, | ||
1515 | sizeof(Mpi2RaidVolPage1_t))); | ||
1516 | |||
1517 | if (mem.config_page_sz > ioc->config_page_sz) | ||
1518 | _config_free_config_dma_memory(ioc, &mem); | ||
1519 | |||
1520 | out: | 1157 | out: |
1521 | return r; | 1158 | return r; |
1522 | } | 1159 | } |
@@ -1535,10 +1172,9 @@ mpt2sas_config_get_number_pds(struct MPT2SAS_ADAPTER *ioc, u16 handle, | |||
1535 | u8 *num_pds) | 1172 | u8 *num_pds) |
1536 | { | 1173 | { |
1537 | Mpi2ConfigRequest_t mpi_request; | 1174 | Mpi2ConfigRequest_t mpi_request; |
1538 | Mpi2RaidVolPage0_t *config_page; | 1175 | Mpi2RaidVolPage0_t config_page; |
1539 | Mpi2ConfigReply_t mpi_reply; | 1176 | Mpi2ConfigReply_t mpi_reply; |
1540 | int r; | 1177 | int r; |
1541 | struct config_request mem; | ||
1542 | u16 ioc_status; | 1178 | u16 ioc_status; |
1543 | 1179 | ||
1544 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 1180 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
@@ -1550,43 +1186,23 @@ mpt2sas_config_get_number_pds(struct MPT2SAS_ADAPTER *ioc, u16 handle, | |||
1550 | mpi_request.Header.PageVersion = MPI2_RAIDVOLPAGE0_PAGEVERSION; | 1186 | mpi_request.Header.PageVersion = MPI2_RAIDVOLPAGE0_PAGEVERSION; |
1551 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 1187 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
1552 | r = _config_request(ioc, &mpi_request, &mpi_reply, | 1188 | r = _config_request(ioc, &mpi_request, &mpi_reply, |
1553 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1189 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
1554 | if (r) | 1190 | if (r) |
1555 | goto out; | 1191 | goto out; |
1556 | 1192 | ||
1557 | mpi_request.PageAddress = | 1193 | mpi_request.PageAddress = |
1558 | cpu_to_le32(MPI2_RAID_VOLUME_PGAD_FORM_HANDLE | handle); | 1194 | cpu_to_le32(MPI2_RAID_VOLUME_PGAD_FORM_HANDLE | handle); |
1559 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 1195 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
1560 | mpi_request.Header.PageVersion = mpi_reply.Header.PageVersion; | ||
1561 | mpi_request.Header.PageNumber = mpi_reply.Header.PageNumber; | ||
1562 | mpi_request.Header.PageType = mpi_reply.Header.PageType; | ||
1563 | mpi_request.Header.PageLength = mpi_reply.Header.PageLength; | ||
1564 | mem.config_page_sz = le16_to_cpu(mpi_reply.Header.PageLength) * 4; | ||
1565 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
1566 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
1567 | if (r) | ||
1568 | goto out; | ||
1569 | } else { | ||
1570 | mem.config_page_dma = ioc->config_page_dma; | ||
1571 | mem.config_page = ioc->config_page; | ||
1572 | } | ||
1573 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
1574 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
1575 | mem.config_page_dma); | ||
1576 | r = _config_request(ioc, &mpi_request, &mpi_reply, | 1196 | r = _config_request(ioc, &mpi_request, &mpi_reply, |
1577 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1197 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, &config_page, |
1198 | sizeof(Mpi2RaidVolPage0_t)); | ||
1578 | if (!r) { | 1199 | if (!r) { |
1579 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & | 1200 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
1580 | MPI2_IOCSTATUS_MASK; | 1201 | MPI2_IOCSTATUS_MASK; |
1581 | if (ioc_status == MPI2_IOCSTATUS_SUCCESS) { | 1202 | if (ioc_status == MPI2_IOCSTATUS_SUCCESS) |
1582 | config_page = mem.config_page; | 1203 | *num_pds = config_page.NumPhysDisks; |
1583 | *num_pds = config_page->NumPhysDisks; | ||
1584 | } | ||
1585 | } | 1204 | } |
1586 | 1205 | ||
1587 | if (mem.config_page_sz > ioc->config_page_sz) | ||
1588 | _config_free_config_dma_memory(ioc, &mem); | ||
1589 | |||
1590 | out: | 1206 | out: |
1591 | return r; | 1207 | return r; |
1592 | } | 1208 | } |
@@ -1610,10 +1226,8 @@ mpt2sas_config_get_raid_volume_pg0(struct MPT2SAS_ADAPTER *ioc, | |||
1610 | { | 1226 | { |
1611 | Mpi2ConfigRequest_t mpi_request; | 1227 | Mpi2ConfigRequest_t mpi_request; |
1612 | int r; | 1228 | int r; |
1613 | struct config_request mem; | ||
1614 | 1229 | ||
1615 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 1230 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
1616 | memset(config_page, 0, sz); | ||
1617 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 1231 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
1618 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 1232 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
1619 | mpi_request.Header.PageType = MPI2_CONFIG_PAGETYPE_RAID_VOLUME; | 1233 | mpi_request.Header.PageType = MPI2_CONFIG_PAGETYPE_RAID_VOLUME; |
@@ -1621,37 +1235,14 @@ mpt2sas_config_get_raid_volume_pg0(struct MPT2SAS_ADAPTER *ioc, | |||
1621 | mpi_request.Header.PageVersion = MPI2_RAIDVOLPAGE0_PAGEVERSION; | 1235 | mpi_request.Header.PageVersion = MPI2_RAIDVOLPAGE0_PAGEVERSION; |
1622 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 1236 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
1623 | r = _config_request(ioc, &mpi_request, mpi_reply, | 1237 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1624 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1238 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
1625 | if (r) | 1239 | if (r) |
1626 | goto out; | 1240 | goto out; |
1627 | 1241 | ||
1628 | mpi_request.PageAddress = cpu_to_le32(form | handle); | 1242 | mpi_request.PageAddress = cpu_to_le32(form | handle); |
1629 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 1243 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
1630 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
1631 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
1632 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
1633 | mpi_request.Header.PageLength = mpi_reply->Header.PageLength; | ||
1634 | mem.config_page_sz = le16_to_cpu(mpi_reply->Header.PageLength) * 4; | ||
1635 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
1636 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
1637 | if (r) | ||
1638 | goto out; | ||
1639 | } else { | ||
1640 | mem.config_page_dma = ioc->config_page_dma; | ||
1641 | mem.config_page = ioc->config_page; | ||
1642 | } | ||
1643 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
1644 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
1645 | mem.config_page_dma); | ||
1646 | r = _config_request(ioc, &mpi_request, mpi_reply, | 1244 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1647 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1245 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, sz); |
1648 | if (!r) | ||
1649 | memcpy(config_page, mem.config_page, | ||
1650 | min_t(u16, sz, mem.config_page_sz)); | ||
1651 | |||
1652 | if (mem.config_page_sz > ioc->config_page_sz) | ||
1653 | _config_free_config_dma_memory(ioc, &mem); | ||
1654 | |||
1655 | out: | 1246 | out: |
1656 | return r; | 1247 | return r; |
1657 | } | 1248 | } |
@@ -1674,10 +1265,8 @@ mpt2sas_config_get_phys_disk_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1674 | { | 1265 | { |
1675 | Mpi2ConfigRequest_t mpi_request; | 1266 | Mpi2ConfigRequest_t mpi_request; |
1676 | int r; | 1267 | int r; |
1677 | struct config_request mem; | ||
1678 | 1268 | ||
1679 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); | 1269 | memset(&mpi_request, 0, sizeof(Mpi2ConfigRequest_t)); |
1680 | memset(config_page, 0, sizeof(Mpi2RaidPhysDiskPage0_t)); | ||
1681 | mpi_request.Function = MPI2_FUNCTION_CONFIG; | 1270 | mpi_request.Function = MPI2_FUNCTION_CONFIG; |
1682 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; | 1271 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_HEADER; |
1683 | mpi_request.Header.PageType = MPI2_CONFIG_PAGETYPE_RAID_PHYSDISK; | 1272 | mpi_request.Header.PageType = MPI2_CONFIG_PAGETYPE_RAID_PHYSDISK; |
@@ -1685,38 +1274,15 @@ mpt2sas_config_get_phys_disk_pg0(struct MPT2SAS_ADAPTER *ioc, Mpi2ConfigReply_t | |||
1685 | mpi_request.Header.PageVersion = MPI2_RAIDPHYSDISKPAGE0_PAGEVERSION; | 1274 | mpi_request.Header.PageVersion = MPI2_RAIDPHYSDISKPAGE0_PAGEVERSION; |
1686 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 1275 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
1687 | r = _config_request(ioc, &mpi_request, mpi_reply, | 1276 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1688 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1277 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
1689 | if (r) | 1278 | if (r) |
1690 | goto out; | 1279 | goto out; |
1691 | 1280 | ||
1692 | mpi_request.PageAddress = cpu_to_le32(form | form_specific); | 1281 | mpi_request.PageAddress = cpu_to_le32(form | form_specific); |
1693 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 1282 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
1694 | mpi_request.Header.PageVersion = mpi_reply->Header.PageVersion; | ||
1695 | mpi_request.Header.PageNumber = mpi_reply->Header.PageNumber; | ||
1696 | mpi_request.Header.PageType = mpi_reply->Header.PageType; | ||
1697 | mpi_request.Header.PageLength = mpi_reply->Header.PageLength; | ||
1698 | mem.config_page_sz = le16_to_cpu(mpi_reply->Header.PageLength) * 4; | ||
1699 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
1700 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
1701 | if (r) | ||
1702 | goto out; | ||
1703 | } else { | ||
1704 | mem.config_page_dma = ioc->config_page_dma; | ||
1705 | mem.config_page = ioc->config_page; | ||
1706 | } | ||
1707 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
1708 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
1709 | mem.config_page_dma); | ||
1710 | r = _config_request(ioc, &mpi_request, mpi_reply, | 1283 | r = _config_request(ioc, &mpi_request, mpi_reply, |
1711 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1284 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
1712 | if (!r) | 1285 | sizeof(*config_page)); |
1713 | memcpy(config_page, mem.config_page, | ||
1714 | min_t(u16, mem.config_page_sz, | ||
1715 | sizeof(Mpi2RaidPhysDiskPage0_t))); | ||
1716 | |||
1717 | if (mem.config_page_sz > ioc->config_page_sz) | ||
1718 | _config_free_config_dma_memory(ioc, &mem); | ||
1719 | |||
1720 | out: | 1286 | out: |
1721 | return r; | 1287 | return r; |
1722 | } | 1288 | } |
@@ -1734,11 +1300,10 @@ int | |||
1734 | mpt2sas_config_get_volume_handle(struct MPT2SAS_ADAPTER *ioc, u16 pd_handle, | 1300 | mpt2sas_config_get_volume_handle(struct MPT2SAS_ADAPTER *ioc, u16 pd_handle, |
1735 | u16 *volume_handle) | 1301 | u16 *volume_handle) |
1736 | { | 1302 | { |
1737 | Mpi2RaidConfigurationPage0_t *config_page; | 1303 | Mpi2RaidConfigurationPage0_t *config_page = NULL; |
1738 | Mpi2ConfigRequest_t mpi_request; | 1304 | Mpi2ConfigRequest_t mpi_request; |
1739 | Mpi2ConfigReply_t mpi_reply; | 1305 | Mpi2ConfigReply_t mpi_reply; |
1740 | int r, i; | 1306 | int r, i, config_page_sz; |
1741 | struct config_request mem; | ||
1742 | u16 ioc_status; | 1307 | u16 ioc_status; |
1743 | 1308 | ||
1744 | *volume_handle = 0; | 1309 | *volume_handle = 0; |
@@ -1751,40 +1316,27 @@ mpt2sas_config_get_volume_handle(struct MPT2SAS_ADAPTER *ioc, u16 pd_handle, | |||
1751 | mpi_request.Header.PageNumber = 0; | 1316 | mpi_request.Header.PageNumber = 0; |
1752 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); | 1317 | mpt2sas_base_build_zero_len_sge(ioc, &mpi_request.PageBufferSGE); |
1753 | r = _config_request(ioc, &mpi_request, &mpi_reply, | 1318 | r = _config_request(ioc, &mpi_request, &mpi_reply, |
1754 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1319 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, NULL, 0); |
1755 | if (r) | 1320 | if (r) |
1756 | goto out; | 1321 | goto out; |
1757 | 1322 | ||
1758 | mpi_request.PageAddress = | 1323 | mpi_request.PageAddress = |
1759 | cpu_to_le32(MPI2_RAID_PGAD_FORM_ACTIVE_CONFIG); | 1324 | cpu_to_le32(MPI2_RAID_PGAD_FORM_ACTIVE_CONFIG); |
1760 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; | 1325 | mpi_request.Action = MPI2_CONFIG_ACTION_PAGE_READ_CURRENT; |
1761 | mpi_request.Header.PageVersion = mpi_reply.Header.PageVersion; | 1326 | config_page_sz = (le16_to_cpu(mpi_reply.ExtPageLength) * 4); |
1762 | mpi_request.Header.PageNumber = mpi_reply.Header.PageNumber; | 1327 | config_page = kmalloc(config_page_sz, GFP_KERNEL); |
1763 | mpi_request.Header.PageType = mpi_reply.Header.PageType; | 1328 | if (!config_page) |
1764 | mpi_request.ExtPageLength = mpi_reply.ExtPageLength; | 1329 | goto out; |
1765 | mpi_request.ExtPageType = mpi_reply.ExtPageType; | ||
1766 | mem.config_page_sz = le16_to_cpu(mpi_reply.ExtPageLength) * 4; | ||
1767 | if (mem.config_page_sz > ioc->config_page_sz) { | ||
1768 | r = _config_alloc_config_dma_memory(ioc, &mem); | ||
1769 | if (r) | ||
1770 | goto out; | ||
1771 | } else { | ||
1772 | mem.config_page_dma = ioc->config_page_dma; | ||
1773 | mem.config_page = ioc->config_page; | ||
1774 | } | ||
1775 | ioc->base_add_sg_single(&mpi_request.PageBufferSGE, | ||
1776 | MPT2_CONFIG_COMMON_SGLFLAGS | mem.config_page_sz, | ||
1777 | mem.config_page_dma); | ||
1778 | r = _config_request(ioc, &mpi_request, &mpi_reply, | 1330 | r = _config_request(ioc, &mpi_request, &mpi_reply, |
1779 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT); | 1331 | MPT2_CONFIG_PAGE_DEFAULT_TIMEOUT, config_page, |
1332 | config_page_sz); | ||
1780 | if (r) | 1333 | if (r) |
1781 | goto out; | 1334 | goto out; |
1782 | 1335 | ||
1783 | r = -1; | 1336 | r = -1; |
1784 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; | 1337 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; |
1785 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) | 1338 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) |
1786 | goto done; | 1339 | goto out; |
1787 | config_page = mem.config_page; | ||
1788 | for (i = 0; i < config_page->NumElements; i++) { | 1340 | for (i = 0; i < config_page->NumElements; i++) { |
1789 | if ((config_page->ConfigElement[i].ElementFlags & | 1341 | if ((config_page->ConfigElement[i].ElementFlags & |
1790 | MPI2_RAIDCONFIG0_EFLAGS_MASK_ELEMENT_TYPE) != | 1342 | MPI2_RAIDCONFIG0_EFLAGS_MASK_ELEMENT_TYPE) != |
@@ -1795,15 +1347,11 @@ mpt2sas_config_get_volume_handle(struct MPT2SAS_ADAPTER *ioc, u16 pd_handle, | |||
1795 | *volume_handle = le16_to_cpu(config_page-> | 1347 | *volume_handle = le16_to_cpu(config_page-> |
1796 | ConfigElement[i].VolDevHandle); | 1348 | ConfigElement[i].VolDevHandle); |
1797 | r = 0; | 1349 | r = 0; |
1798 | goto done; | 1350 | goto out; |
1799 | } | 1351 | } |
1800 | } | 1352 | } |
1801 | |||
1802 | done: | ||
1803 | if (mem.config_page_sz > ioc->config_page_sz) | ||
1804 | _config_free_config_dma_memory(ioc, &mem); | ||
1805 | |||
1806 | out: | 1353 | out: |
1354 | kfree(config_page); | ||
1807 | return r; | 1355 | return r; |
1808 | } | 1356 | } |
1809 | 1357 | ||
diff --git a/drivers/scsi/mpt2sas/mpt2sas_ctl.c b/drivers/scsi/mpt2sas/mpt2sas_ctl.c index 14e473d1fa7b..c2a51018910f 100644 --- a/drivers/scsi/mpt2sas/mpt2sas_ctl.c +++ b/drivers/scsi/mpt2sas/mpt2sas_ctl.c | |||
@@ -1963,7 +1963,6 @@ _ctl_ioctl_main(struct file *file, unsigned int cmd, void __user *arg) | |||
1963 | { | 1963 | { |
1964 | enum block_state state; | 1964 | enum block_state state; |
1965 | long ret = -EINVAL; | 1965 | long ret = -EINVAL; |
1966 | unsigned long flags; | ||
1967 | 1966 | ||
1968 | state = (file->f_flags & O_NONBLOCK) ? NON_BLOCKING : | 1967 | state = (file->f_flags & O_NONBLOCK) ? NON_BLOCKING : |
1969 | BLOCKING; | 1968 | BLOCKING; |
@@ -1989,13 +1988,8 @@ _ctl_ioctl_main(struct file *file, unsigned int cmd, void __user *arg) | |||
1989 | !ioc) | 1988 | !ioc) |
1990 | return -ENODEV; | 1989 | return -ENODEV; |
1991 | 1990 | ||
1992 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); | 1991 | if (ioc->shost_recovery) |
1993 | if (ioc->shost_recovery) { | ||
1994 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, | ||
1995 | flags); | ||
1996 | return -EAGAIN; | 1992 | return -EAGAIN; |
1997 | } | ||
1998 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | ||
1999 | 1993 | ||
2000 | if (_IOC_SIZE(cmd) == sizeof(struct mpt2_ioctl_command)) { | 1994 | if (_IOC_SIZE(cmd) == sizeof(struct mpt2_ioctl_command)) { |
2001 | uarg = arg; | 1995 | uarg = arg; |
@@ -2098,7 +2092,6 @@ _ctl_compat_mpt_command(struct file *file, unsigned cmd, unsigned long arg) | |||
2098 | struct mpt2_ioctl_command karg; | 2092 | struct mpt2_ioctl_command karg; |
2099 | struct MPT2SAS_ADAPTER *ioc; | 2093 | struct MPT2SAS_ADAPTER *ioc; |
2100 | enum block_state state; | 2094 | enum block_state state; |
2101 | unsigned long flags; | ||
2102 | 2095 | ||
2103 | if (_IOC_SIZE(cmd) != sizeof(struct mpt2_ioctl_command32)) | 2096 | if (_IOC_SIZE(cmd) != sizeof(struct mpt2_ioctl_command32)) |
2104 | return -EINVAL; | 2097 | return -EINVAL; |
@@ -2113,13 +2106,8 @@ _ctl_compat_mpt_command(struct file *file, unsigned cmd, unsigned long arg) | |||
2113 | if (_ctl_verify_adapter(karg32.hdr.ioc_number, &ioc) == -1 || !ioc) | 2106 | if (_ctl_verify_adapter(karg32.hdr.ioc_number, &ioc) == -1 || !ioc) |
2114 | return -ENODEV; | 2107 | return -ENODEV; |
2115 | 2108 | ||
2116 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); | 2109 | if (ioc->shost_recovery) |
2117 | if (ioc->shost_recovery) { | ||
2118 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, | ||
2119 | flags); | ||
2120 | return -EAGAIN; | 2110 | return -EAGAIN; |
2121 | } | ||
2122 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | ||
2123 | 2111 | ||
2124 | memset(&karg, 0, sizeof(struct mpt2_ioctl_command)); | 2112 | memset(&karg, 0, sizeof(struct mpt2_ioctl_command)); |
2125 | karg.hdr.ioc_number = karg32.hdr.ioc_number; | 2113 | karg.hdr.ioc_number = karg32.hdr.ioc_number; |
diff --git a/drivers/scsi/mpt2sas/mpt2sas_scsih.c b/drivers/scsi/mpt2sas/mpt2sas_scsih.c index 2e9a4445596f..774b34525bba 100644 --- a/drivers/scsi/mpt2sas/mpt2sas_scsih.c +++ b/drivers/scsi/mpt2sas/mpt2sas_scsih.c | |||
@@ -103,7 +103,6 @@ struct sense_info { | |||
103 | }; | 103 | }; |
104 | 104 | ||
105 | 105 | ||
106 | #define MPT2SAS_RESCAN_AFTER_HOST_RESET (0xFFFF) | ||
107 | /** | 106 | /** |
108 | * struct fw_event_work - firmware event struct | 107 | * struct fw_event_work - firmware event struct |
109 | * @list: link list framework | 108 | * @list: link list framework |
@@ -1502,7 +1501,13 @@ _scsih_slave_configure(struct scsi_device *sdev) | |||
1502 | break; | 1501 | break; |
1503 | case MPI2_RAID_VOL_TYPE_RAID1E: | 1502 | case MPI2_RAID_VOL_TYPE_RAID1E: |
1504 | qdepth = MPT2SAS_RAID_QUEUE_DEPTH; | 1503 | qdepth = MPT2SAS_RAID_QUEUE_DEPTH; |
1505 | r_level = "RAID1E"; | 1504 | if (ioc->manu_pg10.OEMIdentifier && |
1505 | (ioc->manu_pg10.GenericFlags0 & | ||
1506 | MFG10_GF0_R10_DISPLAY) && | ||
1507 | !(raid_device->num_pds % 2)) | ||
1508 | r_level = "RAID10"; | ||
1509 | else | ||
1510 | r_level = "RAID1E"; | ||
1506 | break; | 1511 | break; |
1507 | case MPI2_RAID_VOL_TYPE_RAID1: | 1512 | case MPI2_RAID_VOL_TYPE_RAID1: |
1508 | qdepth = MPT2SAS_RAID_QUEUE_DEPTH; | 1513 | qdepth = MPT2SAS_RAID_QUEUE_DEPTH; |
@@ -1786,17 +1791,18 @@ mpt2sas_scsih_issue_tm(struct MPT2SAS_ADAPTER *ioc, u16 handle, uint lun, | |||
1786 | u32 ioc_state; | 1791 | u32 ioc_state; |
1787 | unsigned long timeleft; | 1792 | unsigned long timeleft; |
1788 | u8 VF_ID = 0; | 1793 | u8 VF_ID = 0; |
1789 | unsigned long flags; | ||
1790 | 1794 | ||
1791 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); | 1795 | if (ioc->tm_cmds.status != MPT2_CMD_NOT_USED) { |
1792 | if (ioc->tm_cmds.status != MPT2_CMD_NOT_USED || | 1796 | printk(MPT2SAS_INFO_FMT "%s: tm_cmd busy!!!\n", |
1793 | ioc->shost_recovery) { | 1797 | __func__, ioc->name); |
1794 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | 1798 | return; |
1799 | } | ||
1800 | |||
1801 | if (ioc->shost_recovery) { | ||
1795 | printk(MPT2SAS_INFO_FMT "%s: host reset in progress!\n", | 1802 | printk(MPT2SAS_INFO_FMT "%s: host reset in progress!\n", |
1796 | __func__, ioc->name); | 1803 | __func__, ioc->name); |
1797 | return; | 1804 | return; |
1798 | } | 1805 | } |
1799 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | ||
1800 | 1806 | ||
1801 | ioc_state = mpt2sas_base_get_iocstate(ioc, 0); | 1807 | ioc_state = mpt2sas_base_get_iocstate(ioc, 0); |
1802 | if (ioc_state & MPI2_DOORBELL_USED) { | 1808 | if (ioc_state & MPI2_DOORBELL_USED) { |
@@ -1830,6 +1836,7 @@ mpt2sas_scsih_issue_tm(struct MPT2SAS_ADAPTER *ioc, u16 handle, uint lun, | |||
1830 | mpi_request->TaskMID = cpu_to_le16(smid_task); | 1836 | mpi_request->TaskMID = cpu_to_le16(smid_task); |
1831 | int_to_scsilun(lun, (struct scsi_lun *)mpi_request->LUN); | 1837 | int_to_scsilun(lun, (struct scsi_lun *)mpi_request->LUN); |
1832 | mpt2sas_scsih_set_tm_flag(ioc, handle); | 1838 | mpt2sas_scsih_set_tm_flag(ioc, handle); |
1839 | init_completion(&ioc->tm_cmds.done); | ||
1833 | mpt2sas_base_put_smid_hi_priority(ioc, smid, VF_ID); | 1840 | mpt2sas_base_put_smid_hi_priority(ioc, smid, VF_ID); |
1834 | timeleft = wait_for_completion_timeout(&ioc->tm_cmds.done, timeout*HZ); | 1841 | timeleft = wait_for_completion_timeout(&ioc->tm_cmds.done, timeout*HZ); |
1835 | mpt2sas_scsih_clear_tm_flag(ioc, handle); | 1842 | mpt2sas_scsih_clear_tm_flag(ioc, handle); |
@@ -2222,7 +2229,7 @@ _scsih_ublock_io_device(struct MPT2SAS_ADAPTER *ioc, u16 handle) | |||
2222 | MPT2SAS_INFO_FMT "SDEV_RUNNING: " | 2229 | MPT2SAS_INFO_FMT "SDEV_RUNNING: " |
2223 | "handle(0x%04x)\n", ioc->name, handle)); | 2230 | "handle(0x%04x)\n", ioc->name, handle)); |
2224 | sas_device_priv_data->block = 0; | 2231 | sas_device_priv_data->block = 0; |
2225 | scsi_device_set_state(sdev, SDEV_RUNNING); | 2232 | scsi_internal_device_unblock(sdev); |
2226 | } | 2233 | } |
2227 | } | 2234 | } |
2228 | } | 2235 | } |
@@ -2251,7 +2258,7 @@ _scsih_block_io_device(struct MPT2SAS_ADAPTER *ioc, u16 handle) | |||
2251 | MPT2SAS_INFO_FMT "SDEV_BLOCK: " | 2258 | MPT2SAS_INFO_FMT "SDEV_BLOCK: " |
2252 | "handle(0x%04x)\n", ioc->name, handle)); | 2259 | "handle(0x%04x)\n", ioc->name, handle)); |
2253 | sas_device_priv_data->block = 1; | 2260 | sas_device_priv_data->block = 1; |
2254 | scsi_device_set_state(sdev, SDEV_BLOCK); | 2261 | scsi_internal_device_block(sdev); |
2255 | } | 2262 | } |
2256 | } | 2263 | } |
2257 | } | 2264 | } |
@@ -2327,6 +2334,7 @@ _scsih_block_io_to_children_attached_directly(struct MPT2SAS_ADAPTER *ioc, | |||
2327 | u16 handle; | 2334 | u16 handle; |
2328 | u16 reason_code; | 2335 | u16 reason_code; |
2329 | u8 phy_number; | 2336 | u8 phy_number; |
2337 | u8 link_rate; | ||
2330 | 2338 | ||
2331 | for (i = 0; i < event_data->NumEntries; i++) { | 2339 | for (i = 0; i < event_data->NumEntries; i++) { |
2332 | handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); | 2340 | handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); |
@@ -2337,6 +2345,11 @@ _scsih_block_io_to_children_attached_directly(struct MPT2SAS_ADAPTER *ioc, | |||
2337 | MPI2_EVENT_SAS_TOPO_RC_MASK; | 2345 | MPI2_EVENT_SAS_TOPO_RC_MASK; |
2338 | if (reason_code == MPI2_EVENT_SAS_TOPO_RC_DELAY_NOT_RESPONDING) | 2346 | if (reason_code == MPI2_EVENT_SAS_TOPO_RC_DELAY_NOT_RESPONDING) |
2339 | _scsih_block_io_device(ioc, handle); | 2347 | _scsih_block_io_device(ioc, handle); |
2348 | if (reason_code == MPI2_EVENT_SAS_TOPO_RC_PHY_CHANGED) { | ||
2349 | link_rate = event_data->PHY[i].LinkRate >> 4; | ||
2350 | if (link_rate >= MPI2_SAS_NEG_LINK_RATE_1_5) | ||
2351 | _scsih_ublock_io_device(ioc, handle); | ||
2352 | } | ||
2340 | } | 2353 | } |
2341 | } | 2354 | } |
2342 | 2355 | ||
@@ -2405,27 +2418,6 @@ _scsih_check_topo_delete_events(struct MPT2SAS_ADAPTER *ioc, | |||
2405 | } | 2418 | } |
2406 | 2419 | ||
2407 | /** | 2420 | /** |
2408 | * _scsih_queue_rescan - queue a topology rescan from user context | ||
2409 | * @ioc: per adapter object | ||
2410 | * | ||
2411 | * Return nothing. | ||
2412 | */ | ||
2413 | static void | ||
2414 | _scsih_queue_rescan(struct MPT2SAS_ADAPTER *ioc) | ||
2415 | { | ||
2416 | struct fw_event_work *fw_event; | ||
2417 | |||
2418 | if (ioc->wait_for_port_enable_to_complete) | ||
2419 | return; | ||
2420 | fw_event = kzalloc(sizeof(struct fw_event_work), GFP_ATOMIC); | ||
2421 | if (!fw_event) | ||
2422 | return; | ||
2423 | fw_event->event = MPT2SAS_RESCAN_AFTER_HOST_RESET; | ||
2424 | fw_event->ioc = ioc; | ||
2425 | _scsih_fw_event_add(ioc, fw_event); | ||
2426 | } | ||
2427 | |||
2428 | /** | ||
2429 | * _scsih_flush_running_cmds - completing outstanding commands. | 2421 | * _scsih_flush_running_cmds - completing outstanding commands. |
2430 | * @ioc: per adapter object | 2422 | * @ioc: per adapter object |
2431 | * | 2423 | * |
@@ -2456,46 +2448,6 @@ _scsih_flush_running_cmds(struct MPT2SAS_ADAPTER *ioc) | |||
2456 | } | 2448 | } |
2457 | 2449 | ||
2458 | /** | 2450 | /** |
2459 | * mpt2sas_scsih_reset_handler - reset callback handler (for scsih) | ||
2460 | * @ioc: per adapter object | ||
2461 | * @reset_phase: phase | ||
2462 | * | ||
2463 | * The handler for doing any required cleanup or initialization. | ||
2464 | * | ||
2465 | * The reset phase can be MPT2_IOC_PRE_RESET, MPT2_IOC_AFTER_RESET, | ||
2466 | * MPT2_IOC_DONE_RESET | ||
2467 | * | ||
2468 | * Return nothing. | ||
2469 | */ | ||
2470 | void | ||
2471 | mpt2sas_scsih_reset_handler(struct MPT2SAS_ADAPTER *ioc, int reset_phase) | ||
2472 | { | ||
2473 | switch (reset_phase) { | ||
2474 | case MPT2_IOC_PRE_RESET: | ||
2475 | dtmprintk(ioc, printk(MPT2SAS_DEBUG_FMT "%s: " | ||
2476 | "MPT2_IOC_PRE_RESET\n", ioc->name, __func__)); | ||
2477 | _scsih_fw_event_off(ioc); | ||
2478 | break; | ||
2479 | case MPT2_IOC_AFTER_RESET: | ||
2480 | dtmprintk(ioc, printk(MPT2SAS_DEBUG_FMT "%s: " | ||
2481 | "MPT2_IOC_AFTER_RESET\n", ioc->name, __func__)); | ||
2482 | if (ioc->tm_cmds.status & MPT2_CMD_PENDING) { | ||
2483 | ioc->tm_cmds.status |= MPT2_CMD_RESET; | ||
2484 | mpt2sas_base_free_smid(ioc, ioc->tm_cmds.smid); | ||
2485 | complete(&ioc->tm_cmds.done); | ||
2486 | } | ||
2487 | _scsih_fw_event_on(ioc); | ||
2488 | _scsih_flush_running_cmds(ioc); | ||
2489 | break; | ||
2490 | case MPT2_IOC_DONE_RESET: | ||
2491 | dtmprintk(ioc, printk(MPT2SAS_DEBUG_FMT "%s: " | ||
2492 | "MPT2_IOC_DONE_RESET\n", ioc->name, __func__)); | ||
2493 | _scsih_queue_rescan(ioc); | ||
2494 | break; | ||
2495 | } | ||
2496 | } | ||
2497 | |||
2498 | /** | ||
2499 | * _scsih_setup_eedp - setup MPI request for EEDP transfer | 2451 | * _scsih_setup_eedp - setup MPI request for EEDP transfer |
2500 | * @scmd: pointer to scsi command object | 2452 | * @scmd: pointer to scsi command object |
2501 | * @mpi_request: pointer to the SCSI_IO reqest message frame | 2453 | * @mpi_request: pointer to the SCSI_IO reqest message frame |
@@ -2615,7 +2567,6 @@ _scsih_qcmd(struct scsi_cmnd *scmd, void (*done)(struct scsi_cmnd *)) | |||
2615 | Mpi2SCSIIORequest_t *mpi_request; | 2567 | Mpi2SCSIIORequest_t *mpi_request; |
2616 | u32 mpi_control; | 2568 | u32 mpi_control; |
2617 | u16 smid; | 2569 | u16 smid; |
2618 | unsigned long flags; | ||
2619 | 2570 | ||
2620 | scmd->scsi_done = done; | 2571 | scmd->scsi_done = done; |
2621 | sas_device_priv_data = scmd->device->hostdata; | 2572 | sas_device_priv_data = scmd->device->hostdata; |
@@ -2634,13 +2585,10 @@ _scsih_qcmd(struct scsi_cmnd *scmd, void (*done)(struct scsi_cmnd *)) | |||
2634 | } | 2585 | } |
2635 | 2586 | ||
2636 | /* see if we are busy with task managment stuff */ | 2587 | /* see if we are busy with task managment stuff */ |
2637 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); | 2588 | if (sas_target_priv_data->tm_busy) |
2638 | if (sas_target_priv_data->tm_busy || | 2589 | return SCSI_MLQUEUE_DEVICE_BUSY; |
2639 | ioc->shost_recovery || ioc->ioc_link_reset_in_progress) { | 2590 | else if (ioc->shost_recovery || ioc->ioc_link_reset_in_progress) |
2640 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | ||
2641 | return SCSI_MLQUEUE_HOST_BUSY; | 2591 | return SCSI_MLQUEUE_HOST_BUSY; |
2642 | } | ||
2643 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | ||
2644 | 2592 | ||
2645 | if (scmd->sc_data_direction == DMA_FROM_DEVICE) | 2593 | if (scmd->sc_data_direction == DMA_FROM_DEVICE) |
2646 | mpi_control = MPI2_SCSIIO_CONTROL_READ; | 2594 | mpi_control = MPI2_SCSIIO_CONTROL_READ; |
@@ -3189,25 +3137,6 @@ _scsih_io_done(struct MPT2SAS_ADAPTER *ioc, u16 smid, u8 VF_ID, u32 reply) | |||
3189 | } | 3137 | } |
3190 | 3138 | ||
3191 | /** | 3139 | /** |
3192 | * _scsih_link_change - process phy link changes | ||
3193 | * @ioc: per adapter object | ||
3194 | * @handle: phy handle | ||
3195 | * @attached_handle: valid for devices attached to link | ||
3196 | * @phy_number: phy number | ||
3197 | * @link_rate: new link rate | ||
3198 | * Context: user. | ||
3199 | * | ||
3200 | * Return nothing. | ||
3201 | */ | ||
3202 | static void | ||
3203 | _scsih_link_change(struct MPT2SAS_ADAPTER *ioc, u16 handle, u16 attached_handle, | ||
3204 | u8 phy_number, u8 link_rate) | ||
3205 | { | ||
3206 | mpt2sas_transport_update_phy_link_change(ioc, handle, attached_handle, | ||
3207 | phy_number, link_rate); | ||
3208 | } | ||
3209 | |||
3210 | /** | ||
3211 | * _scsih_sas_host_refresh - refreshing sas host object contents | 3140 | * _scsih_sas_host_refresh - refreshing sas host object contents |
3212 | * @ioc: per adapter object | 3141 | * @ioc: per adapter object |
3213 | * @update: update link information | 3142 | * @update: update link information |
@@ -3251,7 +3180,8 @@ _scsih_sas_host_refresh(struct MPT2SAS_ADAPTER *ioc, u8 update) | |||
3251 | le16_to_cpu(sas_iounit_pg0->PhyData[i]. | 3180 | le16_to_cpu(sas_iounit_pg0->PhyData[i]. |
3252 | ControllerDevHandle); | 3181 | ControllerDevHandle); |
3253 | if (update) | 3182 | if (update) |
3254 | _scsih_link_change(ioc, | 3183 | mpt2sas_transport_update_links( |
3184 | ioc, | ||
3255 | ioc->sas_hba.phy[i].handle, | 3185 | ioc->sas_hba.phy[i].handle, |
3256 | le16_to_cpu(sas_iounit_pg0->PhyData[i]. | 3186 | le16_to_cpu(sas_iounit_pg0->PhyData[i]. |
3257 | AttachedDevHandle), i, | 3187 | AttachedDevHandle), i, |
@@ -3436,6 +3366,9 @@ _scsih_expander_add(struct MPT2SAS_ADAPTER *ioc, u16 handle) | |||
3436 | if (!handle) | 3366 | if (!handle) |
3437 | return -1; | 3367 | return -1; |
3438 | 3368 | ||
3369 | if (ioc->shost_recovery) | ||
3370 | return -1; | ||
3371 | |||
3439 | if ((mpt2sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0, | 3372 | if ((mpt2sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0, |
3440 | MPI2_SAS_EXPAND_PGAD_FORM_HNDL, handle))) { | 3373 | MPI2_SAS_EXPAND_PGAD_FORM_HNDL, handle))) { |
3441 | printk(MPT2SAS_ERR_FMT "failure at %s:%d/%s()!\n", | 3374 | printk(MPT2SAS_ERR_FMT "failure at %s:%d/%s()!\n", |
@@ -3572,6 +3505,9 @@ _scsih_expander_remove(struct MPT2SAS_ADAPTER *ioc, u16 handle) | |||
3572 | struct _sas_node *sas_expander; | 3505 | struct _sas_node *sas_expander; |
3573 | unsigned long flags; | 3506 | unsigned long flags; |
3574 | 3507 | ||
3508 | if (ioc->shost_recovery) | ||
3509 | return; | ||
3510 | |||
3575 | spin_lock_irqsave(&ioc->sas_node_lock, flags); | 3511 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
3576 | sas_expander = mpt2sas_scsih_expander_find_by_handle(ioc, handle); | 3512 | sas_expander = mpt2sas_scsih_expander_find_by_handle(ioc, handle); |
3577 | spin_unlock_irqrestore(&ioc->sas_node_lock, flags); | 3513 | spin_unlock_irqrestore(&ioc->sas_node_lock, flags); |
@@ -3743,6 +3679,8 @@ _scsih_remove_device(struct MPT2SAS_ADAPTER *ioc, u16 handle) | |||
3743 | mutex_unlock(&ioc->tm_cmds.mutex); | 3679 | mutex_unlock(&ioc->tm_cmds.mutex); |
3744 | dewtprintk(ioc, printk(MPT2SAS_INFO_FMT "issue target reset " | 3680 | dewtprintk(ioc, printk(MPT2SAS_INFO_FMT "issue target reset " |
3745 | "done: handle(0x%04x)\n", ioc->name, device_handle)); | 3681 | "done: handle(0x%04x)\n", ioc->name, device_handle)); |
3682 | if (ioc->shost_recovery) | ||
3683 | goto out; | ||
3746 | } | 3684 | } |
3747 | 3685 | ||
3748 | /* SAS_IO_UNIT_CNTR - send REMOVE_DEVICE */ | 3686 | /* SAS_IO_UNIT_CNTR - send REMOVE_DEVICE */ |
@@ -3765,6 +3703,9 @@ _scsih_remove_device(struct MPT2SAS_ADAPTER *ioc, u16 handle) | |||
3765 | le32_to_cpu(mpi_reply.IOCLogInfo))); | 3703 | le32_to_cpu(mpi_reply.IOCLogInfo))); |
3766 | 3704 | ||
3767 | out: | 3705 | out: |
3706 | |||
3707 | _scsih_ublock_io_device(ioc, handle); | ||
3708 | |||
3768 | mpt2sas_transport_port_remove(ioc, sas_device->sas_address, | 3709 | mpt2sas_transport_port_remove(ioc, sas_device->sas_address, |
3769 | sas_device->parent_handle); | 3710 | sas_device->parent_handle); |
3770 | 3711 | ||
@@ -3908,6 +3849,8 @@ _scsih_sas_topology_change_event(struct MPT2SAS_ADAPTER *ioc, u8 VF_ID, | |||
3908 | "expander event\n", ioc->name)); | 3849 | "expander event\n", ioc->name)); |
3909 | return; | 3850 | return; |
3910 | } | 3851 | } |
3852 | if (ioc->shost_recovery) | ||
3853 | return; | ||
3911 | if (event_data->PHY[i].PhyStatus & | 3854 | if (event_data->PHY[i].PhyStatus & |
3912 | MPI2_EVENT_SAS_TOPO_PHYSTATUS_VACANT) | 3855 | MPI2_EVENT_SAS_TOPO_PHYSTATUS_VACANT) |
3913 | continue; | 3856 | continue; |
@@ -3923,9 +3866,10 @@ _scsih_sas_topology_change_event(struct MPT2SAS_ADAPTER *ioc, u8 VF_ID, | |||
3923 | case MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED: | 3866 | case MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED: |
3924 | if (!parent_handle) { | 3867 | if (!parent_handle) { |
3925 | if (phy_number < ioc->sas_hba.num_phys) | 3868 | if (phy_number < ioc->sas_hba.num_phys) |
3926 | _scsih_link_change(ioc, | 3869 | mpt2sas_transport_update_links( |
3927 | ioc->sas_hba.phy[phy_number].handle, | 3870 | ioc, |
3928 | handle, phy_number, link_rate_); | 3871 | ioc->sas_hba.phy[phy_number].handle, |
3872 | handle, phy_number, link_rate_); | ||
3929 | } else { | 3873 | } else { |
3930 | spin_lock_irqsave(&ioc->sas_node_lock, flags); | 3874 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
3931 | sas_expander = | 3875 | sas_expander = |
@@ -3935,17 +3879,14 @@ _scsih_sas_topology_change_event(struct MPT2SAS_ADAPTER *ioc, u8 VF_ID, | |||
3935 | flags); | 3879 | flags); |
3936 | if (sas_expander) { | 3880 | if (sas_expander) { |
3937 | if (phy_number < sas_expander->num_phys) | 3881 | if (phy_number < sas_expander->num_phys) |
3938 | _scsih_link_change(ioc, | 3882 | mpt2sas_transport_update_links( |
3939 | sas_expander-> | 3883 | ioc, |
3940 | phy[phy_number].handle, | 3884 | sas_expander-> |
3941 | handle, phy_number, | 3885 | phy[phy_number].handle, |
3942 | link_rate_); | 3886 | handle, phy_number, |
3887 | link_rate_); | ||
3943 | } | 3888 | } |
3944 | } | 3889 | } |
3945 | if (reason_code == MPI2_EVENT_SAS_TOPO_RC_PHY_CHANGED) { | ||
3946 | if (link_rate_ >= MPI2_SAS_NEG_LINK_RATE_1_5) | ||
3947 | _scsih_ublock_io_device(ioc, handle); | ||
3948 | } | ||
3949 | if (reason_code == MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED) { | 3890 | if (reason_code == MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED) { |
3950 | if (link_rate_ < MPI2_SAS_NEG_LINK_RATE_1_5) | 3891 | if (link_rate_ < MPI2_SAS_NEG_LINK_RATE_1_5) |
3951 | break; | 3892 | break; |
@@ -4455,7 +4396,7 @@ _scsih_sas_pd_add(struct MPT2SAS_ADAPTER *ioc, | |||
4455 | return; | 4396 | return; |
4456 | } | 4397 | } |
4457 | 4398 | ||
4458 | _scsih_link_change(ioc, | 4399 | mpt2sas_transport_update_links(ioc, |
4459 | le16_to_cpu(sas_device_pg0.ParentDevHandle), | 4400 | le16_to_cpu(sas_device_pg0.ParentDevHandle), |
4460 | handle, sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5); | 4401 | handle, sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5); |
4461 | 4402 | ||
@@ -4744,7 +4685,7 @@ _scsih_sas_ir_physical_disk_event(struct MPT2SAS_ADAPTER *ioc, u8 VF_ID, | |||
4744 | return; | 4685 | return; |
4745 | } | 4686 | } |
4746 | 4687 | ||
4747 | _scsih_link_change(ioc, | 4688 | mpt2sas_transport_update_links(ioc, |
4748 | le16_to_cpu(sas_device_pg0.ParentDevHandle), | 4689 | le16_to_cpu(sas_device_pg0.ParentDevHandle), |
4749 | handle, sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5); | 4690 | handle, sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5); |
4750 | 4691 | ||
@@ -5156,22 +5097,9 @@ static void | |||
5156 | _scsih_remove_unresponding_devices(struct MPT2SAS_ADAPTER *ioc) | 5097 | _scsih_remove_unresponding_devices(struct MPT2SAS_ADAPTER *ioc) |
5157 | { | 5098 | { |
5158 | struct _sas_device *sas_device, *sas_device_next; | 5099 | struct _sas_device *sas_device, *sas_device_next; |
5159 | struct _sas_node *sas_expander, *sas_expander_next; | 5100 | struct _sas_node *sas_expander; |
5160 | struct _raid_device *raid_device, *raid_device_next; | 5101 | struct _raid_device *raid_device, *raid_device_next; |
5161 | unsigned long flags; | ||
5162 | 5102 | ||
5163 | _scsih_search_responding_sas_devices(ioc); | ||
5164 | _scsih_search_responding_raid_devices(ioc); | ||
5165 | _scsih_search_responding_expanders(ioc); | ||
5166 | |||
5167 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); | ||
5168 | ioc->shost_recovery = 0; | ||
5169 | if (ioc->shost->shost_state == SHOST_RECOVERY) { | ||
5170 | printk(MPT2SAS_INFO_FMT "putting controller into " | ||
5171 | "SHOST_RUNNING\n", ioc->name); | ||
5172 | scsi_host_set_state(ioc->shost, SHOST_RUNNING); | ||
5173 | } | ||
5174 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | ||
5175 | 5103 | ||
5176 | list_for_each_entry_safe(sas_device, sas_device_next, | 5104 | list_for_each_entry_safe(sas_device, sas_device_next, |
5177 | &ioc->sas_device_list, list) { | 5105 | &ioc->sas_device_list, list) { |
@@ -5207,16 +5135,63 @@ _scsih_remove_unresponding_devices(struct MPT2SAS_ADAPTER *ioc) | |||
5207 | _scsih_raid_device_remove(ioc, raid_device); | 5135 | _scsih_raid_device_remove(ioc, raid_device); |
5208 | } | 5136 | } |
5209 | 5137 | ||
5210 | list_for_each_entry_safe(sas_expander, sas_expander_next, | 5138 | retry_expander_search: |
5211 | &ioc->sas_expander_list, list) { | 5139 | sas_expander = NULL; |
5140 | list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) { | ||
5212 | if (sas_expander->responding) { | 5141 | if (sas_expander->responding) { |
5213 | sas_expander->responding = 0; | 5142 | sas_expander->responding = 0; |
5214 | continue; | 5143 | continue; |
5215 | } | 5144 | } |
5216 | printk("\tremoving expander: handle(0x%04x), " | ||
5217 | " sas_addr(0x%016llx)\n", sas_expander->handle, | ||
5218 | (unsigned long long)sas_expander->sas_address); | ||
5219 | _scsih_expander_remove(ioc, sas_expander->handle); | 5145 | _scsih_expander_remove(ioc, sas_expander->handle); |
5146 | goto retry_expander_search; | ||
5147 | } | ||
5148 | } | ||
5149 | |||
5150 | /** | ||
5151 | * mpt2sas_scsih_reset_handler - reset callback handler (for scsih) | ||
5152 | * @ioc: per adapter object | ||
5153 | * @reset_phase: phase | ||
5154 | * | ||
5155 | * The handler for doing any required cleanup or initialization. | ||
5156 | * | ||
5157 | * The reset phase can be MPT2_IOC_PRE_RESET, MPT2_IOC_AFTER_RESET, | ||
5158 | * MPT2_IOC_DONE_RESET | ||
5159 | * | ||
5160 | * Return nothing. | ||
5161 | */ | ||
5162 | void | ||
5163 | mpt2sas_scsih_reset_handler(struct MPT2SAS_ADAPTER *ioc, int reset_phase) | ||
5164 | { | ||
5165 | switch (reset_phase) { | ||
5166 | case MPT2_IOC_PRE_RESET: | ||
5167 | dtmprintk(ioc, printk(MPT2SAS_DEBUG_FMT "%s: " | ||
5168 | "MPT2_IOC_PRE_RESET\n", ioc->name, __func__)); | ||
5169 | _scsih_fw_event_off(ioc); | ||
5170 | break; | ||
5171 | case MPT2_IOC_AFTER_RESET: | ||
5172 | dtmprintk(ioc, printk(MPT2SAS_DEBUG_FMT "%s: " | ||
5173 | "MPT2_IOC_AFTER_RESET\n", ioc->name, __func__)); | ||
5174 | if (ioc->tm_cmds.status & MPT2_CMD_PENDING) { | ||
5175 | ioc->tm_cmds.status |= MPT2_CMD_RESET; | ||
5176 | mpt2sas_base_free_smid(ioc, ioc->tm_cmds.smid); | ||
5177 | complete(&ioc->tm_cmds.done); | ||
5178 | } | ||
5179 | _scsih_fw_event_on(ioc); | ||
5180 | _scsih_flush_running_cmds(ioc); | ||
5181 | break; | ||
5182 | case MPT2_IOC_DONE_RESET: | ||
5183 | dtmprintk(ioc, printk(MPT2SAS_DEBUG_FMT "%s: " | ||
5184 | "MPT2_IOC_DONE_RESET\n", ioc->name, __func__)); | ||
5185 | _scsih_sas_host_refresh(ioc, 0); | ||
5186 | _scsih_search_responding_sas_devices(ioc); | ||
5187 | _scsih_search_responding_raid_devices(ioc); | ||
5188 | _scsih_search_responding_expanders(ioc); | ||
5189 | break; | ||
5190 | case MPT2_IOC_RUNNING: | ||
5191 | dtmprintk(ioc, printk(MPT2SAS_DEBUG_FMT "%s: " | ||
5192 | "MPT2_IOC_RUNNING\n", ioc->name, __func__)); | ||
5193 | _scsih_remove_unresponding_devices(ioc); | ||
5194 | break; | ||
5220 | } | 5195 | } |
5221 | } | 5196 | } |
5222 | 5197 | ||
@@ -5236,14 +5211,6 @@ _firmware_event_work(struct work_struct *work) | |||
5236 | unsigned long flags; | 5211 | unsigned long flags; |
5237 | struct MPT2SAS_ADAPTER *ioc = fw_event->ioc; | 5212 | struct MPT2SAS_ADAPTER *ioc = fw_event->ioc; |
5238 | 5213 | ||
5239 | /* This is invoked by calling _scsih_queue_rescan(). */ | ||
5240 | if (fw_event->event == MPT2SAS_RESCAN_AFTER_HOST_RESET) { | ||
5241 | _scsih_fw_event_free(ioc, fw_event); | ||
5242 | _scsih_sas_host_refresh(ioc, 1); | ||
5243 | _scsih_remove_unresponding_devices(ioc); | ||
5244 | return; | ||
5245 | } | ||
5246 | |||
5247 | /* the queue is being flushed so ignore this event */ | 5214 | /* the queue is being flushed so ignore this event */ |
5248 | spin_lock_irqsave(&ioc->fw_event_lock, flags); | 5215 | spin_lock_irqsave(&ioc->fw_event_lock, flags); |
5249 | if (ioc->fw_events_off || ioc->remove_host) { | 5216 | if (ioc->fw_events_off || ioc->remove_host) { |
@@ -5253,13 +5220,10 @@ _firmware_event_work(struct work_struct *work) | |||
5253 | } | 5220 | } |
5254 | spin_unlock_irqrestore(&ioc->fw_event_lock, flags); | 5221 | spin_unlock_irqrestore(&ioc->fw_event_lock, flags); |
5255 | 5222 | ||
5256 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); | ||
5257 | if (ioc->shost_recovery) { | 5223 | if (ioc->shost_recovery) { |
5258 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | ||
5259 | _scsih_fw_event_requeue(ioc, fw_event, 1000); | 5224 | _scsih_fw_event_requeue(ioc, fw_event, 1000); |
5260 | return; | 5225 | return; |
5261 | } | 5226 | } |
5262 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | ||
5263 | 5227 | ||
5264 | switch (fw_event->event) { | 5228 | switch (fw_event->event) { |
5265 | case MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST: | 5229 | case MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST: |
@@ -5461,6 +5425,8 @@ _scsih_expander_node_remove(struct MPT2SAS_ADAPTER *ioc, | |||
5461 | if (!sas_device) | 5425 | if (!sas_device) |
5462 | continue; | 5426 | continue; |
5463 | _scsih_remove_device(ioc, sas_device->handle); | 5427 | _scsih_remove_device(ioc, sas_device->handle); |
5428 | if (ioc->shost_recovery) | ||
5429 | return; | ||
5464 | goto retry_device_search; | 5430 | goto retry_device_search; |
5465 | } | 5431 | } |
5466 | } | 5432 | } |
@@ -5482,6 +5448,8 @@ _scsih_expander_node_remove(struct MPT2SAS_ADAPTER *ioc, | |||
5482 | if (!expander_sibling) | 5448 | if (!expander_sibling) |
5483 | continue; | 5449 | continue; |
5484 | _scsih_expander_remove(ioc, expander_sibling->handle); | 5450 | _scsih_expander_remove(ioc, expander_sibling->handle); |
5451 | if (ioc->shost_recovery) | ||
5452 | return; | ||
5485 | goto retry_expander_search; | 5453 | goto retry_expander_search; |
5486 | } | 5454 | } |
5487 | } | 5455 | } |
diff --git a/drivers/scsi/mpt2sas/mpt2sas_transport.c b/drivers/scsi/mpt2sas/mpt2sas_transport.c index 686695b155c7..742324a0a11e 100644 --- a/drivers/scsi/mpt2sas/mpt2sas_transport.c +++ b/drivers/scsi/mpt2sas/mpt2sas_transport.c | |||
@@ -140,11 +140,18 @@ _transport_set_identify(struct MPT2SAS_ADAPTER *ioc, u16 handle, | |||
140 | u32 device_info; | 140 | u32 device_info; |
141 | u32 ioc_status; | 141 | u32 ioc_status; |
142 | 142 | ||
143 | if (ioc->shost_recovery) { | ||
144 | printk(MPT2SAS_INFO_FMT "%s: host reset in progress!\n", | ||
145 | __func__, ioc->name); | ||
146 | return -EFAULT; | ||
147 | } | ||
148 | |||
143 | if ((mpt2sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, | 149 | if ((mpt2sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, |
144 | MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { | 150 | MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { |
145 | printk(MPT2SAS_ERR_FMT "failure at %s:%d/%s()!\n", | 151 | printk(MPT2SAS_ERR_FMT "failure at %s:%d/%s()!\n", |
152 | |||
146 | ioc->name, __FILE__, __LINE__, __func__); | 153 | ioc->name, __FILE__, __LINE__, __func__); |
147 | return -1; | 154 | return -ENXIO; |
148 | } | 155 | } |
149 | 156 | ||
150 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & | 157 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
@@ -153,7 +160,7 @@ _transport_set_identify(struct MPT2SAS_ADAPTER *ioc, u16 handle, | |||
153 | printk(MPT2SAS_ERR_FMT "handle(0x%04x), ioc_status(0x%04x)" | 160 | printk(MPT2SAS_ERR_FMT "handle(0x%04x), ioc_status(0x%04x)" |
154 | "\nfailure at %s:%d/%s()!\n", ioc->name, handle, ioc_status, | 161 | "\nfailure at %s:%d/%s()!\n", ioc->name, handle, ioc_status, |
155 | __FILE__, __LINE__, __func__); | 162 | __FILE__, __LINE__, __func__); |
156 | return -1; | 163 | return -EIO; |
157 | } | 164 | } |
158 | 165 | ||
159 | memset(identify, 0, sizeof(identify)); | 166 | memset(identify, 0, sizeof(identify)); |
@@ -288,21 +295,17 @@ _transport_expander_report_manufacture(struct MPT2SAS_ADAPTER *ioc, | |||
288 | void *psge; | 295 | void *psge; |
289 | u32 sgl_flags; | 296 | u32 sgl_flags; |
290 | u8 issue_reset = 0; | 297 | u8 issue_reset = 0; |
291 | unsigned long flags; | ||
292 | void *data_out = NULL; | 298 | void *data_out = NULL; |
293 | dma_addr_t data_out_dma; | 299 | dma_addr_t data_out_dma; |
294 | u32 sz; | 300 | u32 sz; |
295 | u64 *sas_address_le; | 301 | u64 *sas_address_le; |
296 | u16 wait_state_count; | 302 | u16 wait_state_count; |
297 | 303 | ||
298 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); | 304 | if (ioc->shost_recovery) { |
299 | if (ioc->ioc_reset_in_progress) { | ||
300 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | ||
301 | printk(MPT2SAS_INFO_FMT "%s: host reset in progress!\n", | 305 | printk(MPT2SAS_INFO_FMT "%s: host reset in progress!\n", |
302 | __func__, ioc->name); | 306 | __func__, ioc->name); |
303 | return -EFAULT; | 307 | return -EFAULT; |
304 | } | 308 | } |
305 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | ||
306 | 309 | ||
307 | mutex_lock(&ioc->transport_cmds.mutex); | 310 | mutex_lock(&ioc->transport_cmds.mutex); |
308 | 311 | ||
@@ -789,7 +792,7 @@ mpt2sas_transport_add_expander_phy(struct MPT2SAS_ADAPTER *ioc, struct _sas_phy | |||
789 | } | 792 | } |
790 | 793 | ||
791 | /** | 794 | /** |
792 | * mpt2sas_transport_update_phy_link_change - refreshing phy link changes and attached devices | 795 | * mpt2sas_transport_update_links - refreshing phy link changes |
793 | * @ioc: per adapter object | 796 | * @ioc: per adapter object |
794 | * @handle: handle to sas_host or expander | 797 | * @handle: handle to sas_host or expander |
795 | * @attached_handle: attached device handle | 798 | * @attached_handle: attached device handle |
@@ -799,13 +802,19 @@ mpt2sas_transport_add_expander_phy(struct MPT2SAS_ADAPTER *ioc, struct _sas_phy | |||
799 | * Returns nothing. | 802 | * Returns nothing. |
800 | */ | 803 | */ |
801 | void | 804 | void |
802 | mpt2sas_transport_update_phy_link_change(struct MPT2SAS_ADAPTER *ioc, | 805 | mpt2sas_transport_update_links(struct MPT2SAS_ADAPTER *ioc, |
803 | u16 handle, u16 attached_handle, u8 phy_number, u8 link_rate) | 806 | u16 handle, u16 attached_handle, u8 phy_number, u8 link_rate) |
804 | { | 807 | { |
805 | unsigned long flags; | 808 | unsigned long flags; |
806 | struct _sas_node *sas_node; | 809 | struct _sas_node *sas_node; |
807 | struct _sas_phy *mpt2sas_phy; | 810 | struct _sas_phy *mpt2sas_phy; |
808 | 811 | ||
812 | if (ioc->shost_recovery) { | ||
813 | printk(MPT2SAS_INFO_FMT "%s: host reset in progress!\n", | ||
814 | __func__, ioc->name); | ||
815 | return; | ||
816 | } | ||
817 | |||
809 | spin_lock_irqsave(&ioc->sas_node_lock, flags); | 818 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
810 | sas_node = _transport_sas_node_find_by_handle(ioc, handle); | 819 | sas_node = _transport_sas_node_find_by_handle(ioc, handle); |
811 | spin_unlock_irqrestore(&ioc->sas_node_lock, flags); | 820 | spin_unlock_irqrestore(&ioc->sas_node_lock, flags); |
@@ -1025,7 +1034,6 @@ _transport_smp_handler(struct Scsi_Host *shost, struct sas_rphy *rphy, | |||
1025 | void *psge; | 1034 | void *psge; |
1026 | u32 sgl_flags; | 1035 | u32 sgl_flags; |
1027 | u8 issue_reset = 0; | 1036 | u8 issue_reset = 0; |
1028 | unsigned long flags; | ||
1029 | dma_addr_t dma_addr_in = 0; | 1037 | dma_addr_t dma_addr_in = 0; |
1030 | dma_addr_t dma_addr_out = 0; | 1038 | dma_addr_t dma_addr_out = 0; |
1031 | u16 wait_state_count; | 1039 | u16 wait_state_count; |
@@ -1045,14 +1053,11 @@ _transport_smp_handler(struct Scsi_Host *shost, struct sas_rphy *rphy, | |||
1045 | return -EINVAL; | 1053 | return -EINVAL; |
1046 | } | 1054 | } |
1047 | 1055 | ||
1048 | spin_lock_irqsave(&ioc->ioc_reset_in_progress_lock, flags); | 1056 | if (ioc->shost_recovery) { |
1049 | if (ioc->ioc_reset_in_progress) { | ||
1050 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | ||
1051 | printk(MPT2SAS_INFO_FMT "%s: host reset in progress!\n", | 1057 | printk(MPT2SAS_INFO_FMT "%s: host reset in progress!\n", |
1052 | __func__, ioc->name); | 1058 | __func__, ioc->name); |
1053 | return -EFAULT; | 1059 | return -EFAULT; |
1054 | } | 1060 | } |
1055 | spin_unlock_irqrestore(&ioc->ioc_reset_in_progress_lock, flags); | ||
1056 | 1061 | ||
1057 | rc = mutex_lock_interruptible(&ioc->transport_cmds.mutex); | 1062 | rc = mutex_lock_interruptible(&ioc->transport_cmds.mutex); |
1058 | if (rc) | 1063 | if (rc) |
diff --git a/drivers/scsi/pcmcia/nsp_cs.c b/drivers/scsi/pcmcia/nsp_cs.c index 70b60ade049e..e32c344d7ad8 100644 --- a/drivers/scsi/pcmcia/nsp_cs.c +++ b/drivers/scsi/pcmcia/nsp_cs.c | |||
@@ -1713,7 +1713,7 @@ static int nsp_cs_config(struct pcmcia_device *link) | |||
1713 | 1713 | ||
1714 | nsp_dbg(NSP_DEBUG_INIT, "in"); | 1714 | nsp_dbg(NSP_DEBUG_INIT, "in"); |
1715 | 1715 | ||
1716 | cfg_mem = kzalloc(sizeof(cfg_mem), GFP_KERNEL); | 1716 | cfg_mem = kzalloc(sizeof(*cfg_mem), GFP_KERNEL); |
1717 | if (!cfg_mem) | 1717 | if (!cfg_mem) |
1718 | return -ENOMEM; | 1718 | return -ENOMEM; |
1719 | cfg_mem->data = data; | 1719 | cfg_mem->data = data; |
diff --git a/drivers/scsi/pmcraid.c b/drivers/scsi/pmcraid.c new file mode 100644 index 000000000000..4302f06e4ec9 --- /dev/null +++ b/drivers/scsi/pmcraid.c | |||
@@ -0,0 +1,5604 @@ | |||
1 | /* | ||
2 | * pmcraid.c -- driver for PMC Sierra MaxRAID controller adapters | ||
3 | * | ||
4 | * Written By: PMC Sierra Corporation | ||
5 | * | ||
6 | * Copyright (C) 2008, 2009 PMC Sierra Inc | ||
7 | * | ||
8 | * This program is free software; you can redistribute it and/or modify | ||
9 | * it under the terms of the GNU General Public License as published by | ||
10 | * the Free Software Foundation; either version 2 of the License, or | ||
11 | * (at your option) any later version. | ||
12 | * | ||
13 | * This program is distributed in the hope that it will be useful, | ||
14 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
15 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
16 | * GNU General Public License for more details. | ||
17 | * | ||
18 | * You should have received a copy of the GNU General Public License | ||
19 | * along with this program; if not, write to the Free Software | ||
20 | * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307, | ||
21 | * USA | ||
22 | * | ||
23 | */ | ||
24 | #include <linux/fs.h> | ||
25 | #include <linux/init.h> | ||
26 | #include <linux/types.h> | ||
27 | #include <linux/errno.h> | ||
28 | #include <linux/kernel.h> | ||
29 | #include <linux/ioport.h> | ||
30 | #include <linux/delay.h> | ||
31 | #include <linux/pci.h> | ||
32 | #include <linux/wait.h> | ||
33 | #include <linux/spinlock.h> | ||
34 | #include <linux/sched.h> | ||
35 | #include <linux/interrupt.h> | ||
36 | #include <linux/blkdev.h> | ||
37 | #include <linux/firmware.h> | ||
38 | #include <linux/module.h> | ||
39 | #include <linux/moduleparam.h> | ||
40 | #include <linux/hdreg.h> | ||
41 | #include <linux/version.h> | ||
42 | #include <linux/io.h> | ||
43 | #include <asm/irq.h> | ||
44 | #include <asm/processor.h> | ||
45 | #include <linux/libata.h> | ||
46 | #include <linux/mutex.h> | ||
47 | #include <scsi/scsi.h> | ||
48 | #include <scsi/scsi_host.h> | ||
49 | #include <scsi/scsi_tcq.h> | ||
50 | #include <scsi/scsi_eh.h> | ||
51 | #include <scsi/scsi_cmnd.h> | ||
52 | #include <scsi/scsicam.h> | ||
53 | |||
54 | #include "pmcraid.h" | ||
55 | |||
56 | /* | ||
57 | * Module configuration parameters | ||
58 | */ | ||
59 | static unsigned int pmcraid_debug_log; | ||
60 | static unsigned int pmcraid_disable_aen; | ||
61 | static unsigned int pmcraid_log_level = IOASC_LOG_LEVEL_MUST; | ||
62 | |||
63 | /* | ||
64 | * Data structures to support multiple adapters by the LLD. | ||
65 | * pmcraid_adapter_count - count of configured adapters | ||
66 | */ | ||
67 | static atomic_t pmcraid_adapter_count = ATOMIC_INIT(0); | ||
68 | |||
69 | /* | ||
70 | * Supporting user-level control interface through IOCTL commands. | ||
71 | * pmcraid_major - major number to use | ||
72 | * pmcraid_minor - minor number(s) to use | ||
73 | */ | ||
74 | static unsigned int pmcraid_major; | ||
75 | static struct class *pmcraid_class; | ||
76 | DECLARE_BITMAP(pmcraid_minor, PMCRAID_MAX_ADAPTERS); | ||
77 | |||
78 | /* | ||
79 | * Module parameters | ||
80 | */ | ||
81 | MODULE_AUTHOR("PMC Sierra Corporation, anil_ravindranath@pmc-sierra.com"); | ||
82 | MODULE_DESCRIPTION("PMC Sierra MaxRAID Controller Driver"); | ||
83 | MODULE_LICENSE("GPL"); | ||
84 | MODULE_VERSION(PMCRAID_DRIVER_VERSION); | ||
85 | |||
86 | module_param_named(log_level, pmcraid_log_level, uint, (S_IRUGO | S_IWUSR)); | ||
87 | MODULE_PARM_DESC(log_level, | ||
88 | "Enables firmware error code logging, default :1 high-severity" | ||
89 | " errors, 2: all errors including high-severity errors," | ||
90 | " 0: disables logging"); | ||
91 | |||
92 | module_param_named(debug, pmcraid_debug_log, uint, (S_IRUGO | S_IWUSR)); | ||
93 | MODULE_PARM_DESC(debug, | ||
94 | "Enable driver verbose message logging. Set 1 to enable." | ||
95 | "(default: 0)"); | ||
96 | |||
97 | module_param_named(disable_aen, pmcraid_disable_aen, uint, (S_IRUGO | S_IWUSR)); | ||
98 | MODULE_PARM_DESC(disable_aen, | ||
99 | "Disable driver aen notifications to apps. Set 1 to disable." | ||
100 | "(default: 0)"); | ||
101 | |||
102 | /* chip specific constants for PMC MaxRAID controllers (same for | ||
103 | * 0x5220 and 0x8010 | ||
104 | */ | ||
105 | static struct pmcraid_chip_details pmcraid_chip_cfg[] = { | ||
106 | { | ||
107 | .ioastatus = 0x0, | ||
108 | .ioarrin = 0x00040, | ||
109 | .mailbox = 0x7FC30, | ||
110 | .global_intr_mask = 0x00034, | ||
111 | .ioa_host_intr = 0x0009C, | ||
112 | .ioa_host_intr_clr = 0x000A0, | ||
113 | .ioa_host_mask = 0x7FC28, | ||
114 | .ioa_host_mask_clr = 0x7FC28, | ||
115 | .host_ioa_intr = 0x00020, | ||
116 | .host_ioa_intr_clr = 0x00020, | ||
117 | .transop_timeout = 300 | ||
118 | } | ||
119 | }; | ||
120 | |||
121 | /* | ||
122 | * PCI device ids supported by pmcraid driver | ||
123 | */ | ||
124 | static struct pci_device_id pmcraid_pci_table[] __devinitdata = { | ||
125 | { PCI_DEVICE(PCI_VENDOR_ID_PMC, PCI_DEVICE_ID_PMC_MAXRAID), | ||
126 | 0, 0, (kernel_ulong_t)&pmcraid_chip_cfg[0] | ||
127 | }, | ||
128 | {} | ||
129 | }; | ||
130 | |||
131 | MODULE_DEVICE_TABLE(pci, pmcraid_pci_table); | ||
132 | |||
133 | |||
134 | |||
135 | /** | ||
136 | * pmcraid_slave_alloc - Prepare for commands to a device | ||
137 | * @scsi_dev: scsi device struct | ||
138 | * | ||
139 | * This function is called by mid-layer prior to sending any command to the new | ||
140 | * device. Stores resource entry details of the device in scsi_device struct. | ||
141 | * Queuecommand uses the resource handle and other details to fill up IOARCB | ||
142 | * while sending commands to the device. | ||
143 | * | ||
144 | * Return value: | ||
145 | * 0 on success / -ENXIO if device does not exist | ||
146 | */ | ||
147 | static int pmcraid_slave_alloc(struct scsi_device *scsi_dev) | ||
148 | { | ||
149 | struct pmcraid_resource_entry *temp, *res = NULL; | ||
150 | struct pmcraid_instance *pinstance; | ||
151 | u8 target, bus, lun; | ||
152 | unsigned long lock_flags; | ||
153 | int rc = -ENXIO; | ||
154 | pinstance = shost_priv(scsi_dev->host); | ||
155 | |||
156 | /* Driver exposes VSET and GSCSI resources only; all other device types | ||
157 | * are not exposed. Resource list is synchronized using resource lock | ||
158 | * so any traversal or modifications to the list should be done inside | ||
159 | * this lock | ||
160 | */ | ||
161 | spin_lock_irqsave(&pinstance->resource_lock, lock_flags); | ||
162 | list_for_each_entry(temp, &pinstance->used_res_q, queue) { | ||
163 | |||
164 | /* do not expose VSETs with order-ids >= 240 */ | ||
165 | if (RES_IS_VSET(temp->cfg_entry)) { | ||
166 | target = temp->cfg_entry.unique_flags1; | ||
167 | if (target >= PMCRAID_MAX_VSET_TARGETS) | ||
168 | continue; | ||
169 | bus = PMCRAID_VSET_BUS_ID; | ||
170 | lun = 0; | ||
171 | } else if (RES_IS_GSCSI(temp->cfg_entry)) { | ||
172 | target = RES_TARGET(temp->cfg_entry.resource_address); | ||
173 | bus = PMCRAID_PHYS_BUS_ID; | ||
174 | lun = RES_LUN(temp->cfg_entry.resource_address); | ||
175 | } else { | ||
176 | continue; | ||
177 | } | ||
178 | |||
179 | if (bus == scsi_dev->channel && | ||
180 | target == scsi_dev->id && | ||
181 | lun == scsi_dev->lun) { | ||
182 | res = temp; | ||
183 | break; | ||
184 | } | ||
185 | } | ||
186 | |||
187 | if (res) { | ||
188 | res->scsi_dev = scsi_dev; | ||
189 | scsi_dev->hostdata = res; | ||
190 | res->change_detected = 0; | ||
191 | atomic_set(&res->read_failures, 0); | ||
192 | atomic_set(&res->write_failures, 0); | ||
193 | rc = 0; | ||
194 | } | ||
195 | spin_unlock_irqrestore(&pinstance->resource_lock, lock_flags); | ||
196 | return rc; | ||
197 | } | ||
198 | |||
199 | /** | ||
200 | * pmcraid_slave_configure - Configures a SCSI device | ||
201 | * @scsi_dev: scsi device struct | ||
202 | * | ||
203 | * This fucntion is executed by SCSI mid layer just after a device is first | ||
204 | * scanned (i.e. it has responded to an INQUIRY). For VSET resources, the | ||
205 | * timeout value (default 30s) will be over-written to a higher value (60s) | ||
206 | * and max_sectors value will be over-written to 512. It also sets queue depth | ||
207 | * to host->cmd_per_lun value | ||
208 | * | ||
209 | * Return value: | ||
210 | * 0 on success | ||
211 | */ | ||
212 | static int pmcraid_slave_configure(struct scsi_device *scsi_dev) | ||
213 | { | ||
214 | struct pmcraid_resource_entry *res = scsi_dev->hostdata; | ||
215 | |||
216 | if (!res) | ||
217 | return 0; | ||
218 | |||
219 | /* LLD exposes VSETs and Enclosure devices only */ | ||
220 | if (RES_IS_GSCSI(res->cfg_entry) && | ||
221 | scsi_dev->type != TYPE_ENCLOSURE) | ||
222 | return -ENXIO; | ||
223 | |||
224 | pmcraid_info("configuring %x:%x:%x:%x\n", | ||
225 | scsi_dev->host->unique_id, | ||
226 | scsi_dev->channel, | ||
227 | scsi_dev->id, | ||
228 | scsi_dev->lun); | ||
229 | |||
230 | if (RES_IS_GSCSI(res->cfg_entry)) { | ||
231 | scsi_dev->allow_restart = 1; | ||
232 | } else if (RES_IS_VSET(res->cfg_entry)) { | ||
233 | scsi_dev->allow_restart = 1; | ||
234 | blk_queue_rq_timeout(scsi_dev->request_queue, | ||
235 | PMCRAID_VSET_IO_TIMEOUT); | ||
236 | blk_queue_max_sectors(scsi_dev->request_queue, | ||
237 | PMCRAID_VSET_MAX_SECTORS); | ||
238 | } | ||
239 | |||
240 | if (scsi_dev->tagged_supported && | ||
241 | (RES_IS_GSCSI(res->cfg_entry) || RES_IS_VSET(res->cfg_entry))) { | ||
242 | scsi_activate_tcq(scsi_dev, scsi_dev->queue_depth); | ||
243 | scsi_adjust_queue_depth(scsi_dev, MSG_SIMPLE_TAG, | ||
244 | scsi_dev->host->cmd_per_lun); | ||
245 | } else { | ||
246 | scsi_adjust_queue_depth(scsi_dev, 0, | ||
247 | scsi_dev->host->cmd_per_lun); | ||
248 | } | ||
249 | |||
250 | return 0; | ||
251 | } | ||
252 | |||
253 | /** | ||
254 | * pmcraid_slave_destroy - Unconfigure a SCSI device before removing it | ||
255 | * | ||
256 | * @scsi_dev: scsi device struct | ||
257 | * | ||
258 | * This is called by mid-layer before removing a device. Pointer assignments | ||
259 | * done in pmcraid_slave_alloc will be reset to NULL here. | ||
260 | * | ||
261 | * Return value | ||
262 | * none | ||
263 | */ | ||
264 | static void pmcraid_slave_destroy(struct scsi_device *scsi_dev) | ||
265 | { | ||
266 | struct pmcraid_resource_entry *res; | ||
267 | |||
268 | res = (struct pmcraid_resource_entry *)scsi_dev->hostdata; | ||
269 | |||
270 | if (res) | ||
271 | res->scsi_dev = NULL; | ||
272 | |||
273 | scsi_dev->hostdata = NULL; | ||
274 | } | ||
275 | |||
276 | /** | ||
277 | * pmcraid_change_queue_depth - Change the device's queue depth | ||
278 | * @scsi_dev: scsi device struct | ||
279 | * @depth: depth to set | ||
280 | * | ||
281 | * Return value | ||
282 | * actual depth set | ||
283 | */ | ||
284 | static int pmcraid_change_queue_depth(struct scsi_device *scsi_dev, int depth) | ||
285 | { | ||
286 | if (depth > PMCRAID_MAX_CMD_PER_LUN) | ||
287 | depth = PMCRAID_MAX_CMD_PER_LUN; | ||
288 | |||
289 | scsi_adjust_queue_depth(scsi_dev, scsi_get_tag_type(scsi_dev), depth); | ||
290 | |||
291 | return scsi_dev->queue_depth; | ||
292 | } | ||
293 | |||
294 | /** | ||
295 | * pmcraid_change_queue_type - Change the device's queue type | ||
296 | * @scsi_dev: scsi device struct | ||
297 | * @tag: type of tags to use | ||
298 | * | ||
299 | * Return value: | ||
300 | * actual queue type set | ||
301 | */ | ||
302 | static int pmcraid_change_queue_type(struct scsi_device *scsi_dev, int tag) | ||
303 | { | ||
304 | struct pmcraid_resource_entry *res; | ||
305 | |||
306 | res = (struct pmcraid_resource_entry *)scsi_dev->hostdata; | ||
307 | |||
308 | if ((res) && scsi_dev->tagged_supported && | ||
309 | (RES_IS_GSCSI(res->cfg_entry) || RES_IS_VSET(res->cfg_entry))) { | ||
310 | scsi_set_tag_type(scsi_dev, tag); | ||
311 | |||
312 | if (tag) | ||
313 | scsi_activate_tcq(scsi_dev, scsi_dev->queue_depth); | ||
314 | else | ||
315 | scsi_deactivate_tcq(scsi_dev, scsi_dev->queue_depth); | ||
316 | } else | ||
317 | tag = 0; | ||
318 | |||
319 | return tag; | ||
320 | } | ||
321 | |||
322 | |||
323 | /** | ||
324 | * pmcraid_init_cmdblk - initializes a command block | ||
325 | * | ||
326 | * @cmd: pointer to struct pmcraid_cmd to be initialized | ||
327 | * @index: if >=0 first time initialization; otherwise reinitialization | ||
328 | * | ||
329 | * Return Value | ||
330 | * None | ||
331 | */ | ||
332 | void pmcraid_init_cmdblk(struct pmcraid_cmd *cmd, int index) | ||
333 | { | ||
334 | struct pmcraid_ioarcb *ioarcb = &(cmd->ioa_cb->ioarcb); | ||
335 | dma_addr_t dma_addr = cmd->ioa_cb_bus_addr; | ||
336 | |||
337 | if (index >= 0) { | ||
338 | /* first time initialization (called from probe) */ | ||
339 | u32 ioasa_offset = | ||
340 | offsetof(struct pmcraid_control_block, ioasa); | ||
341 | |||
342 | cmd->index = index; | ||
343 | ioarcb->response_handle = cpu_to_le32(index << 2); | ||
344 | ioarcb->ioarcb_bus_addr = cpu_to_le64(dma_addr); | ||
345 | ioarcb->ioasa_bus_addr = cpu_to_le64(dma_addr + ioasa_offset); | ||
346 | ioarcb->ioasa_len = cpu_to_le16(sizeof(struct pmcraid_ioasa)); | ||
347 | } else { | ||
348 | /* re-initialization of various lengths, called once command is | ||
349 | * processed by IOA | ||
350 | */ | ||
351 | memset(&cmd->ioa_cb->ioarcb.cdb, 0, PMCRAID_MAX_CDB_LEN); | ||
352 | ioarcb->request_flags0 = 0; | ||
353 | ioarcb->request_flags1 = 0; | ||
354 | ioarcb->cmd_timeout = 0; | ||
355 | ioarcb->ioarcb_bus_addr &= (~0x1FULL); | ||
356 | ioarcb->ioadl_bus_addr = 0; | ||
357 | ioarcb->ioadl_length = 0; | ||
358 | ioarcb->data_transfer_length = 0; | ||
359 | ioarcb->add_cmd_param_length = 0; | ||
360 | ioarcb->add_cmd_param_offset = 0; | ||
361 | cmd->ioa_cb->ioasa.ioasc = 0; | ||
362 | cmd->ioa_cb->ioasa.residual_data_length = 0; | ||
363 | cmd->u.time_left = 0; | ||
364 | } | ||
365 | |||
366 | cmd->cmd_done = NULL; | ||
367 | cmd->scsi_cmd = NULL; | ||
368 | cmd->release = 0; | ||
369 | cmd->completion_req = 0; | ||
370 | cmd->dma_handle = 0; | ||
371 | init_timer(&cmd->timer); | ||
372 | } | ||
373 | |||
374 | /** | ||
375 | * pmcraid_reinit_cmdblk - reinitialize a command block | ||
376 | * | ||
377 | * @cmd: pointer to struct pmcraid_cmd to be reinitialized | ||
378 | * | ||
379 | * Return Value | ||
380 | * None | ||
381 | */ | ||
382 | static void pmcraid_reinit_cmdblk(struct pmcraid_cmd *cmd) | ||
383 | { | ||
384 | pmcraid_init_cmdblk(cmd, -1); | ||
385 | } | ||
386 | |||
387 | /** | ||
388 | * pmcraid_get_free_cmd - get a free cmd block from command block pool | ||
389 | * @pinstance: adapter instance structure | ||
390 | * | ||
391 | * Return Value: | ||
392 | * returns pointer to cmd block or NULL if no blocks are available | ||
393 | */ | ||
394 | static struct pmcraid_cmd *pmcraid_get_free_cmd( | ||
395 | struct pmcraid_instance *pinstance | ||
396 | ) | ||
397 | { | ||
398 | struct pmcraid_cmd *cmd = NULL; | ||
399 | unsigned long lock_flags; | ||
400 | |||
401 | /* free cmd block list is protected by free_pool_lock */ | ||
402 | spin_lock_irqsave(&pinstance->free_pool_lock, lock_flags); | ||
403 | |||
404 | if (!list_empty(&pinstance->free_cmd_pool)) { | ||
405 | cmd = list_entry(pinstance->free_cmd_pool.next, | ||
406 | struct pmcraid_cmd, free_list); | ||
407 | list_del(&cmd->free_list); | ||
408 | } | ||
409 | spin_unlock_irqrestore(&pinstance->free_pool_lock, lock_flags); | ||
410 | |||
411 | /* Initialize the command block before giving it the caller */ | ||
412 | if (cmd != NULL) | ||
413 | pmcraid_reinit_cmdblk(cmd); | ||
414 | return cmd; | ||
415 | } | ||
416 | |||
417 | /** | ||
418 | * pmcraid_return_cmd - return a completed command block back into free pool | ||
419 | * @cmd: pointer to the command block | ||
420 | * | ||
421 | * Return Value: | ||
422 | * nothing | ||
423 | */ | ||
424 | void pmcraid_return_cmd(struct pmcraid_cmd *cmd) | ||
425 | { | ||
426 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
427 | unsigned long lock_flags; | ||
428 | |||
429 | spin_lock_irqsave(&pinstance->free_pool_lock, lock_flags); | ||
430 | list_add_tail(&cmd->free_list, &pinstance->free_cmd_pool); | ||
431 | spin_unlock_irqrestore(&pinstance->free_pool_lock, lock_flags); | ||
432 | } | ||
433 | |||
434 | /** | ||
435 | * pmcraid_read_interrupts - reads IOA interrupts | ||
436 | * | ||
437 | * @pinstance: pointer to adapter instance structure | ||
438 | * | ||
439 | * Return value | ||
440 | * interrupts read from IOA | ||
441 | */ | ||
442 | static u32 pmcraid_read_interrupts(struct pmcraid_instance *pinstance) | ||
443 | { | ||
444 | return ioread32(pinstance->int_regs.ioa_host_interrupt_reg); | ||
445 | } | ||
446 | |||
447 | /** | ||
448 | * pmcraid_disable_interrupts - Masks and clears all specified interrupts | ||
449 | * | ||
450 | * @pinstance: pointer to per adapter instance structure | ||
451 | * @intrs: interrupts to disable | ||
452 | * | ||
453 | * Return Value | ||
454 | * None | ||
455 | */ | ||
456 | static void pmcraid_disable_interrupts( | ||
457 | struct pmcraid_instance *pinstance, | ||
458 | u32 intrs | ||
459 | ) | ||
460 | { | ||
461 | u32 gmask = ioread32(pinstance->int_regs.global_interrupt_mask_reg); | ||
462 | u32 nmask = gmask | GLOBAL_INTERRUPT_MASK; | ||
463 | |||
464 | iowrite32(nmask, pinstance->int_regs.global_interrupt_mask_reg); | ||
465 | iowrite32(intrs, pinstance->int_regs.ioa_host_interrupt_clr_reg); | ||
466 | iowrite32(intrs, pinstance->int_regs.ioa_host_interrupt_mask_reg); | ||
467 | ioread32(pinstance->int_regs.ioa_host_interrupt_mask_reg); | ||
468 | } | ||
469 | |||
470 | /** | ||
471 | * pmcraid_enable_interrupts - Enables specified interrupts | ||
472 | * | ||
473 | * @pinstance: pointer to per adapter instance structure | ||
474 | * @intr: interrupts to enable | ||
475 | * | ||
476 | * Return Value | ||
477 | * None | ||
478 | */ | ||
479 | static void pmcraid_enable_interrupts( | ||
480 | struct pmcraid_instance *pinstance, | ||
481 | u32 intrs | ||
482 | ) | ||
483 | { | ||
484 | u32 gmask = ioread32(pinstance->int_regs.global_interrupt_mask_reg); | ||
485 | u32 nmask = gmask & (~GLOBAL_INTERRUPT_MASK); | ||
486 | |||
487 | iowrite32(nmask, pinstance->int_regs.global_interrupt_mask_reg); | ||
488 | iowrite32(~intrs, pinstance->int_regs.ioa_host_interrupt_mask_reg); | ||
489 | ioread32(pinstance->int_regs.ioa_host_interrupt_mask_reg); | ||
490 | |||
491 | pmcraid_info("enabled interrupts global mask = %x intr_mask = %x\n", | ||
492 | ioread32(pinstance->int_regs.global_interrupt_mask_reg), | ||
493 | ioread32(pinstance->int_regs.ioa_host_interrupt_mask_reg)); | ||
494 | } | ||
495 | |||
496 | /** | ||
497 | * pmcraid_reset_type - Determine the required reset type | ||
498 | * @pinstance: pointer to adapter instance structure | ||
499 | * | ||
500 | * IOA requires hard reset if any of the following conditions is true. | ||
501 | * 1. If HRRQ valid interrupt is not masked | ||
502 | * 2. IOA reset alert doorbell is set | ||
503 | * 3. If there are any error interrupts | ||
504 | */ | ||
505 | static void pmcraid_reset_type(struct pmcraid_instance *pinstance) | ||
506 | { | ||
507 | u32 mask; | ||
508 | u32 intrs; | ||
509 | u32 alerts; | ||
510 | |||
511 | mask = ioread32(pinstance->int_regs.ioa_host_interrupt_mask_reg); | ||
512 | intrs = ioread32(pinstance->int_regs.ioa_host_interrupt_reg); | ||
513 | alerts = ioread32(pinstance->int_regs.host_ioa_interrupt_reg); | ||
514 | |||
515 | if ((mask & INTRS_HRRQ_VALID) == 0 || | ||
516 | (alerts & DOORBELL_IOA_RESET_ALERT) || | ||
517 | (intrs & PMCRAID_ERROR_INTERRUPTS)) { | ||
518 | pmcraid_info("IOA requires hard reset\n"); | ||
519 | pinstance->ioa_hard_reset = 1; | ||
520 | } | ||
521 | |||
522 | /* If unit check is active, trigger the dump */ | ||
523 | if (intrs & INTRS_IOA_UNIT_CHECK) | ||
524 | pinstance->ioa_unit_check = 1; | ||
525 | } | ||
526 | |||
527 | /** | ||
528 | * pmcraid_bist_done - completion function for PCI BIST | ||
529 | * @cmd: pointer to reset command | ||
530 | * Return Value | ||
531 | * none | ||
532 | */ | ||
533 | |||
534 | static void pmcraid_ioa_reset(struct pmcraid_cmd *); | ||
535 | |||
536 | static void pmcraid_bist_done(struct pmcraid_cmd *cmd) | ||
537 | { | ||
538 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
539 | unsigned long lock_flags; | ||
540 | int rc; | ||
541 | u16 pci_reg; | ||
542 | |||
543 | rc = pci_read_config_word(pinstance->pdev, PCI_COMMAND, &pci_reg); | ||
544 | |||
545 | /* If PCI config space can't be accessed wait for another two secs */ | ||
546 | if ((rc != PCIBIOS_SUCCESSFUL || (!(pci_reg & PCI_COMMAND_MEMORY))) && | ||
547 | cmd->u.time_left > 0) { | ||
548 | pmcraid_info("BIST not complete, waiting another 2 secs\n"); | ||
549 | cmd->timer.expires = jiffies + cmd->u.time_left; | ||
550 | cmd->u.time_left = 0; | ||
551 | cmd->timer.data = (unsigned long)cmd; | ||
552 | cmd->timer.function = | ||
553 | (void (*)(unsigned long))pmcraid_bist_done; | ||
554 | add_timer(&cmd->timer); | ||
555 | } else { | ||
556 | cmd->u.time_left = 0; | ||
557 | pmcraid_info("BIST is complete, proceeding with reset\n"); | ||
558 | spin_lock_irqsave(pinstance->host->host_lock, lock_flags); | ||
559 | pmcraid_ioa_reset(cmd); | ||
560 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
561 | } | ||
562 | } | ||
563 | |||
564 | /** | ||
565 | * pmcraid_start_bist - starts BIST | ||
566 | * @cmd: pointer to reset cmd | ||
567 | * Return Value | ||
568 | * none | ||
569 | */ | ||
570 | static void pmcraid_start_bist(struct pmcraid_cmd *cmd) | ||
571 | { | ||
572 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
573 | u32 doorbells, intrs; | ||
574 | |||
575 | /* proceed with bist and wait for 2 seconds */ | ||
576 | iowrite32(DOORBELL_IOA_START_BIST, | ||
577 | pinstance->int_regs.host_ioa_interrupt_reg); | ||
578 | doorbells = ioread32(pinstance->int_regs.host_ioa_interrupt_reg); | ||
579 | intrs = ioread32(pinstance->int_regs.ioa_host_interrupt_reg); | ||
580 | pmcraid_info("doorbells after start bist: %x intrs: %x \n", | ||
581 | doorbells, intrs); | ||
582 | |||
583 | cmd->u.time_left = msecs_to_jiffies(PMCRAID_BIST_TIMEOUT); | ||
584 | cmd->timer.data = (unsigned long)cmd; | ||
585 | cmd->timer.expires = jiffies + msecs_to_jiffies(PMCRAID_BIST_TIMEOUT); | ||
586 | cmd->timer.function = (void (*)(unsigned long))pmcraid_bist_done; | ||
587 | add_timer(&cmd->timer); | ||
588 | } | ||
589 | |||
590 | /** | ||
591 | * pmcraid_reset_alert_done - completion routine for reset_alert | ||
592 | * @cmd: pointer to command block used in reset sequence | ||
593 | * Return value | ||
594 | * None | ||
595 | */ | ||
596 | static void pmcraid_reset_alert_done(struct pmcraid_cmd *cmd) | ||
597 | { | ||
598 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
599 | u32 status = ioread32(pinstance->ioa_status); | ||
600 | unsigned long lock_flags; | ||
601 | |||
602 | /* if the critical operation in progress bit is set or the wait times | ||
603 | * out, invoke reset engine to proceed with hard reset. If there is | ||
604 | * some more time to wait, restart the timer | ||
605 | */ | ||
606 | if (((status & INTRS_CRITICAL_OP_IN_PROGRESS) == 0) || | ||
607 | cmd->u.time_left <= 0) { | ||
608 | pmcraid_info("critical op is reset proceeding with reset\n"); | ||
609 | spin_lock_irqsave(pinstance->host->host_lock, lock_flags); | ||
610 | pmcraid_ioa_reset(cmd); | ||
611 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
612 | } else { | ||
613 | pmcraid_info("critical op is not yet reset waiting again\n"); | ||
614 | /* restart timer if some more time is available to wait */ | ||
615 | cmd->u.time_left -= PMCRAID_CHECK_FOR_RESET_TIMEOUT; | ||
616 | cmd->timer.data = (unsigned long)cmd; | ||
617 | cmd->timer.expires = jiffies + PMCRAID_CHECK_FOR_RESET_TIMEOUT; | ||
618 | cmd->timer.function = | ||
619 | (void (*)(unsigned long))pmcraid_reset_alert_done; | ||
620 | add_timer(&cmd->timer); | ||
621 | } | ||
622 | } | ||
623 | |||
624 | /** | ||
625 | * pmcraid_reset_alert - alerts IOA for a possible reset | ||
626 | * @cmd : command block to be used for reset sequence. | ||
627 | * | ||
628 | * Return Value | ||
629 | * returns 0 if pci config-space is accessible and RESET_DOORBELL is | ||
630 | * successfully written to IOA. Returns non-zero in case pci_config_space | ||
631 | * is not accessible | ||
632 | */ | ||
633 | static void pmcraid_reset_alert(struct pmcraid_cmd *cmd) | ||
634 | { | ||
635 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
636 | u32 doorbells; | ||
637 | int rc; | ||
638 | u16 pci_reg; | ||
639 | |||
640 | /* If we are able to access IOA PCI config space, alert IOA that we are | ||
641 | * going to reset it soon. This enables IOA to preserv persistent error | ||
642 | * data if any. In case memory space is not accessible, proceed with | ||
643 | * BIST or slot_reset | ||
644 | */ | ||
645 | rc = pci_read_config_word(pinstance->pdev, PCI_COMMAND, &pci_reg); | ||
646 | if ((rc == PCIBIOS_SUCCESSFUL) && (pci_reg & PCI_COMMAND_MEMORY)) { | ||
647 | |||
648 | /* wait for IOA permission i.e until CRITICAL_OPERATION bit is | ||
649 | * reset IOA doesn't generate any interrupts when CRITICAL | ||
650 | * OPERATION bit is reset. A timer is started to wait for this | ||
651 | * bit to be reset. | ||
652 | */ | ||
653 | cmd->u.time_left = PMCRAID_RESET_TIMEOUT; | ||
654 | cmd->timer.data = (unsigned long)cmd; | ||
655 | cmd->timer.expires = jiffies + PMCRAID_CHECK_FOR_RESET_TIMEOUT; | ||
656 | cmd->timer.function = | ||
657 | (void (*)(unsigned long))pmcraid_reset_alert_done; | ||
658 | add_timer(&cmd->timer); | ||
659 | |||
660 | iowrite32(DOORBELL_IOA_RESET_ALERT, | ||
661 | pinstance->int_regs.host_ioa_interrupt_reg); | ||
662 | doorbells = | ||
663 | ioread32(pinstance->int_regs.host_ioa_interrupt_reg); | ||
664 | pmcraid_info("doorbells after reset alert: %x\n", doorbells); | ||
665 | } else { | ||
666 | pmcraid_info("PCI config is not accessible starting BIST\n"); | ||
667 | pinstance->ioa_state = IOA_STATE_IN_HARD_RESET; | ||
668 | pmcraid_start_bist(cmd); | ||
669 | } | ||
670 | } | ||
671 | |||
672 | /** | ||
673 | * pmcraid_timeout_handler - Timeout handler for internally generated ops | ||
674 | * | ||
675 | * @cmd : pointer to command structure, that got timedout | ||
676 | * | ||
677 | * This function blocks host requests and initiates an adapter reset. | ||
678 | * | ||
679 | * Return value: | ||
680 | * None | ||
681 | */ | ||
682 | static void pmcraid_timeout_handler(struct pmcraid_cmd *cmd) | ||
683 | { | ||
684 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
685 | unsigned long lock_flags; | ||
686 | |||
687 | dev_err(&pinstance->pdev->dev, | ||
688 | "Adapter being reset due to command timeout.\n"); | ||
689 | |||
690 | /* Command timeouts result in hard reset sequence. The command that got | ||
691 | * timed out may be the one used as part of reset sequence. In this | ||
692 | * case restart reset sequence using the same command block even if | ||
693 | * reset is in progress. Otherwise fail this command and get a free | ||
694 | * command block to restart the reset sequence. | ||
695 | */ | ||
696 | spin_lock_irqsave(pinstance->host->host_lock, lock_flags); | ||
697 | if (!pinstance->ioa_reset_in_progress) { | ||
698 | pinstance->ioa_reset_attempts = 0; | ||
699 | cmd = pmcraid_get_free_cmd(pinstance); | ||
700 | |||
701 | /* If we are out of command blocks, just return here itself. | ||
702 | * Some other command's timeout handler can do the reset job | ||
703 | */ | ||
704 | if (cmd == NULL) { | ||
705 | spin_unlock_irqrestore(pinstance->host->host_lock, | ||
706 | lock_flags); | ||
707 | pmcraid_err("no free cmnd block for timeout handler\n"); | ||
708 | return; | ||
709 | } | ||
710 | |||
711 | pinstance->reset_cmd = cmd; | ||
712 | pinstance->ioa_reset_in_progress = 1; | ||
713 | } else { | ||
714 | pmcraid_info("reset is already in progress\n"); | ||
715 | |||
716 | if (pinstance->reset_cmd != cmd) { | ||
717 | /* This command should have been given to IOA, this | ||
718 | * command will be completed by fail_outstanding_cmds | ||
719 | * anyway | ||
720 | */ | ||
721 | pmcraid_err("cmd is pending but reset in progress\n"); | ||
722 | } | ||
723 | |||
724 | /* If this command was being used as part of the reset | ||
725 | * sequence, set cmd_done pointer to pmcraid_ioa_reset. This | ||
726 | * causes fail_outstanding_commands not to return the command | ||
727 | * block back to free pool | ||
728 | */ | ||
729 | if (cmd == pinstance->reset_cmd) | ||
730 | cmd->cmd_done = pmcraid_ioa_reset; | ||
731 | |||
732 | } | ||
733 | |||
734 | pinstance->ioa_state = IOA_STATE_IN_RESET_ALERT; | ||
735 | scsi_block_requests(pinstance->host); | ||
736 | pmcraid_reset_alert(cmd); | ||
737 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
738 | } | ||
739 | |||
740 | /** | ||
741 | * pmcraid_internal_done - completion routine for internally generated cmds | ||
742 | * | ||
743 | * @cmd: command that got response from IOA | ||
744 | * | ||
745 | * Return Value: | ||
746 | * none | ||
747 | */ | ||
748 | static void pmcraid_internal_done(struct pmcraid_cmd *cmd) | ||
749 | { | ||
750 | pmcraid_info("response internal cmd CDB[0] = %x ioasc = %x\n", | ||
751 | cmd->ioa_cb->ioarcb.cdb[0], | ||
752 | le32_to_cpu(cmd->ioa_cb->ioasa.ioasc)); | ||
753 | |||
754 | /* Some of the internal commands are sent with callers blocking for the | ||
755 | * response. Same will be indicated as part of cmd->completion_req | ||
756 | * field. Response path needs to wake up any waiters waiting for cmd | ||
757 | * completion if this flag is set. | ||
758 | */ | ||
759 | if (cmd->completion_req) { | ||
760 | cmd->completion_req = 0; | ||
761 | complete(&cmd->wait_for_completion); | ||
762 | } | ||
763 | |||
764 | /* most of the internal commands are completed by caller itself, so | ||
765 | * no need to return the command block back to free pool until we are | ||
766 | * required to do so (e.g once done with initialization). | ||
767 | */ | ||
768 | if (cmd->release) { | ||
769 | cmd->release = 0; | ||
770 | pmcraid_return_cmd(cmd); | ||
771 | } | ||
772 | } | ||
773 | |||
774 | /** | ||
775 | * pmcraid_reinit_cfgtable_done - done function for cfg table reinitialization | ||
776 | * | ||
777 | * @cmd: command that got response from IOA | ||
778 | * | ||
779 | * This routine is called after driver re-reads configuration table due to a | ||
780 | * lost CCN. It returns the command block back to free pool and schedules | ||
781 | * worker thread to add/delete devices into the system. | ||
782 | * | ||
783 | * Return Value: | ||
784 | * none | ||
785 | */ | ||
786 | static void pmcraid_reinit_cfgtable_done(struct pmcraid_cmd *cmd) | ||
787 | { | ||
788 | pmcraid_info("response internal cmd CDB[0] = %x ioasc = %x\n", | ||
789 | cmd->ioa_cb->ioarcb.cdb[0], | ||
790 | le32_to_cpu(cmd->ioa_cb->ioasa.ioasc)); | ||
791 | |||
792 | if (cmd->release) { | ||
793 | cmd->release = 0; | ||
794 | pmcraid_return_cmd(cmd); | ||
795 | } | ||
796 | pmcraid_info("scheduling worker for config table reinitialization\n"); | ||
797 | schedule_work(&cmd->drv_inst->worker_q); | ||
798 | } | ||
799 | |||
800 | /** | ||
801 | * pmcraid_erp_done - Process completion of SCSI error response from device | ||
802 | * @cmd: pmcraid_command | ||
803 | * | ||
804 | * This function copies the sense buffer into the scsi_cmd struct and completes | ||
805 | * scsi_cmd by calling scsi_done function. | ||
806 | * | ||
807 | * Return value: | ||
808 | * none | ||
809 | */ | ||
810 | static void pmcraid_erp_done(struct pmcraid_cmd *cmd) | ||
811 | { | ||
812 | struct scsi_cmnd *scsi_cmd = cmd->scsi_cmd; | ||
813 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
814 | u32 ioasc = le32_to_cpu(cmd->ioa_cb->ioasa.ioasc); | ||
815 | |||
816 | if (PMCRAID_IOASC_SENSE_KEY(ioasc) > 0) { | ||
817 | scsi_cmd->result |= (DID_ERROR << 16); | ||
818 | pmcraid_err("command CDB[0] = %x failed with IOASC: 0x%08X\n", | ||
819 | cmd->ioa_cb->ioarcb.cdb[0], ioasc); | ||
820 | } | ||
821 | |||
822 | /* if we had allocated sense buffers for request sense, copy the sense | ||
823 | * release the buffers | ||
824 | */ | ||
825 | if (cmd->sense_buffer != NULL) { | ||
826 | memcpy(scsi_cmd->sense_buffer, | ||
827 | cmd->sense_buffer, | ||
828 | SCSI_SENSE_BUFFERSIZE); | ||
829 | pci_free_consistent(pinstance->pdev, | ||
830 | SCSI_SENSE_BUFFERSIZE, | ||
831 | cmd->sense_buffer, cmd->sense_buffer_dma); | ||
832 | cmd->sense_buffer = NULL; | ||
833 | cmd->sense_buffer_dma = 0; | ||
834 | } | ||
835 | |||
836 | scsi_dma_unmap(scsi_cmd); | ||
837 | pmcraid_return_cmd(cmd); | ||
838 | scsi_cmd->scsi_done(scsi_cmd); | ||
839 | } | ||
840 | |||
841 | /** | ||
842 | * pmcraid_fire_command - sends an IOA command to adapter | ||
843 | * | ||
844 | * This function adds the given block into pending command list | ||
845 | * and returns without waiting | ||
846 | * | ||
847 | * @cmd : command to be sent to the device | ||
848 | * | ||
849 | * Return Value | ||
850 | * None | ||
851 | */ | ||
852 | static void _pmcraid_fire_command(struct pmcraid_cmd *cmd) | ||
853 | { | ||
854 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
855 | unsigned long lock_flags; | ||
856 | |||
857 | /* Add this command block to pending cmd pool. We do this prior to | ||
858 | * writting IOARCB to ioarrin because IOA might complete the command | ||
859 | * by the time we are about to add it to the list. Response handler | ||
860 | * (isr/tasklet) looks for cmb block in the pending pending list. | ||
861 | */ | ||
862 | spin_lock_irqsave(&pinstance->pending_pool_lock, lock_flags); | ||
863 | list_add_tail(&cmd->free_list, &pinstance->pending_cmd_pool); | ||
864 | spin_unlock_irqrestore(&pinstance->pending_pool_lock, lock_flags); | ||
865 | atomic_inc(&pinstance->outstanding_cmds); | ||
866 | |||
867 | /* driver writes lower 32-bit value of IOARCB address only */ | ||
868 | mb(); | ||
869 | iowrite32(le32_to_cpu(cmd->ioa_cb->ioarcb.ioarcb_bus_addr), | ||
870 | pinstance->ioarrin); | ||
871 | } | ||
872 | |||
873 | /** | ||
874 | * pmcraid_send_cmd - fires a command to IOA | ||
875 | * | ||
876 | * This function also sets up timeout function, and command completion | ||
877 | * function | ||
878 | * | ||
879 | * @cmd: pointer to the command block to be fired to IOA | ||
880 | * @cmd_done: command completion function, called once IOA responds | ||
881 | * @timeout: timeout to wait for this command completion | ||
882 | * @timeout_func: timeout handler | ||
883 | * | ||
884 | * Return value | ||
885 | * none | ||
886 | */ | ||
887 | static void pmcraid_send_cmd( | ||
888 | struct pmcraid_cmd *cmd, | ||
889 | void (*cmd_done) (struct pmcraid_cmd *), | ||
890 | unsigned long timeout, | ||
891 | void (*timeout_func) (struct pmcraid_cmd *) | ||
892 | ) | ||
893 | { | ||
894 | /* initialize done function */ | ||
895 | cmd->cmd_done = cmd_done; | ||
896 | |||
897 | if (timeout_func) { | ||
898 | /* setup timeout handler */ | ||
899 | cmd->timer.data = (unsigned long)cmd; | ||
900 | cmd->timer.expires = jiffies + timeout; | ||
901 | cmd->timer.function = (void (*)(unsigned long))timeout_func; | ||
902 | add_timer(&cmd->timer); | ||
903 | } | ||
904 | |||
905 | /* fire the command to IOA */ | ||
906 | _pmcraid_fire_command(cmd); | ||
907 | } | ||
908 | |||
909 | /** | ||
910 | * pmcraid_ioa_shutdown - sends SHUTDOWN command to ioa | ||
911 | * | ||
912 | * @cmd: pointer to the command block used as part of reset sequence | ||
913 | * | ||
914 | * Return Value | ||
915 | * None | ||
916 | */ | ||
917 | static void pmcraid_ioa_shutdown(struct pmcraid_cmd *cmd) | ||
918 | { | ||
919 | pmcraid_info("response for Cancel CCN CDB[0] = %x ioasc = %x\n", | ||
920 | cmd->ioa_cb->ioarcb.cdb[0], | ||
921 | le32_to_cpu(cmd->ioa_cb->ioasa.ioasc)); | ||
922 | |||
923 | /* Note that commands sent during reset require next command to be sent | ||
924 | * to IOA. Hence reinit the done function as well as timeout function | ||
925 | */ | ||
926 | pmcraid_reinit_cmdblk(cmd); | ||
927 | cmd->ioa_cb->ioarcb.request_type = REQ_TYPE_IOACMD; | ||
928 | cmd->ioa_cb->ioarcb.resource_handle = | ||
929 | cpu_to_le32(PMCRAID_IOA_RES_HANDLE); | ||
930 | cmd->ioa_cb->ioarcb.cdb[0] = PMCRAID_IOA_SHUTDOWN; | ||
931 | cmd->ioa_cb->ioarcb.cdb[1] = PMCRAID_SHUTDOWN_NORMAL; | ||
932 | |||
933 | /* fire shutdown command to hardware. */ | ||
934 | pmcraid_info("firing normal shutdown command (%d) to IOA\n", | ||
935 | le32_to_cpu(cmd->ioa_cb->ioarcb.response_handle)); | ||
936 | |||
937 | pmcraid_send_cmd(cmd, pmcraid_ioa_reset, | ||
938 | PMCRAID_SHUTDOWN_TIMEOUT, | ||
939 | pmcraid_timeout_handler); | ||
940 | } | ||
941 | |||
942 | /** | ||
943 | * pmcraid_identify_hrrq - registers host rrq buffers with IOA | ||
944 | * @cmd: pointer to command block to be used for identify hrrq | ||
945 | * | ||
946 | * Return Value | ||
947 | * 0 in case of success, otherwise non-zero failure code | ||
948 | */ | ||
949 | |||
950 | static void pmcraid_querycfg(struct pmcraid_cmd *); | ||
951 | |||
952 | static void pmcraid_identify_hrrq(struct pmcraid_cmd *cmd) | ||
953 | { | ||
954 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
955 | struct pmcraid_ioarcb *ioarcb = &cmd->ioa_cb->ioarcb; | ||
956 | int index = 0; | ||
957 | __be64 hrrq_addr = cpu_to_be64(pinstance->hrrq_start_bus_addr[index]); | ||
958 | u32 hrrq_size = cpu_to_be32(sizeof(u32) * PMCRAID_MAX_CMD); | ||
959 | |||
960 | pmcraid_reinit_cmdblk(cmd); | ||
961 | |||
962 | /* Initialize ioarcb */ | ||
963 | ioarcb->request_type = REQ_TYPE_IOACMD; | ||
964 | ioarcb->resource_handle = cpu_to_le32(PMCRAID_IOA_RES_HANDLE); | ||
965 | |||
966 | /* initialize the hrrq number where IOA will respond to this command */ | ||
967 | ioarcb->hrrq_id = index; | ||
968 | ioarcb->cdb[0] = PMCRAID_IDENTIFY_HRRQ; | ||
969 | ioarcb->cdb[1] = index; | ||
970 | |||
971 | /* IOA expects 64-bit pci address to be written in B.E format | ||
972 | * (i.e cdb[2]=MSByte..cdb[9]=LSB. | ||
973 | */ | ||
974 | pmcraid_info("HRRQ_IDENTIFY with hrrq:ioarcb => %llx:%llx\n", | ||
975 | hrrq_addr, ioarcb->ioarcb_bus_addr); | ||
976 | |||
977 | memcpy(&(ioarcb->cdb[2]), &hrrq_addr, sizeof(hrrq_addr)); | ||
978 | memcpy(&(ioarcb->cdb[10]), &hrrq_size, sizeof(hrrq_size)); | ||
979 | |||
980 | /* Subsequent commands require HRRQ identification to be successful. | ||
981 | * Note that this gets called even during reset from SCSI mid-layer | ||
982 | * or tasklet | ||
983 | */ | ||
984 | pmcraid_send_cmd(cmd, pmcraid_querycfg, | ||
985 | PMCRAID_INTERNAL_TIMEOUT, | ||
986 | pmcraid_timeout_handler); | ||
987 | } | ||
988 | |||
989 | static void pmcraid_process_ccn(struct pmcraid_cmd *cmd); | ||
990 | static void pmcraid_process_ldn(struct pmcraid_cmd *cmd); | ||
991 | |||
992 | /** | ||
993 | * pmcraid_send_hcam_cmd - send an initialized command block(HCAM) to IOA | ||
994 | * | ||
995 | * @cmd: initialized command block pointer | ||
996 | * | ||
997 | * Return Value | ||
998 | * none | ||
999 | */ | ||
1000 | static void pmcraid_send_hcam_cmd(struct pmcraid_cmd *cmd) | ||
1001 | { | ||
1002 | if (cmd->ioa_cb->ioarcb.cdb[1] == PMCRAID_HCAM_CODE_CONFIG_CHANGE) | ||
1003 | atomic_set(&(cmd->drv_inst->ccn.ignore), 0); | ||
1004 | else | ||
1005 | atomic_set(&(cmd->drv_inst->ldn.ignore), 0); | ||
1006 | |||
1007 | pmcraid_send_cmd(cmd, cmd->cmd_done, 0, NULL); | ||
1008 | } | ||
1009 | |||
1010 | /** | ||
1011 | * pmcraid_init_hcam - send an initialized command block(HCAM) to IOA | ||
1012 | * | ||
1013 | * @pinstance: pointer to adapter instance structure | ||
1014 | * @type: HCAM type | ||
1015 | * | ||
1016 | * Return Value | ||
1017 | * pointer to initialized pmcraid_cmd structure or NULL | ||
1018 | */ | ||
1019 | static struct pmcraid_cmd *pmcraid_init_hcam | ||
1020 | ( | ||
1021 | struct pmcraid_instance *pinstance, | ||
1022 | u8 type | ||
1023 | ) | ||
1024 | { | ||
1025 | struct pmcraid_cmd *cmd; | ||
1026 | struct pmcraid_ioarcb *ioarcb; | ||
1027 | struct pmcraid_ioadl_desc *ioadl; | ||
1028 | struct pmcraid_hostrcb *hcam; | ||
1029 | void (*cmd_done) (struct pmcraid_cmd *); | ||
1030 | dma_addr_t dma; | ||
1031 | int rcb_size; | ||
1032 | |||
1033 | cmd = pmcraid_get_free_cmd(pinstance); | ||
1034 | |||
1035 | if (!cmd) { | ||
1036 | pmcraid_err("no free command blocks for hcam\n"); | ||
1037 | return cmd; | ||
1038 | } | ||
1039 | |||
1040 | if (type == PMCRAID_HCAM_CODE_CONFIG_CHANGE) { | ||
1041 | rcb_size = sizeof(struct pmcraid_hcam_ccn); | ||
1042 | cmd_done = pmcraid_process_ccn; | ||
1043 | dma = pinstance->ccn.baddr + PMCRAID_AEN_HDR_SIZE; | ||
1044 | hcam = &pinstance->ccn; | ||
1045 | } else { | ||
1046 | rcb_size = sizeof(struct pmcraid_hcam_ldn); | ||
1047 | cmd_done = pmcraid_process_ldn; | ||
1048 | dma = pinstance->ldn.baddr + PMCRAID_AEN_HDR_SIZE; | ||
1049 | hcam = &pinstance->ldn; | ||
1050 | } | ||
1051 | |||
1052 | /* initialize command pointer used for HCAM registration */ | ||
1053 | hcam->cmd = cmd; | ||
1054 | |||
1055 | ioarcb = &cmd->ioa_cb->ioarcb; | ||
1056 | ioarcb->ioadl_bus_addr = cpu_to_le64((cmd->ioa_cb_bus_addr) + | ||
1057 | offsetof(struct pmcraid_ioarcb, | ||
1058 | add_data.u.ioadl[0])); | ||
1059 | ioarcb->ioadl_length = cpu_to_le32(sizeof(struct pmcraid_ioadl_desc)); | ||
1060 | ioadl = ioarcb->add_data.u.ioadl; | ||
1061 | |||
1062 | /* Initialize ioarcb */ | ||
1063 | ioarcb->request_type = REQ_TYPE_HCAM; | ||
1064 | ioarcb->resource_handle = cpu_to_le32(PMCRAID_IOA_RES_HANDLE); | ||
1065 | ioarcb->cdb[0] = PMCRAID_HOST_CONTROLLED_ASYNC; | ||
1066 | ioarcb->cdb[1] = type; | ||
1067 | ioarcb->cdb[7] = (rcb_size >> 8) & 0xFF; | ||
1068 | ioarcb->cdb[8] = (rcb_size) & 0xFF; | ||
1069 | |||
1070 | ioarcb->data_transfer_length = cpu_to_le32(rcb_size); | ||
1071 | |||
1072 | ioadl[0].flags |= cpu_to_le32(IOADL_FLAGS_READ_LAST); | ||
1073 | ioadl[0].data_len = cpu_to_le32(rcb_size); | ||
1074 | ioadl[0].address = cpu_to_le32(dma); | ||
1075 | |||
1076 | cmd->cmd_done = cmd_done; | ||
1077 | return cmd; | ||
1078 | } | ||
1079 | |||
1080 | /** | ||
1081 | * pmcraid_send_hcam - Send an HCAM to IOA | ||
1082 | * @pinstance: ioa config struct | ||
1083 | * @type: HCAM type | ||
1084 | * | ||
1085 | * This function will send a Host Controlled Async command to IOA. | ||
1086 | * | ||
1087 | * Return value: | ||
1088 | * none | ||
1089 | */ | ||
1090 | static void pmcraid_send_hcam(struct pmcraid_instance *pinstance, u8 type) | ||
1091 | { | ||
1092 | struct pmcraid_cmd *cmd = pmcraid_init_hcam(pinstance, type); | ||
1093 | pmcraid_send_hcam_cmd(cmd); | ||
1094 | } | ||
1095 | |||
1096 | |||
1097 | /** | ||
1098 | * pmcraid_prepare_cancel_cmd - prepares a command block to abort another | ||
1099 | * | ||
1100 | * @cmd: pointer to cmd that is used as cancelling command | ||
1101 | * @cmd_to_cancel: pointer to the command that needs to be cancelled | ||
1102 | */ | ||
1103 | static void pmcraid_prepare_cancel_cmd( | ||
1104 | struct pmcraid_cmd *cmd, | ||
1105 | struct pmcraid_cmd *cmd_to_cancel | ||
1106 | ) | ||
1107 | { | ||
1108 | struct pmcraid_ioarcb *ioarcb = &cmd->ioa_cb->ioarcb; | ||
1109 | __be64 ioarcb_addr = cmd_to_cancel->ioa_cb->ioarcb.ioarcb_bus_addr; | ||
1110 | |||
1111 | /* Get the resource handle to where the command to be aborted has been | ||
1112 | * sent. | ||
1113 | */ | ||
1114 | ioarcb->resource_handle = cmd_to_cancel->ioa_cb->ioarcb.resource_handle; | ||
1115 | ioarcb->request_type = REQ_TYPE_IOACMD; | ||
1116 | memset(ioarcb->cdb, 0, PMCRAID_MAX_CDB_LEN); | ||
1117 | ioarcb->cdb[0] = PMCRAID_ABORT_CMD; | ||
1118 | |||
1119 | /* IOARCB address of the command to be cancelled is given in | ||
1120 | * cdb[2]..cdb[9] is Big-Endian format. Note that length bits in | ||
1121 | * IOARCB address are not masked. | ||
1122 | */ | ||
1123 | ioarcb_addr = cpu_to_be64(ioarcb_addr); | ||
1124 | memcpy(&(ioarcb->cdb[2]), &ioarcb_addr, sizeof(ioarcb_addr)); | ||
1125 | } | ||
1126 | |||
1127 | /** | ||
1128 | * pmcraid_cancel_hcam - sends ABORT task to abort a given HCAM | ||
1129 | * | ||
1130 | * @cmd: command to be used as cancelling command | ||
1131 | * @type: HCAM type | ||
1132 | * @cmd_done: op done function for the cancelling command | ||
1133 | */ | ||
1134 | static void pmcraid_cancel_hcam( | ||
1135 | struct pmcraid_cmd *cmd, | ||
1136 | u8 type, | ||
1137 | void (*cmd_done) (struct pmcraid_cmd *) | ||
1138 | ) | ||
1139 | { | ||
1140 | struct pmcraid_instance *pinstance; | ||
1141 | struct pmcraid_hostrcb *hcam; | ||
1142 | |||
1143 | pinstance = cmd->drv_inst; | ||
1144 | hcam = (type == PMCRAID_HCAM_CODE_LOG_DATA) ? | ||
1145 | &pinstance->ldn : &pinstance->ccn; | ||
1146 | |||
1147 | /* prepare for cancelling previous hcam command. If the HCAM is | ||
1148 | * currently not pending with IOA, we would have hcam->cmd as non-null | ||
1149 | */ | ||
1150 | if (hcam->cmd == NULL) | ||
1151 | return; | ||
1152 | |||
1153 | pmcraid_prepare_cancel_cmd(cmd, hcam->cmd); | ||
1154 | |||
1155 | /* writing to IOARRIN must be protected by host_lock, as mid-layer | ||
1156 | * schedule queuecommand while we are doing this | ||
1157 | */ | ||
1158 | pmcraid_send_cmd(cmd, cmd_done, | ||
1159 | PMCRAID_INTERNAL_TIMEOUT, | ||
1160 | pmcraid_timeout_handler); | ||
1161 | } | ||
1162 | |||
1163 | /** | ||
1164 | * pmcraid_cancel_ccn - cancel CCN HCAM already registered with IOA | ||
1165 | * | ||
1166 | * @cmd: command block to be used for cancelling the HCAM | ||
1167 | */ | ||
1168 | static void pmcraid_cancel_ccn(struct pmcraid_cmd *cmd) | ||
1169 | { | ||
1170 | pmcraid_info("response for Cancel LDN CDB[0] = %x ioasc = %x\n", | ||
1171 | cmd->ioa_cb->ioarcb.cdb[0], | ||
1172 | le32_to_cpu(cmd->ioa_cb->ioasa.ioasc)); | ||
1173 | |||
1174 | pmcraid_reinit_cmdblk(cmd); | ||
1175 | |||
1176 | pmcraid_cancel_hcam(cmd, | ||
1177 | PMCRAID_HCAM_CODE_CONFIG_CHANGE, | ||
1178 | pmcraid_ioa_shutdown); | ||
1179 | } | ||
1180 | |||
1181 | /** | ||
1182 | * pmcraid_cancel_ldn - cancel LDN HCAM already registered with IOA | ||
1183 | * | ||
1184 | * @cmd: command block to be used for cancelling the HCAM | ||
1185 | */ | ||
1186 | static void pmcraid_cancel_ldn(struct pmcraid_cmd *cmd) | ||
1187 | { | ||
1188 | pmcraid_cancel_hcam(cmd, | ||
1189 | PMCRAID_HCAM_CODE_LOG_DATA, | ||
1190 | pmcraid_cancel_ccn); | ||
1191 | } | ||
1192 | |||
1193 | /** | ||
1194 | * pmcraid_expose_resource - check if the resource can be exposed to OS | ||
1195 | * | ||
1196 | * @cfgte: pointer to configuration table entry of the resource | ||
1197 | * | ||
1198 | * Return value: | ||
1199 | * true if resource can be added to midlayer, false(0) otherwise | ||
1200 | */ | ||
1201 | static int pmcraid_expose_resource(struct pmcraid_config_table_entry *cfgte) | ||
1202 | { | ||
1203 | int retval = 0; | ||
1204 | |||
1205 | if (cfgte->resource_type == RES_TYPE_VSET) | ||
1206 | retval = ((cfgte->unique_flags1 & 0xFF) < 0xFE); | ||
1207 | else if (cfgte->resource_type == RES_TYPE_GSCSI) | ||
1208 | retval = (RES_BUS(cfgte->resource_address) != | ||
1209 | PMCRAID_VIRTUAL_ENCL_BUS_ID); | ||
1210 | return retval; | ||
1211 | } | ||
1212 | |||
1213 | /* attributes supported by pmcraid_event_family */ | ||
1214 | enum { | ||
1215 | PMCRAID_AEN_ATTR_UNSPEC, | ||
1216 | PMCRAID_AEN_ATTR_EVENT, | ||
1217 | __PMCRAID_AEN_ATTR_MAX, | ||
1218 | }; | ||
1219 | #define PMCRAID_AEN_ATTR_MAX (__PMCRAID_AEN_ATTR_MAX - 1) | ||
1220 | |||
1221 | /* commands supported by pmcraid_event_family */ | ||
1222 | enum { | ||
1223 | PMCRAID_AEN_CMD_UNSPEC, | ||
1224 | PMCRAID_AEN_CMD_EVENT, | ||
1225 | __PMCRAID_AEN_CMD_MAX, | ||
1226 | }; | ||
1227 | #define PMCRAID_AEN_CMD_MAX (__PMCRAID_AEN_CMD_MAX - 1) | ||
1228 | |||
1229 | static struct genl_family pmcraid_event_family = { | ||
1230 | .id = GENL_ID_GENERATE, | ||
1231 | .name = "pmcraid", | ||
1232 | .version = 1, | ||
1233 | .maxattr = PMCRAID_AEN_ATTR_MAX | ||
1234 | }; | ||
1235 | |||
1236 | /** | ||
1237 | * pmcraid_netlink_init - registers pmcraid_event_family | ||
1238 | * | ||
1239 | * Return value: | ||
1240 | * 0 if the pmcraid_event_family is successfully registered | ||
1241 | * with netlink generic, non-zero otherwise | ||
1242 | */ | ||
1243 | static int pmcraid_netlink_init(void) | ||
1244 | { | ||
1245 | int result; | ||
1246 | |||
1247 | result = genl_register_family(&pmcraid_event_family); | ||
1248 | |||
1249 | if (result) | ||
1250 | return result; | ||
1251 | |||
1252 | pmcraid_info("registered NETLINK GENERIC group: %d\n", | ||
1253 | pmcraid_event_family.id); | ||
1254 | |||
1255 | return result; | ||
1256 | } | ||
1257 | |||
1258 | /** | ||
1259 | * pmcraid_netlink_release - unregisters pmcraid_event_family | ||
1260 | * | ||
1261 | * Return value: | ||
1262 | * none | ||
1263 | */ | ||
1264 | static void pmcraid_netlink_release(void) | ||
1265 | { | ||
1266 | genl_unregister_family(&pmcraid_event_family); | ||
1267 | } | ||
1268 | |||
1269 | /** | ||
1270 | * pmcraid_notify_aen - sends event msg to user space application | ||
1271 | * @pinstance: pointer to adapter instance structure | ||
1272 | * @type: HCAM type | ||
1273 | * | ||
1274 | * Return value: | ||
1275 | * 0 if success, error value in case of any failure. | ||
1276 | */ | ||
1277 | static int pmcraid_notify_aen(struct pmcraid_instance *pinstance, u8 type) | ||
1278 | { | ||
1279 | struct sk_buff *skb; | ||
1280 | struct pmcraid_aen_msg *aen_msg; | ||
1281 | void *msg_header; | ||
1282 | int data_size, total_size; | ||
1283 | int result; | ||
1284 | |||
1285 | |||
1286 | if (type == PMCRAID_HCAM_CODE_LOG_DATA) { | ||
1287 | aen_msg = pinstance->ldn.msg; | ||
1288 | data_size = pinstance->ldn.hcam->data_len; | ||
1289 | } else { | ||
1290 | aen_msg = pinstance->ccn.msg; | ||
1291 | data_size = pinstance->ccn.hcam->data_len; | ||
1292 | } | ||
1293 | |||
1294 | data_size += sizeof(struct pmcraid_hcam_hdr); | ||
1295 | aen_msg->hostno = (pinstance->host->unique_id << 16 | | ||
1296 | MINOR(pinstance->cdev.dev)); | ||
1297 | aen_msg->length = data_size; | ||
1298 | data_size += sizeof(*aen_msg); | ||
1299 | |||
1300 | total_size = nla_total_size(data_size); | ||
1301 | skb = genlmsg_new(total_size, GFP_ATOMIC); | ||
1302 | |||
1303 | |||
1304 | if (!skb) { | ||
1305 | pmcraid_err("Failed to allocate aen data SKB of size: %x\n", | ||
1306 | total_size); | ||
1307 | return -ENOMEM; | ||
1308 | } | ||
1309 | |||
1310 | /* add the genetlink message header */ | ||
1311 | msg_header = genlmsg_put(skb, 0, 0, | ||
1312 | &pmcraid_event_family, 0, | ||
1313 | PMCRAID_AEN_CMD_EVENT); | ||
1314 | if (!msg_header) { | ||
1315 | pmcraid_err("failed to copy command details\n"); | ||
1316 | nlmsg_free(skb); | ||
1317 | return -ENOMEM; | ||
1318 | } | ||
1319 | |||
1320 | result = nla_put(skb, PMCRAID_AEN_ATTR_EVENT, data_size, aen_msg); | ||
1321 | |||
1322 | if (result) { | ||
1323 | pmcraid_err("failed to copy AEN attribute data \n"); | ||
1324 | nlmsg_free(skb); | ||
1325 | return -EINVAL; | ||
1326 | } | ||
1327 | |||
1328 | /* send genetlink multicast message to notify appplications */ | ||
1329 | result = genlmsg_end(skb, msg_header); | ||
1330 | |||
1331 | if (result < 0) { | ||
1332 | pmcraid_err("genlmsg_end failed\n"); | ||
1333 | nlmsg_free(skb); | ||
1334 | return result; | ||
1335 | } | ||
1336 | |||
1337 | result = | ||
1338 | genlmsg_multicast(skb, 0, pmcraid_event_family.id, GFP_ATOMIC); | ||
1339 | |||
1340 | /* If there are no listeners, genlmsg_multicast may return non-zero | ||
1341 | * value. | ||
1342 | */ | ||
1343 | if (result) | ||
1344 | pmcraid_info("failed to send %s event message %x!\n", | ||
1345 | type == PMCRAID_HCAM_CODE_LOG_DATA ? "LDN" : "CCN", | ||
1346 | result); | ||
1347 | return result; | ||
1348 | } | ||
1349 | |||
1350 | /** | ||
1351 | * pmcraid_handle_config_change - Handle a config change from the adapter | ||
1352 | * @pinstance: pointer to per adapter instance structure | ||
1353 | * | ||
1354 | * Return value: | ||
1355 | * none | ||
1356 | */ | ||
1357 | static void pmcraid_handle_config_change(struct pmcraid_instance *pinstance) | ||
1358 | { | ||
1359 | struct pmcraid_config_table_entry *cfg_entry; | ||
1360 | struct pmcraid_hcam_ccn *ccn_hcam; | ||
1361 | struct pmcraid_cmd *cmd; | ||
1362 | struct pmcraid_cmd *cfgcmd; | ||
1363 | struct pmcraid_resource_entry *res = NULL; | ||
1364 | u32 new_entry = 1; | ||
1365 | unsigned long lock_flags; | ||
1366 | unsigned long host_lock_flags; | ||
1367 | int rc; | ||
1368 | |||
1369 | ccn_hcam = (struct pmcraid_hcam_ccn *)pinstance->ccn.hcam; | ||
1370 | cfg_entry = &ccn_hcam->cfg_entry; | ||
1371 | |||
1372 | pmcraid_info | ||
1373 | ("CCN(%x): %x type: %x lost: %x flags: %x res: %x:%x:%x:%x\n", | ||
1374 | pinstance->ccn.hcam->ilid, | ||
1375 | pinstance->ccn.hcam->op_code, | ||
1376 | pinstance->ccn.hcam->notification_type, | ||
1377 | pinstance->ccn.hcam->notification_lost, | ||
1378 | pinstance->ccn.hcam->flags, | ||
1379 | pinstance->host->unique_id, | ||
1380 | RES_IS_VSET(*cfg_entry) ? PMCRAID_VSET_BUS_ID : | ||
1381 | (RES_IS_GSCSI(*cfg_entry) ? PMCRAID_PHYS_BUS_ID : | ||
1382 | RES_BUS(cfg_entry->resource_address)), | ||
1383 | RES_IS_VSET(*cfg_entry) ? cfg_entry->unique_flags1 : | ||
1384 | RES_TARGET(cfg_entry->resource_address), | ||
1385 | RES_LUN(cfg_entry->resource_address)); | ||
1386 | |||
1387 | |||
1388 | /* If this HCAM indicates a lost notification, read the config table */ | ||
1389 | if (pinstance->ccn.hcam->notification_lost) { | ||
1390 | cfgcmd = pmcraid_get_free_cmd(pinstance); | ||
1391 | if (cfgcmd) { | ||
1392 | pmcraid_info("lost CCN, reading config table\b"); | ||
1393 | pinstance->reinit_cfg_table = 1; | ||
1394 | pmcraid_querycfg(cfgcmd); | ||
1395 | } else { | ||
1396 | pmcraid_err("lost CCN, no free cmd for querycfg\n"); | ||
1397 | } | ||
1398 | goto out_notify_apps; | ||
1399 | } | ||
1400 | |||
1401 | /* If this resource is not going to be added to mid-layer, just notify | ||
1402 | * applications and return | ||
1403 | */ | ||
1404 | if (!pmcraid_expose_resource(cfg_entry)) | ||
1405 | goto out_notify_apps; | ||
1406 | |||
1407 | spin_lock_irqsave(&pinstance->resource_lock, lock_flags); | ||
1408 | list_for_each_entry(res, &pinstance->used_res_q, queue) { | ||
1409 | rc = memcmp(&res->cfg_entry.resource_address, | ||
1410 | &cfg_entry->resource_address, | ||
1411 | sizeof(cfg_entry->resource_address)); | ||
1412 | if (!rc) { | ||
1413 | new_entry = 0; | ||
1414 | break; | ||
1415 | } | ||
1416 | } | ||
1417 | |||
1418 | if (new_entry) { | ||
1419 | |||
1420 | /* If there are more number of resources than what driver can | ||
1421 | * manage, do not notify the applications about the CCN. Just | ||
1422 | * ignore this notifications and re-register the same HCAM | ||
1423 | */ | ||
1424 | if (list_empty(&pinstance->free_res_q)) { | ||
1425 | spin_unlock_irqrestore(&pinstance->resource_lock, | ||
1426 | lock_flags); | ||
1427 | pmcraid_err("too many resources attached\n"); | ||
1428 | spin_lock_irqsave(pinstance->host->host_lock, | ||
1429 | host_lock_flags); | ||
1430 | pmcraid_send_hcam(pinstance, | ||
1431 | PMCRAID_HCAM_CODE_CONFIG_CHANGE); | ||
1432 | spin_unlock_irqrestore(pinstance->host->host_lock, | ||
1433 | host_lock_flags); | ||
1434 | return; | ||
1435 | } | ||
1436 | |||
1437 | res = list_entry(pinstance->free_res_q.next, | ||
1438 | struct pmcraid_resource_entry, queue); | ||
1439 | |||
1440 | list_del(&res->queue); | ||
1441 | res->scsi_dev = NULL; | ||
1442 | res->reset_progress = 0; | ||
1443 | list_add_tail(&res->queue, &pinstance->used_res_q); | ||
1444 | } | ||
1445 | |||
1446 | memcpy(&res->cfg_entry, cfg_entry, | ||
1447 | sizeof(struct pmcraid_config_table_entry)); | ||
1448 | |||
1449 | if (pinstance->ccn.hcam->notification_type == | ||
1450 | NOTIFICATION_TYPE_ENTRY_DELETED) { | ||
1451 | if (res->scsi_dev) { | ||
1452 | res->change_detected = RES_CHANGE_DEL; | ||
1453 | res->cfg_entry.resource_handle = | ||
1454 | PMCRAID_INVALID_RES_HANDLE; | ||
1455 | schedule_work(&pinstance->worker_q); | ||
1456 | } else { | ||
1457 | /* This may be one of the non-exposed resources */ | ||
1458 | list_move_tail(&res->queue, &pinstance->free_res_q); | ||
1459 | } | ||
1460 | } else if (!res->scsi_dev) { | ||
1461 | res->change_detected = RES_CHANGE_ADD; | ||
1462 | schedule_work(&pinstance->worker_q); | ||
1463 | } | ||
1464 | spin_unlock_irqrestore(&pinstance->resource_lock, lock_flags); | ||
1465 | |||
1466 | out_notify_apps: | ||
1467 | |||
1468 | /* Notify configuration changes to registered applications.*/ | ||
1469 | if (!pmcraid_disable_aen) | ||
1470 | pmcraid_notify_aen(pinstance, PMCRAID_HCAM_CODE_CONFIG_CHANGE); | ||
1471 | |||
1472 | cmd = pmcraid_init_hcam(pinstance, PMCRAID_HCAM_CODE_CONFIG_CHANGE); | ||
1473 | if (cmd) | ||
1474 | pmcraid_send_hcam_cmd(cmd); | ||
1475 | } | ||
1476 | |||
1477 | /** | ||
1478 | * pmcraid_get_error_info - return error string for an ioasc | ||
1479 | * @ioasc: ioasc code | ||
1480 | * Return Value | ||
1481 | * none | ||
1482 | */ | ||
1483 | static struct pmcraid_ioasc_error *pmcraid_get_error_info(u32 ioasc) | ||
1484 | { | ||
1485 | int i; | ||
1486 | for (i = 0; i < ARRAY_SIZE(pmcraid_ioasc_error_table); i++) { | ||
1487 | if (pmcraid_ioasc_error_table[i].ioasc_code == ioasc) | ||
1488 | return &pmcraid_ioasc_error_table[i]; | ||
1489 | } | ||
1490 | return NULL; | ||
1491 | } | ||
1492 | |||
1493 | /** | ||
1494 | * pmcraid_ioasc_logger - log IOASC information based user-settings | ||
1495 | * @ioasc: ioasc code | ||
1496 | * @cmd: pointer to command that resulted in 'ioasc' | ||
1497 | */ | ||
1498 | void pmcraid_ioasc_logger(u32 ioasc, struct pmcraid_cmd *cmd) | ||
1499 | { | ||
1500 | struct pmcraid_ioasc_error *error_info = pmcraid_get_error_info(ioasc); | ||
1501 | |||
1502 | if (error_info == NULL || | ||
1503 | cmd->drv_inst->current_log_level < error_info->log_level) | ||
1504 | return; | ||
1505 | |||
1506 | /* log the error string */ | ||
1507 | pmcraid_err("cmd [%d] for resource %x failed with %x(%s)\n", | ||
1508 | cmd->ioa_cb->ioarcb.cdb[0], | ||
1509 | cmd->ioa_cb->ioarcb.resource_handle, | ||
1510 | le32_to_cpu(ioasc), error_info->error_string); | ||
1511 | } | ||
1512 | |||
1513 | /** | ||
1514 | * pmcraid_handle_error_log - Handle a config change (error log) from the IOA | ||
1515 | * | ||
1516 | * @pinstance: pointer to per adapter instance structure | ||
1517 | * | ||
1518 | * Return value: | ||
1519 | * none | ||
1520 | */ | ||
1521 | static void pmcraid_handle_error_log(struct pmcraid_instance *pinstance) | ||
1522 | { | ||
1523 | struct pmcraid_hcam_ldn *hcam_ldn; | ||
1524 | u32 ioasc; | ||
1525 | |||
1526 | hcam_ldn = (struct pmcraid_hcam_ldn *)pinstance->ldn.hcam; | ||
1527 | |||
1528 | pmcraid_info | ||
1529 | ("LDN(%x): %x type: %x lost: %x flags: %x overlay id: %x\n", | ||
1530 | pinstance->ldn.hcam->ilid, | ||
1531 | pinstance->ldn.hcam->op_code, | ||
1532 | pinstance->ldn.hcam->notification_type, | ||
1533 | pinstance->ldn.hcam->notification_lost, | ||
1534 | pinstance->ldn.hcam->flags, | ||
1535 | pinstance->ldn.hcam->overlay_id); | ||
1536 | |||
1537 | /* log only the errors, no need to log informational log entries */ | ||
1538 | if (pinstance->ldn.hcam->notification_type != | ||
1539 | NOTIFICATION_TYPE_ERROR_LOG) | ||
1540 | return; | ||
1541 | |||
1542 | if (pinstance->ldn.hcam->notification_lost == | ||
1543 | HOSTRCB_NOTIFICATIONS_LOST) | ||
1544 | dev_err(&pinstance->pdev->dev, "Error notifications lost\n"); | ||
1545 | |||
1546 | ioasc = le32_to_cpu(hcam_ldn->error_log.fd_ioasc); | ||
1547 | |||
1548 | if (ioasc == PMCRAID_IOASC_UA_BUS_WAS_RESET || | ||
1549 | ioasc == PMCRAID_IOASC_UA_BUS_WAS_RESET_BY_OTHER) { | ||
1550 | dev_err(&pinstance->pdev->dev, | ||
1551 | "UnitAttention due to IOA Bus Reset\n"); | ||
1552 | scsi_report_bus_reset( | ||
1553 | pinstance->host, | ||
1554 | RES_BUS(hcam_ldn->error_log.fd_ra)); | ||
1555 | } | ||
1556 | |||
1557 | return; | ||
1558 | } | ||
1559 | |||
1560 | /** | ||
1561 | * pmcraid_process_ccn - Op done function for a CCN. | ||
1562 | * @cmd: pointer to command struct | ||
1563 | * | ||
1564 | * This function is the op done function for a configuration | ||
1565 | * change notification | ||
1566 | * | ||
1567 | * Return value: | ||
1568 | * none | ||
1569 | */ | ||
1570 | static void pmcraid_process_ccn(struct pmcraid_cmd *cmd) | ||
1571 | { | ||
1572 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
1573 | u32 ioasc = le32_to_cpu(cmd->ioa_cb->ioasa.ioasc); | ||
1574 | unsigned long lock_flags; | ||
1575 | |||
1576 | pinstance->ccn.cmd = NULL; | ||
1577 | pmcraid_return_cmd(cmd); | ||
1578 | |||
1579 | /* If driver initiated IOA reset happened while this hcam was pending | ||
1580 | * with IOA, or IOA bringdown sequence is in progress, no need to | ||
1581 | * re-register the hcam | ||
1582 | */ | ||
1583 | if (ioasc == PMCRAID_IOASC_IOA_WAS_RESET || | ||
1584 | atomic_read(&pinstance->ccn.ignore) == 1) { | ||
1585 | return; | ||
1586 | } else if (ioasc) { | ||
1587 | dev_err(&pinstance->pdev->dev, | ||
1588 | "Host RCB (CCN) failed with IOASC: 0x%08X\n", ioasc); | ||
1589 | spin_lock_irqsave(pinstance->host->host_lock, lock_flags); | ||
1590 | pmcraid_send_hcam(pinstance, PMCRAID_HCAM_CODE_CONFIG_CHANGE); | ||
1591 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
1592 | } else { | ||
1593 | pmcraid_handle_config_change(pinstance); | ||
1594 | } | ||
1595 | } | ||
1596 | |||
1597 | /** | ||
1598 | * pmcraid_process_ldn - op done function for an LDN | ||
1599 | * @cmd: pointer to command block | ||
1600 | * | ||
1601 | * Return value | ||
1602 | * none | ||
1603 | */ | ||
1604 | static void pmcraid_initiate_reset(struct pmcraid_instance *); | ||
1605 | |||
1606 | static void pmcraid_process_ldn(struct pmcraid_cmd *cmd) | ||
1607 | { | ||
1608 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
1609 | struct pmcraid_hcam_ldn *ldn_hcam = | ||
1610 | (struct pmcraid_hcam_ldn *)pinstance->ldn.hcam; | ||
1611 | u32 ioasc = le32_to_cpu(cmd->ioa_cb->ioasa.ioasc); | ||
1612 | u32 fd_ioasc = le32_to_cpu(ldn_hcam->error_log.fd_ioasc); | ||
1613 | unsigned long lock_flags; | ||
1614 | |||
1615 | /* return the command block back to freepool */ | ||
1616 | pinstance->ldn.cmd = NULL; | ||
1617 | pmcraid_return_cmd(cmd); | ||
1618 | |||
1619 | /* If driver initiated IOA reset happened while this hcam was pending | ||
1620 | * with IOA, no need to re-register the hcam as reset engine will do it | ||
1621 | * once reset sequence is complete | ||
1622 | */ | ||
1623 | if (ioasc == PMCRAID_IOASC_IOA_WAS_RESET || | ||
1624 | atomic_read(&pinstance->ccn.ignore) == 1) { | ||
1625 | return; | ||
1626 | } else if (!ioasc) { | ||
1627 | pmcraid_handle_error_log(pinstance); | ||
1628 | if (fd_ioasc == PMCRAID_IOASC_NR_IOA_RESET_REQUIRED) { | ||
1629 | spin_lock_irqsave(pinstance->host->host_lock, | ||
1630 | lock_flags); | ||
1631 | pmcraid_initiate_reset(pinstance); | ||
1632 | spin_unlock_irqrestore(pinstance->host->host_lock, | ||
1633 | lock_flags); | ||
1634 | return; | ||
1635 | } | ||
1636 | } else { | ||
1637 | dev_err(&pinstance->pdev->dev, | ||
1638 | "Host RCB(LDN) failed with IOASC: 0x%08X\n", ioasc); | ||
1639 | } | ||
1640 | /* send netlink message for HCAM notification if enabled */ | ||
1641 | if (!pmcraid_disable_aen) | ||
1642 | pmcraid_notify_aen(pinstance, PMCRAID_HCAM_CODE_LOG_DATA); | ||
1643 | |||
1644 | cmd = pmcraid_init_hcam(pinstance, PMCRAID_HCAM_CODE_LOG_DATA); | ||
1645 | if (cmd) | ||
1646 | pmcraid_send_hcam_cmd(cmd); | ||
1647 | } | ||
1648 | |||
1649 | /** | ||
1650 | * pmcraid_register_hcams - register HCAMs for CCN and LDN | ||
1651 | * | ||
1652 | * @pinstance: pointer per adapter instance structure | ||
1653 | * | ||
1654 | * Return Value | ||
1655 | * none | ||
1656 | */ | ||
1657 | static void pmcraid_register_hcams(struct pmcraid_instance *pinstance) | ||
1658 | { | ||
1659 | pmcraid_send_hcam(pinstance, PMCRAID_HCAM_CODE_CONFIG_CHANGE); | ||
1660 | pmcraid_send_hcam(pinstance, PMCRAID_HCAM_CODE_LOG_DATA); | ||
1661 | } | ||
1662 | |||
1663 | /** | ||
1664 | * pmcraid_unregister_hcams - cancel HCAMs registered already | ||
1665 | * @cmd: pointer to command used as part of reset sequence | ||
1666 | */ | ||
1667 | static void pmcraid_unregister_hcams(struct pmcraid_cmd *cmd) | ||
1668 | { | ||
1669 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
1670 | |||
1671 | /* During IOA bringdown, HCAM gets fired and tasklet proceeds with | ||
1672 | * handling hcam response though it is not necessary. In order to | ||
1673 | * prevent this, set 'ignore', so that bring-down sequence doesn't | ||
1674 | * re-send any more hcams | ||
1675 | */ | ||
1676 | atomic_set(&pinstance->ccn.ignore, 1); | ||
1677 | atomic_set(&pinstance->ldn.ignore, 1); | ||
1678 | |||
1679 | /* If adapter reset was forced as part of runtime reset sequence, | ||
1680 | * start the reset sequence. | ||
1681 | */ | ||
1682 | if (pinstance->force_ioa_reset && !pinstance->ioa_bringdown) { | ||
1683 | pinstance->force_ioa_reset = 0; | ||
1684 | pinstance->ioa_state = IOA_STATE_IN_RESET_ALERT; | ||
1685 | pmcraid_reset_alert(cmd); | ||
1686 | return; | ||
1687 | } | ||
1688 | |||
1689 | /* Driver tries to cancel HCAMs by sending ABORT TASK for each HCAM | ||
1690 | * one after the other. So CCN cancellation will be triggered by | ||
1691 | * pmcraid_cancel_ldn itself. | ||
1692 | */ | ||
1693 | pmcraid_cancel_ldn(cmd); | ||
1694 | } | ||
1695 | |||
1696 | /** | ||
1697 | * pmcraid_reset_enable_ioa - re-enable IOA after a hard reset | ||
1698 | * @pinstance: pointer to adapter instance structure | ||
1699 | * Return Value | ||
1700 | * 1 if TRANSITION_TO_OPERATIONAL is active, otherwise 0 | ||
1701 | */ | ||
1702 | static void pmcraid_reinit_buffers(struct pmcraid_instance *); | ||
1703 | |||
1704 | static int pmcraid_reset_enable_ioa(struct pmcraid_instance *pinstance) | ||
1705 | { | ||
1706 | u32 intrs; | ||
1707 | |||
1708 | pmcraid_reinit_buffers(pinstance); | ||
1709 | intrs = pmcraid_read_interrupts(pinstance); | ||
1710 | |||
1711 | pmcraid_enable_interrupts(pinstance, PMCRAID_PCI_INTERRUPTS); | ||
1712 | |||
1713 | if (intrs & INTRS_TRANSITION_TO_OPERATIONAL) { | ||
1714 | iowrite32(INTRS_TRANSITION_TO_OPERATIONAL, | ||
1715 | pinstance->int_regs.ioa_host_interrupt_mask_reg); | ||
1716 | iowrite32(INTRS_TRANSITION_TO_OPERATIONAL, | ||
1717 | pinstance->int_regs.ioa_host_interrupt_clr_reg); | ||
1718 | return 1; | ||
1719 | } else { | ||
1720 | return 0; | ||
1721 | } | ||
1722 | } | ||
1723 | |||
1724 | /** | ||
1725 | * pmcraid_soft_reset - performs a soft reset and makes IOA become ready | ||
1726 | * @cmd : pointer to reset command block | ||
1727 | * | ||
1728 | * Return Value | ||
1729 | * none | ||
1730 | */ | ||
1731 | static void pmcraid_soft_reset(struct pmcraid_cmd *cmd) | ||
1732 | { | ||
1733 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
1734 | u32 int_reg; | ||
1735 | u32 doorbell; | ||
1736 | |||
1737 | /* There will be an interrupt when Transition to Operational bit is | ||
1738 | * set so tasklet would execute next reset task. The timeout handler | ||
1739 | * would re-initiate a reset | ||
1740 | */ | ||
1741 | cmd->cmd_done = pmcraid_ioa_reset; | ||
1742 | cmd->timer.data = (unsigned long)cmd; | ||
1743 | cmd->timer.expires = jiffies + | ||
1744 | msecs_to_jiffies(PMCRAID_TRANSOP_TIMEOUT); | ||
1745 | cmd->timer.function = (void (*)(unsigned long))pmcraid_timeout_handler; | ||
1746 | |||
1747 | if (!timer_pending(&cmd->timer)) | ||
1748 | add_timer(&cmd->timer); | ||
1749 | |||
1750 | /* Enable destructive diagnostics on IOA if it is not yet in | ||
1751 | * operational state | ||
1752 | */ | ||
1753 | doorbell = DOORBELL_RUNTIME_RESET | | ||
1754 | DOORBELL_ENABLE_DESTRUCTIVE_DIAGS; | ||
1755 | |||
1756 | iowrite32(doorbell, pinstance->int_regs.host_ioa_interrupt_reg); | ||
1757 | int_reg = ioread32(pinstance->int_regs.ioa_host_interrupt_reg); | ||
1758 | pmcraid_info("Waiting for IOA to become operational %x:%x\n", | ||
1759 | ioread32(pinstance->int_regs.host_ioa_interrupt_reg), | ||
1760 | int_reg); | ||
1761 | } | ||
1762 | |||
1763 | /** | ||
1764 | * pmcraid_get_dump - retrieves IOA dump in case of Unit Check interrupt | ||
1765 | * | ||
1766 | * @pinstance: pointer to adapter instance structure | ||
1767 | * | ||
1768 | * Return Value | ||
1769 | * none | ||
1770 | */ | ||
1771 | static void pmcraid_get_dump(struct pmcraid_instance *pinstance) | ||
1772 | { | ||
1773 | pmcraid_info("%s is not yet implemented\n", __func__); | ||
1774 | } | ||
1775 | |||
1776 | /** | ||
1777 | * pmcraid_fail_outstanding_cmds - Fails all outstanding ops. | ||
1778 | * @pinstance: pointer to adapter instance structure | ||
1779 | * | ||
1780 | * This function fails all outstanding ops. If they are submitted to IOA | ||
1781 | * already, it sends cancel all messages if IOA is still accepting IOARCBs, | ||
1782 | * otherwise just completes the commands and returns the cmd blocks to free | ||
1783 | * pool. | ||
1784 | * | ||
1785 | * Return value: | ||
1786 | * none | ||
1787 | */ | ||
1788 | static void pmcraid_fail_outstanding_cmds(struct pmcraid_instance *pinstance) | ||
1789 | { | ||
1790 | struct pmcraid_cmd *cmd, *temp; | ||
1791 | unsigned long lock_flags; | ||
1792 | |||
1793 | /* pending command list is protected by pending_pool_lock. Its | ||
1794 | * traversal must be done as within this lock | ||
1795 | */ | ||
1796 | spin_lock_irqsave(&pinstance->pending_pool_lock, lock_flags); | ||
1797 | list_for_each_entry_safe(cmd, temp, &pinstance->pending_cmd_pool, | ||
1798 | free_list) { | ||
1799 | list_del(&cmd->free_list); | ||
1800 | spin_unlock_irqrestore(&pinstance->pending_pool_lock, | ||
1801 | lock_flags); | ||
1802 | cmd->ioa_cb->ioasa.ioasc = | ||
1803 | cpu_to_le32(PMCRAID_IOASC_IOA_WAS_RESET); | ||
1804 | cmd->ioa_cb->ioasa.ilid = | ||
1805 | cpu_to_be32(PMCRAID_DRIVER_ILID); | ||
1806 | |||
1807 | /* In case the command timer is still running */ | ||
1808 | del_timer(&cmd->timer); | ||
1809 | |||
1810 | /* If this is an IO command, complete it by invoking scsi_done | ||
1811 | * function. If this is one of the internal commands other | ||
1812 | * than pmcraid_ioa_reset and HCAM commands invoke cmd_done to | ||
1813 | * complete it | ||
1814 | */ | ||
1815 | if (cmd->scsi_cmd) { | ||
1816 | |||
1817 | struct scsi_cmnd *scsi_cmd = cmd->scsi_cmd; | ||
1818 | __le32 resp = cmd->ioa_cb->ioarcb.response_handle; | ||
1819 | |||
1820 | scsi_cmd->result |= DID_ERROR << 16; | ||
1821 | |||
1822 | scsi_dma_unmap(scsi_cmd); | ||
1823 | pmcraid_return_cmd(cmd); | ||
1824 | |||
1825 | |||
1826 | pmcraid_info("failing(%d) CDB[0] = %x result: %x\n", | ||
1827 | le32_to_cpu(resp) >> 2, | ||
1828 | cmd->ioa_cb->ioarcb.cdb[0], | ||
1829 | scsi_cmd->result); | ||
1830 | scsi_cmd->scsi_done(scsi_cmd); | ||
1831 | } else if (cmd->cmd_done == pmcraid_internal_done || | ||
1832 | cmd->cmd_done == pmcraid_erp_done) { | ||
1833 | cmd->cmd_done(cmd); | ||
1834 | } else if (cmd->cmd_done != pmcraid_ioa_reset) { | ||
1835 | pmcraid_return_cmd(cmd); | ||
1836 | } | ||
1837 | |||
1838 | atomic_dec(&pinstance->outstanding_cmds); | ||
1839 | spin_lock_irqsave(&pinstance->pending_pool_lock, lock_flags); | ||
1840 | } | ||
1841 | |||
1842 | spin_unlock_irqrestore(&pinstance->pending_pool_lock, lock_flags); | ||
1843 | } | ||
1844 | |||
1845 | /** | ||
1846 | * pmcraid_ioa_reset - Implementation of IOA reset logic | ||
1847 | * | ||
1848 | * @cmd: pointer to the cmd block to be used for entire reset process | ||
1849 | * | ||
1850 | * This function executes most of the steps required for IOA reset. This gets | ||
1851 | * called by user threads (modprobe/insmod/rmmod) timer, tasklet and midlayer's | ||
1852 | * 'eh_' thread. Access to variables used for controling the reset sequence is | ||
1853 | * synchronized using host lock. Various functions called during reset process | ||
1854 | * would make use of a single command block, pointer to which is also stored in | ||
1855 | * adapter instance structure. | ||
1856 | * | ||
1857 | * Return Value | ||
1858 | * None | ||
1859 | */ | ||
1860 | static void pmcraid_ioa_reset(struct pmcraid_cmd *cmd) | ||
1861 | { | ||
1862 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
1863 | u8 reset_complete = 0; | ||
1864 | |||
1865 | pinstance->ioa_reset_in_progress = 1; | ||
1866 | |||
1867 | if (pinstance->reset_cmd != cmd) { | ||
1868 | pmcraid_err("reset is called with different command block\n"); | ||
1869 | pinstance->reset_cmd = cmd; | ||
1870 | } | ||
1871 | |||
1872 | pmcraid_info("reset_engine: state = %d, command = %p\n", | ||
1873 | pinstance->ioa_state, cmd); | ||
1874 | |||
1875 | switch (pinstance->ioa_state) { | ||
1876 | |||
1877 | case IOA_STATE_DEAD: | ||
1878 | /* If IOA is offline, whatever may be the reset reason, just | ||
1879 | * return. callers might be waiting on the reset wait_q, wake | ||
1880 | * up them | ||
1881 | */ | ||
1882 | pmcraid_err("IOA is offline no reset is possible\n"); | ||
1883 | reset_complete = 1; | ||
1884 | break; | ||
1885 | |||
1886 | case IOA_STATE_IN_BRINGDOWN: | ||
1887 | /* we enter here, once ioa shutdown command is processed by IOA | ||
1888 | * Alert IOA for a possible reset. If reset alert fails, IOA | ||
1889 | * goes through hard-reset | ||
1890 | */ | ||
1891 | pmcraid_disable_interrupts(pinstance, ~0); | ||
1892 | pinstance->ioa_state = IOA_STATE_IN_RESET_ALERT; | ||
1893 | pmcraid_reset_alert(cmd); | ||
1894 | break; | ||
1895 | |||
1896 | case IOA_STATE_UNKNOWN: | ||
1897 | /* We may be called during probe or resume. Some pre-processing | ||
1898 | * is required for prior to reset | ||
1899 | */ | ||
1900 | scsi_block_requests(pinstance->host); | ||
1901 | |||
1902 | /* If asked to reset while IOA was processing responses or | ||
1903 | * there are any error responses then IOA may require | ||
1904 | * hard-reset. | ||
1905 | */ | ||
1906 | if (pinstance->ioa_hard_reset == 0) { | ||
1907 | if (ioread32(pinstance->ioa_status) & | ||
1908 | INTRS_TRANSITION_TO_OPERATIONAL) { | ||
1909 | pmcraid_info("sticky bit set, bring-up\n"); | ||
1910 | pinstance->ioa_state = IOA_STATE_IN_BRINGUP; | ||
1911 | pmcraid_reinit_cmdblk(cmd); | ||
1912 | pmcraid_identify_hrrq(cmd); | ||
1913 | } else { | ||
1914 | pinstance->ioa_state = IOA_STATE_IN_SOFT_RESET; | ||
1915 | pmcraid_soft_reset(cmd); | ||
1916 | } | ||
1917 | } else { | ||
1918 | /* Alert IOA of a possible reset and wait for critical | ||
1919 | * operation in progress bit to reset | ||
1920 | */ | ||
1921 | pinstance->ioa_state = IOA_STATE_IN_RESET_ALERT; | ||
1922 | pmcraid_reset_alert(cmd); | ||
1923 | } | ||
1924 | break; | ||
1925 | |||
1926 | case IOA_STATE_IN_RESET_ALERT: | ||
1927 | /* If critical operation in progress bit is reset or wait gets | ||
1928 | * timed out, reset proceeds with starting BIST on the IOA. | ||
1929 | * pmcraid_ioa_hard_reset keeps a count of reset attempts. If | ||
1930 | * they are 3 or more, reset engine marks IOA dead and returns | ||
1931 | */ | ||
1932 | pinstance->ioa_state = IOA_STATE_IN_HARD_RESET; | ||
1933 | pmcraid_start_bist(cmd); | ||
1934 | break; | ||
1935 | |||
1936 | case IOA_STATE_IN_HARD_RESET: | ||
1937 | pinstance->ioa_reset_attempts++; | ||
1938 | |||
1939 | /* retry reset if we haven't reached maximum allowed limit */ | ||
1940 | if (pinstance->ioa_reset_attempts > PMCRAID_RESET_ATTEMPTS) { | ||
1941 | pinstance->ioa_reset_attempts = 0; | ||
1942 | pmcraid_err("IOA didn't respond marking it as dead\n"); | ||
1943 | pinstance->ioa_state = IOA_STATE_DEAD; | ||
1944 | reset_complete = 1; | ||
1945 | break; | ||
1946 | } | ||
1947 | |||
1948 | /* Once either bist or pci reset is done, restore PCI config | ||
1949 | * space. If this fails, proceed with hard reset again | ||
1950 | */ | ||
1951 | |||
1952 | if (pci_restore_state(pinstance->pdev)) { | ||
1953 | pmcraid_info("config-space error resetting again\n"); | ||
1954 | pinstance->ioa_state = IOA_STATE_IN_RESET_ALERT; | ||
1955 | pmcraid_reset_alert(cmd); | ||
1956 | break; | ||
1957 | } | ||
1958 | |||
1959 | /* fail all pending commands */ | ||
1960 | pmcraid_fail_outstanding_cmds(pinstance); | ||
1961 | |||
1962 | /* check if unit check is active, if so extract dump */ | ||
1963 | if (pinstance->ioa_unit_check) { | ||
1964 | pmcraid_info("unit check is active\n"); | ||
1965 | pinstance->ioa_unit_check = 0; | ||
1966 | pmcraid_get_dump(pinstance); | ||
1967 | pinstance->ioa_reset_attempts--; | ||
1968 | pinstance->ioa_state = IOA_STATE_IN_RESET_ALERT; | ||
1969 | pmcraid_reset_alert(cmd); | ||
1970 | break; | ||
1971 | } | ||
1972 | |||
1973 | /* if the reset reason is to bring-down the ioa, we might be | ||
1974 | * done with the reset restore pci_config_space and complete | ||
1975 | * the reset | ||
1976 | */ | ||
1977 | if (pinstance->ioa_bringdown) { | ||
1978 | pmcraid_info("bringing down the adapter\n"); | ||
1979 | pinstance->ioa_shutdown_type = SHUTDOWN_NONE; | ||
1980 | pinstance->ioa_bringdown = 0; | ||
1981 | pinstance->ioa_state = IOA_STATE_UNKNOWN; | ||
1982 | reset_complete = 1; | ||
1983 | } else { | ||
1984 | /* bring-up IOA, so proceed with soft reset | ||
1985 | * Reinitialize hrrq_buffers and their indices also | ||
1986 | * enable interrupts after a pci_restore_state | ||
1987 | */ | ||
1988 | if (pmcraid_reset_enable_ioa(pinstance)) { | ||
1989 | pinstance->ioa_state = IOA_STATE_IN_BRINGUP; | ||
1990 | pmcraid_info("bringing up the adapter\n"); | ||
1991 | pmcraid_reinit_cmdblk(cmd); | ||
1992 | pmcraid_identify_hrrq(cmd); | ||
1993 | } else { | ||
1994 | pinstance->ioa_state = IOA_STATE_IN_SOFT_RESET; | ||
1995 | pmcraid_soft_reset(cmd); | ||
1996 | } | ||
1997 | } | ||
1998 | break; | ||
1999 | |||
2000 | case IOA_STATE_IN_SOFT_RESET: | ||
2001 | /* TRANSITION TO OPERATIONAL is on so start initialization | ||
2002 | * sequence | ||
2003 | */ | ||
2004 | pmcraid_info("In softreset proceeding with bring-up\n"); | ||
2005 | pinstance->ioa_state = IOA_STATE_IN_BRINGUP; | ||
2006 | |||
2007 | /* Initialization commands start with HRRQ identification. From | ||
2008 | * now on tasklet completes most of the commands as IOA is up | ||
2009 | * and intrs are enabled | ||
2010 | */ | ||
2011 | pmcraid_identify_hrrq(cmd); | ||
2012 | break; | ||
2013 | |||
2014 | case IOA_STATE_IN_BRINGUP: | ||
2015 | /* we are done with bringing up of IOA, change the ioa_state to | ||
2016 | * operational and wake up any waiters | ||
2017 | */ | ||
2018 | pinstance->ioa_state = IOA_STATE_OPERATIONAL; | ||
2019 | reset_complete = 1; | ||
2020 | break; | ||
2021 | |||
2022 | case IOA_STATE_OPERATIONAL: | ||
2023 | default: | ||
2024 | /* When IOA is operational and a reset is requested, check for | ||
2025 | * the reset reason. If reset is to bring down IOA, unregister | ||
2026 | * HCAMs and initiate shutdown; if adapter reset is forced then | ||
2027 | * restart reset sequence again | ||
2028 | */ | ||
2029 | if (pinstance->ioa_shutdown_type == SHUTDOWN_NONE && | ||
2030 | pinstance->force_ioa_reset == 0) { | ||
2031 | reset_complete = 1; | ||
2032 | } else { | ||
2033 | if (pinstance->ioa_shutdown_type != SHUTDOWN_NONE) | ||
2034 | pinstance->ioa_state = IOA_STATE_IN_BRINGDOWN; | ||
2035 | pmcraid_reinit_cmdblk(cmd); | ||
2036 | pmcraid_unregister_hcams(cmd); | ||
2037 | } | ||
2038 | break; | ||
2039 | } | ||
2040 | |||
2041 | /* reset will be completed if ioa_state is either DEAD or UNKNOWN or | ||
2042 | * OPERATIONAL. Reset all control variables used during reset, wake up | ||
2043 | * any waiting threads and let the SCSI mid-layer send commands. Note | ||
2044 | * that host_lock must be held before invoking scsi_report_bus_reset. | ||
2045 | */ | ||
2046 | if (reset_complete) { | ||
2047 | pinstance->ioa_reset_in_progress = 0; | ||
2048 | pinstance->ioa_reset_attempts = 0; | ||
2049 | pinstance->reset_cmd = NULL; | ||
2050 | pinstance->ioa_shutdown_type = SHUTDOWN_NONE; | ||
2051 | pinstance->ioa_bringdown = 0; | ||
2052 | pmcraid_return_cmd(cmd); | ||
2053 | |||
2054 | /* If target state is to bring up the adapter, proceed with | ||
2055 | * hcam registration and resource exposure to mid-layer. | ||
2056 | */ | ||
2057 | if (pinstance->ioa_state == IOA_STATE_OPERATIONAL) | ||
2058 | pmcraid_register_hcams(pinstance); | ||
2059 | |||
2060 | wake_up_all(&pinstance->reset_wait_q); | ||
2061 | } | ||
2062 | |||
2063 | return; | ||
2064 | } | ||
2065 | |||
2066 | /** | ||
2067 | * pmcraid_initiate_reset - initiates reset sequence. This is called from | ||
2068 | * ISR/tasklet during error interrupts including IOA unit check. If reset | ||
2069 | * is already in progress, it just returns, otherwise initiates IOA reset | ||
2070 | * to bring IOA up to operational state. | ||
2071 | * | ||
2072 | * @pinstance: pointer to adapter instance structure | ||
2073 | * | ||
2074 | * Return value | ||
2075 | * none | ||
2076 | */ | ||
2077 | static void pmcraid_initiate_reset(struct pmcraid_instance *pinstance) | ||
2078 | { | ||
2079 | struct pmcraid_cmd *cmd; | ||
2080 | |||
2081 | /* If the reset is already in progress, just return, otherwise start | ||
2082 | * reset sequence and return | ||
2083 | */ | ||
2084 | if (!pinstance->ioa_reset_in_progress) { | ||
2085 | scsi_block_requests(pinstance->host); | ||
2086 | cmd = pmcraid_get_free_cmd(pinstance); | ||
2087 | |||
2088 | if (cmd == NULL) { | ||
2089 | pmcraid_err("no cmnd blocks for initiate_reset\n"); | ||
2090 | return; | ||
2091 | } | ||
2092 | |||
2093 | pinstance->ioa_shutdown_type = SHUTDOWN_NONE; | ||
2094 | pinstance->reset_cmd = cmd; | ||
2095 | pinstance->force_ioa_reset = 1; | ||
2096 | pmcraid_ioa_reset(cmd); | ||
2097 | } | ||
2098 | } | ||
2099 | |||
2100 | /** | ||
2101 | * pmcraid_reset_reload - utility routine for doing IOA reset either to bringup | ||
2102 | * or bringdown IOA | ||
2103 | * @pinstance: pointer adapter instance structure | ||
2104 | * @shutdown_type: shutdown type to be used NONE, NORMAL or ABRREV | ||
2105 | * @target_state: expected target state after reset | ||
2106 | * | ||
2107 | * Note: This command initiates reset and waits for its completion. Hence this | ||
2108 | * should not be called from isr/timer/tasklet functions (timeout handlers, | ||
2109 | * error response handlers and interrupt handlers). | ||
2110 | * | ||
2111 | * Return Value | ||
2112 | * 1 in case ioa_state is not target_state, 0 otherwise. | ||
2113 | */ | ||
2114 | static int pmcraid_reset_reload( | ||
2115 | struct pmcraid_instance *pinstance, | ||
2116 | u8 shutdown_type, | ||
2117 | u8 target_state | ||
2118 | ) | ||
2119 | { | ||
2120 | struct pmcraid_cmd *reset_cmd = NULL; | ||
2121 | unsigned long lock_flags; | ||
2122 | int reset = 1; | ||
2123 | |||
2124 | spin_lock_irqsave(pinstance->host->host_lock, lock_flags); | ||
2125 | |||
2126 | if (pinstance->ioa_reset_in_progress) { | ||
2127 | pmcraid_info("reset_reload: reset is already in progress\n"); | ||
2128 | |||
2129 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
2130 | |||
2131 | wait_event(pinstance->reset_wait_q, | ||
2132 | !pinstance->ioa_reset_in_progress); | ||
2133 | |||
2134 | spin_lock_irqsave(pinstance->host->host_lock, lock_flags); | ||
2135 | |||
2136 | if (pinstance->ioa_state == IOA_STATE_DEAD) { | ||
2137 | spin_unlock_irqrestore(pinstance->host->host_lock, | ||
2138 | lock_flags); | ||
2139 | pmcraid_info("reset_reload: IOA is dead\n"); | ||
2140 | return reset; | ||
2141 | } else if (pinstance->ioa_state == target_state) { | ||
2142 | reset = 0; | ||
2143 | } | ||
2144 | } | ||
2145 | |||
2146 | if (reset) { | ||
2147 | pmcraid_info("reset_reload: proceeding with reset\n"); | ||
2148 | scsi_block_requests(pinstance->host); | ||
2149 | reset_cmd = pmcraid_get_free_cmd(pinstance); | ||
2150 | |||
2151 | if (reset_cmd == NULL) { | ||
2152 | pmcraid_err("no free cmnd for reset_reload\n"); | ||
2153 | spin_unlock_irqrestore(pinstance->host->host_lock, | ||
2154 | lock_flags); | ||
2155 | return reset; | ||
2156 | } | ||
2157 | |||
2158 | if (shutdown_type == SHUTDOWN_NORMAL) | ||
2159 | pinstance->ioa_bringdown = 1; | ||
2160 | |||
2161 | pinstance->ioa_shutdown_type = shutdown_type; | ||
2162 | pinstance->reset_cmd = reset_cmd; | ||
2163 | pinstance->force_ioa_reset = reset; | ||
2164 | pmcraid_info("reset_reload: initiating reset\n"); | ||
2165 | pmcraid_ioa_reset(reset_cmd); | ||
2166 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
2167 | pmcraid_info("reset_reload: waiting for reset to complete\n"); | ||
2168 | wait_event(pinstance->reset_wait_q, | ||
2169 | !pinstance->ioa_reset_in_progress); | ||
2170 | |||
2171 | pmcraid_info("reset_reload: reset is complete !! \n"); | ||
2172 | scsi_unblock_requests(pinstance->host); | ||
2173 | if (pinstance->ioa_state == target_state) | ||
2174 | reset = 0; | ||
2175 | } | ||
2176 | |||
2177 | return reset; | ||
2178 | } | ||
2179 | |||
2180 | /** | ||
2181 | * pmcraid_reset_bringdown - wrapper over pmcraid_reset_reload to bringdown IOA | ||
2182 | * | ||
2183 | * @pinstance: pointer to adapter instance structure | ||
2184 | * | ||
2185 | * Return Value | ||
2186 | * whatever is returned from pmcraid_reset_reload | ||
2187 | */ | ||
2188 | static int pmcraid_reset_bringdown(struct pmcraid_instance *pinstance) | ||
2189 | { | ||
2190 | return pmcraid_reset_reload(pinstance, | ||
2191 | SHUTDOWN_NORMAL, | ||
2192 | IOA_STATE_UNKNOWN); | ||
2193 | } | ||
2194 | |||
2195 | /** | ||
2196 | * pmcraid_reset_bringup - wrapper over pmcraid_reset_reload to bring up IOA | ||
2197 | * | ||
2198 | * @pinstance: pointer to adapter instance structure | ||
2199 | * | ||
2200 | * Return Value | ||
2201 | * whatever is returned from pmcraid_reset_reload | ||
2202 | */ | ||
2203 | static int pmcraid_reset_bringup(struct pmcraid_instance *pinstance) | ||
2204 | { | ||
2205 | return pmcraid_reset_reload(pinstance, | ||
2206 | SHUTDOWN_NONE, | ||
2207 | IOA_STATE_OPERATIONAL); | ||
2208 | } | ||
2209 | |||
2210 | /** | ||
2211 | * pmcraid_request_sense - Send request sense to a device | ||
2212 | * @cmd: pmcraid command struct | ||
2213 | * | ||
2214 | * This function sends a request sense to a device as a result of a check | ||
2215 | * condition. This method re-uses the same command block that failed earlier. | ||
2216 | */ | ||
2217 | static void pmcraid_request_sense(struct pmcraid_cmd *cmd) | ||
2218 | { | ||
2219 | struct pmcraid_ioarcb *ioarcb = &cmd->ioa_cb->ioarcb; | ||
2220 | struct pmcraid_ioadl_desc *ioadl = ioarcb->add_data.u.ioadl; | ||
2221 | |||
2222 | /* allocate DMAable memory for sense buffers */ | ||
2223 | cmd->sense_buffer = pci_alloc_consistent(cmd->drv_inst->pdev, | ||
2224 | SCSI_SENSE_BUFFERSIZE, | ||
2225 | &cmd->sense_buffer_dma); | ||
2226 | |||
2227 | if (cmd->sense_buffer == NULL) { | ||
2228 | pmcraid_err | ||
2229 | ("couldn't allocate sense buffer for request sense\n"); | ||
2230 | pmcraid_erp_done(cmd); | ||
2231 | return; | ||
2232 | } | ||
2233 | |||
2234 | /* re-use the command block */ | ||
2235 | memset(&cmd->ioa_cb->ioasa, 0, sizeof(struct pmcraid_ioasa)); | ||
2236 | memset(ioarcb->cdb, 0, PMCRAID_MAX_CDB_LEN); | ||
2237 | ioarcb->request_flags0 = (SYNC_COMPLETE | | ||
2238 | NO_LINK_DESCS | | ||
2239 | INHIBIT_UL_CHECK); | ||
2240 | ioarcb->request_type = REQ_TYPE_SCSI; | ||
2241 | ioarcb->cdb[0] = REQUEST_SENSE; | ||
2242 | ioarcb->cdb[4] = SCSI_SENSE_BUFFERSIZE; | ||
2243 | |||
2244 | ioarcb->ioadl_bus_addr = cpu_to_le64((cmd->ioa_cb_bus_addr) + | ||
2245 | offsetof(struct pmcraid_ioarcb, | ||
2246 | add_data.u.ioadl[0])); | ||
2247 | ioarcb->ioadl_length = cpu_to_le32(sizeof(struct pmcraid_ioadl_desc)); | ||
2248 | |||
2249 | ioarcb->data_transfer_length = cpu_to_le32(SCSI_SENSE_BUFFERSIZE); | ||
2250 | |||
2251 | ioadl->address = cpu_to_le64(cmd->sense_buffer_dma); | ||
2252 | ioadl->data_len = cpu_to_le32(SCSI_SENSE_BUFFERSIZE); | ||
2253 | ioadl->flags = cpu_to_le32(IOADL_FLAGS_LAST_DESC); | ||
2254 | |||
2255 | /* request sense might be called as part of error response processing | ||
2256 | * which runs in tasklets context. It is possible that mid-layer might | ||
2257 | * schedule queuecommand during this time, hence, writting to IOARRIN | ||
2258 | * must be protect by host_lock | ||
2259 | */ | ||
2260 | pmcraid_send_cmd(cmd, pmcraid_erp_done, | ||
2261 | PMCRAID_REQUEST_SENSE_TIMEOUT, | ||
2262 | pmcraid_timeout_handler); | ||
2263 | } | ||
2264 | |||
2265 | /** | ||
2266 | * pmcraid_cancel_all - cancel all outstanding IOARCBs as part of error recovery | ||
2267 | * @cmd: command that failed | ||
2268 | * @sense: true if request_sense is required after cancel all | ||
2269 | * | ||
2270 | * This function sends a cancel all to a device to clear the queue. | ||
2271 | */ | ||
2272 | static void pmcraid_cancel_all(struct pmcraid_cmd *cmd, u32 sense) | ||
2273 | { | ||
2274 | struct scsi_cmnd *scsi_cmd = cmd->scsi_cmd; | ||
2275 | struct pmcraid_ioarcb *ioarcb = &cmd->ioa_cb->ioarcb; | ||
2276 | struct pmcraid_resource_entry *res = scsi_cmd->device->hostdata; | ||
2277 | void (*cmd_done) (struct pmcraid_cmd *) = sense ? pmcraid_erp_done | ||
2278 | : pmcraid_request_sense; | ||
2279 | |||
2280 | memset(ioarcb->cdb, 0, PMCRAID_MAX_CDB_LEN); | ||
2281 | ioarcb->request_flags0 = SYNC_OVERRIDE; | ||
2282 | ioarcb->request_type = REQ_TYPE_IOACMD; | ||
2283 | ioarcb->cdb[0] = PMCRAID_CANCEL_ALL_REQUESTS; | ||
2284 | |||
2285 | if (RES_IS_GSCSI(res->cfg_entry)) | ||
2286 | ioarcb->cdb[1] = PMCRAID_SYNC_COMPLETE_AFTER_CANCEL; | ||
2287 | |||
2288 | ioarcb->ioadl_bus_addr = 0; | ||
2289 | ioarcb->ioadl_length = 0; | ||
2290 | ioarcb->data_transfer_length = 0; | ||
2291 | ioarcb->ioarcb_bus_addr &= (~0x1FULL); | ||
2292 | |||
2293 | /* writing to IOARRIN must be protected by host_lock, as mid-layer | ||
2294 | * schedule queuecommand while we are doing this | ||
2295 | */ | ||
2296 | pmcraid_send_cmd(cmd, cmd_done, | ||
2297 | PMCRAID_REQUEST_SENSE_TIMEOUT, | ||
2298 | pmcraid_timeout_handler); | ||
2299 | } | ||
2300 | |||
2301 | /** | ||
2302 | * pmcraid_frame_auto_sense: frame fixed format sense information | ||
2303 | * | ||
2304 | * @cmd: pointer to failing command block | ||
2305 | * | ||
2306 | * Return value | ||
2307 | * none | ||
2308 | */ | ||
2309 | static void pmcraid_frame_auto_sense(struct pmcraid_cmd *cmd) | ||
2310 | { | ||
2311 | u8 *sense_buf = cmd->scsi_cmd->sense_buffer; | ||
2312 | struct pmcraid_resource_entry *res = cmd->scsi_cmd->device->hostdata; | ||
2313 | struct pmcraid_ioasa *ioasa = &cmd->ioa_cb->ioasa; | ||
2314 | u32 ioasc = le32_to_cpu(ioasa->ioasc); | ||
2315 | u32 failing_lba = 0; | ||
2316 | |||
2317 | memset(sense_buf, 0, SCSI_SENSE_BUFFERSIZE); | ||
2318 | cmd->scsi_cmd->result = SAM_STAT_CHECK_CONDITION; | ||
2319 | |||
2320 | if (RES_IS_VSET(res->cfg_entry) && | ||
2321 | ioasc == PMCRAID_IOASC_ME_READ_ERROR_NO_REALLOC && | ||
2322 | ioasa->u.vset.failing_lba_hi != 0) { | ||
2323 | |||
2324 | sense_buf[0] = 0x72; | ||
2325 | sense_buf[1] = PMCRAID_IOASC_SENSE_KEY(ioasc); | ||
2326 | sense_buf[2] = PMCRAID_IOASC_SENSE_CODE(ioasc); | ||
2327 | sense_buf[3] = PMCRAID_IOASC_SENSE_QUAL(ioasc); | ||
2328 | |||
2329 | sense_buf[7] = 12; | ||
2330 | sense_buf[8] = 0; | ||
2331 | sense_buf[9] = 0x0A; | ||
2332 | sense_buf[10] = 0x80; | ||
2333 | |||
2334 | failing_lba = le32_to_cpu(ioasa->u.vset.failing_lba_hi); | ||
2335 | |||
2336 | sense_buf[12] = (failing_lba & 0xff000000) >> 24; | ||
2337 | sense_buf[13] = (failing_lba & 0x00ff0000) >> 16; | ||
2338 | sense_buf[14] = (failing_lba & 0x0000ff00) >> 8; | ||
2339 | sense_buf[15] = failing_lba & 0x000000ff; | ||
2340 | |||
2341 | failing_lba = le32_to_cpu(ioasa->u.vset.failing_lba_lo); | ||
2342 | |||
2343 | sense_buf[16] = (failing_lba & 0xff000000) >> 24; | ||
2344 | sense_buf[17] = (failing_lba & 0x00ff0000) >> 16; | ||
2345 | sense_buf[18] = (failing_lba & 0x0000ff00) >> 8; | ||
2346 | sense_buf[19] = failing_lba & 0x000000ff; | ||
2347 | } else { | ||
2348 | sense_buf[0] = 0x70; | ||
2349 | sense_buf[2] = PMCRAID_IOASC_SENSE_KEY(ioasc); | ||
2350 | sense_buf[12] = PMCRAID_IOASC_SENSE_CODE(ioasc); | ||
2351 | sense_buf[13] = PMCRAID_IOASC_SENSE_QUAL(ioasc); | ||
2352 | |||
2353 | if (ioasc == PMCRAID_IOASC_ME_READ_ERROR_NO_REALLOC) { | ||
2354 | if (RES_IS_VSET(res->cfg_entry)) | ||
2355 | failing_lba = | ||
2356 | le32_to_cpu(ioasa->u. | ||
2357 | vset.failing_lba_lo); | ||
2358 | sense_buf[0] |= 0x80; | ||
2359 | sense_buf[3] = (failing_lba >> 24) & 0xff; | ||
2360 | sense_buf[4] = (failing_lba >> 16) & 0xff; | ||
2361 | sense_buf[5] = (failing_lba >> 8) & 0xff; | ||
2362 | sense_buf[6] = failing_lba & 0xff; | ||
2363 | } | ||
2364 | |||
2365 | sense_buf[7] = 6; /* additional length */ | ||
2366 | } | ||
2367 | } | ||
2368 | |||
2369 | /** | ||
2370 | * pmcraid_error_handler - Error response handlers for a SCSI op | ||
2371 | * @cmd: pointer to pmcraid_cmd that has failed | ||
2372 | * | ||
2373 | * This function determines whether or not to initiate ERP on the affected | ||
2374 | * device. This is called from a tasklet, which doesn't hold any locks. | ||
2375 | * | ||
2376 | * Return value: | ||
2377 | * 0 it caller can complete the request, otherwise 1 where in error | ||
2378 | * handler itself completes the request and returns the command block | ||
2379 | * back to free-pool | ||
2380 | */ | ||
2381 | static int pmcraid_error_handler(struct pmcraid_cmd *cmd) | ||
2382 | { | ||
2383 | struct scsi_cmnd *scsi_cmd = cmd->scsi_cmd; | ||
2384 | struct pmcraid_resource_entry *res = scsi_cmd->device->hostdata; | ||
2385 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
2386 | struct pmcraid_ioasa *ioasa = &cmd->ioa_cb->ioasa; | ||
2387 | u32 ioasc = le32_to_cpu(ioasa->ioasc); | ||
2388 | u32 masked_ioasc = ioasc & PMCRAID_IOASC_SENSE_MASK; | ||
2389 | u32 sense_copied = 0; | ||
2390 | |||
2391 | if (!res) { | ||
2392 | pmcraid_info("resource pointer is NULL\n"); | ||
2393 | return 0; | ||
2394 | } | ||
2395 | |||
2396 | /* If this was a SCSI read/write command keep count of errors */ | ||
2397 | if (SCSI_CMD_TYPE(scsi_cmd->cmnd[0]) == SCSI_READ_CMD) | ||
2398 | atomic_inc(&res->read_failures); | ||
2399 | else if (SCSI_CMD_TYPE(scsi_cmd->cmnd[0]) == SCSI_WRITE_CMD) | ||
2400 | atomic_inc(&res->write_failures); | ||
2401 | |||
2402 | if (!RES_IS_GSCSI(res->cfg_entry) && | ||
2403 | masked_ioasc != PMCRAID_IOASC_HW_DEVICE_BUS_STATUS_ERROR) { | ||
2404 | pmcraid_frame_auto_sense(cmd); | ||
2405 | } | ||
2406 | |||
2407 | /* Log IOASC/IOASA information based on user settings */ | ||
2408 | pmcraid_ioasc_logger(ioasc, cmd); | ||
2409 | |||
2410 | switch (masked_ioasc) { | ||
2411 | |||
2412 | case PMCRAID_IOASC_AC_TERMINATED_BY_HOST: | ||
2413 | scsi_cmd->result |= (DID_ABORT << 16); | ||
2414 | break; | ||
2415 | |||
2416 | case PMCRAID_IOASC_IR_INVALID_RESOURCE_HANDLE: | ||
2417 | case PMCRAID_IOASC_HW_CANNOT_COMMUNICATE: | ||
2418 | scsi_cmd->result |= (DID_NO_CONNECT << 16); | ||
2419 | break; | ||
2420 | |||
2421 | case PMCRAID_IOASC_NR_SYNC_REQUIRED: | ||
2422 | res->sync_reqd = 1; | ||
2423 | scsi_cmd->result |= (DID_IMM_RETRY << 16); | ||
2424 | break; | ||
2425 | |||
2426 | case PMCRAID_IOASC_ME_READ_ERROR_NO_REALLOC: | ||
2427 | scsi_cmd->result |= (DID_PASSTHROUGH << 16); | ||
2428 | break; | ||
2429 | |||
2430 | case PMCRAID_IOASC_UA_BUS_WAS_RESET: | ||
2431 | case PMCRAID_IOASC_UA_BUS_WAS_RESET_BY_OTHER: | ||
2432 | if (!res->reset_progress) | ||
2433 | scsi_report_bus_reset(pinstance->host, | ||
2434 | scsi_cmd->device->channel); | ||
2435 | scsi_cmd->result |= (DID_ERROR << 16); | ||
2436 | break; | ||
2437 | |||
2438 | case PMCRAID_IOASC_HW_DEVICE_BUS_STATUS_ERROR: | ||
2439 | scsi_cmd->result |= PMCRAID_IOASC_SENSE_STATUS(ioasc); | ||
2440 | res->sync_reqd = 1; | ||
2441 | |||
2442 | /* if check_condition is not active return with error otherwise | ||
2443 | * get/frame the sense buffer | ||
2444 | */ | ||
2445 | if (PMCRAID_IOASC_SENSE_STATUS(ioasc) != | ||
2446 | SAM_STAT_CHECK_CONDITION && | ||
2447 | PMCRAID_IOASC_SENSE_STATUS(ioasc) != SAM_STAT_ACA_ACTIVE) | ||
2448 | return 0; | ||
2449 | |||
2450 | /* If we have auto sense data as part of IOASA pass it to | ||
2451 | * mid-layer | ||
2452 | */ | ||
2453 | if (ioasa->auto_sense_length != 0) { | ||
2454 | short sense_len = ioasa->auto_sense_length; | ||
2455 | int data_size = min_t(u16, le16_to_cpu(sense_len), | ||
2456 | SCSI_SENSE_BUFFERSIZE); | ||
2457 | |||
2458 | memcpy(scsi_cmd->sense_buffer, | ||
2459 | ioasa->sense_data, | ||
2460 | data_size); | ||
2461 | sense_copied = 1; | ||
2462 | } | ||
2463 | |||
2464 | if (RES_IS_GSCSI(res->cfg_entry)) { | ||
2465 | pmcraid_cancel_all(cmd, sense_copied); | ||
2466 | } else if (sense_copied) { | ||
2467 | pmcraid_erp_done(cmd); | ||
2468 | return 0; | ||
2469 | } else { | ||
2470 | pmcraid_request_sense(cmd); | ||
2471 | } | ||
2472 | |||
2473 | return 1; | ||
2474 | |||
2475 | case PMCRAID_IOASC_NR_INIT_CMD_REQUIRED: | ||
2476 | break; | ||
2477 | |||
2478 | default: | ||
2479 | if (PMCRAID_IOASC_SENSE_KEY(ioasc) > RECOVERED_ERROR) | ||
2480 | scsi_cmd->result |= (DID_ERROR << 16); | ||
2481 | break; | ||
2482 | } | ||
2483 | return 0; | ||
2484 | } | ||
2485 | |||
2486 | /** | ||
2487 | * pmcraid_reset_device - device reset handler functions | ||
2488 | * | ||
2489 | * @scsi_cmd: scsi command struct | ||
2490 | * @modifier: reset modifier indicating the reset sequence to be performed | ||
2491 | * | ||
2492 | * This function issues a device reset to the affected device. | ||
2493 | * A LUN reset will be sent to the device first. If that does | ||
2494 | * not work, a target reset will be sent. | ||
2495 | * | ||
2496 | * Return value: | ||
2497 | * SUCCESS / FAILED | ||
2498 | */ | ||
2499 | static int pmcraid_reset_device( | ||
2500 | struct scsi_cmnd *scsi_cmd, | ||
2501 | unsigned long timeout, | ||
2502 | u8 modifier | ||
2503 | ) | ||
2504 | { | ||
2505 | struct pmcraid_cmd *cmd; | ||
2506 | struct pmcraid_instance *pinstance; | ||
2507 | struct pmcraid_resource_entry *res; | ||
2508 | struct pmcraid_ioarcb *ioarcb; | ||
2509 | unsigned long lock_flags; | ||
2510 | u32 ioasc; | ||
2511 | |||
2512 | pinstance = | ||
2513 | (struct pmcraid_instance *)scsi_cmd->device->host->hostdata; | ||
2514 | res = scsi_cmd->device->hostdata; | ||
2515 | |||
2516 | if (!res) { | ||
2517 | pmcraid_err("reset_device: NULL resource pointer\n"); | ||
2518 | return FAILED; | ||
2519 | } | ||
2520 | |||
2521 | /* If adapter is currently going through reset/reload, return failed. | ||
2522 | * This will force the mid-layer to call _eh_bus/host reset, which | ||
2523 | * will then go to sleep and wait for the reset to complete | ||
2524 | */ | ||
2525 | spin_lock_irqsave(pinstance->host->host_lock, lock_flags); | ||
2526 | if (pinstance->ioa_reset_in_progress || | ||
2527 | pinstance->ioa_state == IOA_STATE_DEAD) { | ||
2528 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
2529 | return FAILED; | ||
2530 | } | ||
2531 | |||
2532 | res->reset_progress = 1; | ||
2533 | pmcraid_info("Resetting %s resource with addr %x\n", | ||
2534 | ((modifier & RESET_DEVICE_LUN) ? "LUN" : | ||
2535 | ((modifier & RESET_DEVICE_TARGET) ? "TARGET" : "BUS")), | ||
2536 | le32_to_cpu(res->cfg_entry.resource_address)); | ||
2537 | |||
2538 | /* get a free cmd block */ | ||
2539 | cmd = pmcraid_get_free_cmd(pinstance); | ||
2540 | |||
2541 | if (cmd == NULL) { | ||
2542 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
2543 | pmcraid_err("%s: no cmd blocks are available\n", __func__); | ||
2544 | return FAILED; | ||
2545 | } | ||
2546 | |||
2547 | ioarcb = &cmd->ioa_cb->ioarcb; | ||
2548 | ioarcb->resource_handle = res->cfg_entry.resource_handle; | ||
2549 | ioarcb->request_type = REQ_TYPE_IOACMD; | ||
2550 | ioarcb->cdb[0] = PMCRAID_RESET_DEVICE; | ||
2551 | |||
2552 | /* Initialize reset modifier bits */ | ||
2553 | if (modifier) | ||
2554 | modifier = ENABLE_RESET_MODIFIER | modifier; | ||
2555 | |||
2556 | ioarcb->cdb[1] = modifier; | ||
2557 | |||
2558 | init_completion(&cmd->wait_for_completion); | ||
2559 | cmd->completion_req = 1; | ||
2560 | |||
2561 | pmcraid_info("cmd(CDB[0] = %x) for %x with index = %d\n", | ||
2562 | cmd->ioa_cb->ioarcb.cdb[0], | ||
2563 | le32_to_cpu(cmd->ioa_cb->ioarcb.resource_handle), | ||
2564 | le32_to_cpu(cmd->ioa_cb->ioarcb.response_handle) >> 2); | ||
2565 | |||
2566 | pmcraid_send_cmd(cmd, | ||
2567 | pmcraid_internal_done, | ||
2568 | timeout, | ||
2569 | pmcraid_timeout_handler); | ||
2570 | |||
2571 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
2572 | |||
2573 | /* RESET_DEVICE command completes after all pending IOARCBs are | ||
2574 | * completed. Once this command is completed, pmcraind_internal_done | ||
2575 | * will wake up the 'completion' queue. | ||
2576 | */ | ||
2577 | wait_for_completion(&cmd->wait_for_completion); | ||
2578 | |||
2579 | /* complete the command here itself and return the command block | ||
2580 | * to free list | ||
2581 | */ | ||
2582 | pmcraid_return_cmd(cmd); | ||
2583 | res->reset_progress = 0; | ||
2584 | ioasc = le32_to_cpu(cmd->ioa_cb->ioasa.ioasc); | ||
2585 | |||
2586 | /* set the return value based on the returned ioasc */ | ||
2587 | return PMCRAID_IOASC_SENSE_KEY(ioasc) ? FAILED : SUCCESS; | ||
2588 | } | ||
2589 | |||
2590 | /** | ||
2591 | * _pmcraid_io_done - helper for pmcraid_io_done function | ||
2592 | * | ||
2593 | * @cmd: pointer to pmcraid command struct | ||
2594 | * @reslen: residual data length to be set in the ioasa | ||
2595 | * @ioasc: ioasc either returned by IOA or set by driver itself. | ||
2596 | * | ||
2597 | * This function is invoked by pmcraid_io_done to complete mid-layer | ||
2598 | * scsi ops. | ||
2599 | * | ||
2600 | * Return value: | ||
2601 | * 0 if caller is required to return it to free_pool. Returns 1 if | ||
2602 | * caller need not worry about freeing command block as error handler | ||
2603 | * will take care of that. | ||
2604 | */ | ||
2605 | |||
2606 | static int _pmcraid_io_done(struct pmcraid_cmd *cmd, int reslen, int ioasc) | ||
2607 | { | ||
2608 | struct scsi_cmnd *scsi_cmd = cmd->scsi_cmd; | ||
2609 | int rc = 0; | ||
2610 | |||
2611 | scsi_set_resid(scsi_cmd, reslen); | ||
2612 | |||
2613 | pmcraid_info("response(%d) CDB[0] = %x ioasc:result: %x:%x\n", | ||
2614 | le32_to_cpu(cmd->ioa_cb->ioarcb.response_handle) >> 2, | ||
2615 | cmd->ioa_cb->ioarcb.cdb[0], | ||
2616 | ioasc, scsi_cmd->result); | ||
2617 | |||
2618 | if (PMCRAID_IOASC_SENSE_KEY(ioasc) != 0) | ||
2619 | rc = pmcraid_error_handler(cmd); | ||
2620 | |||
2621 | if (rc == 0) { | ||
2622 | scsi_dma_unmap(scsi_cmd); | ||
2623 | scsi_cmd->scsi_done(scsi_cmd); | ||
2624 | } | ||
2625 | |||
2626 | return rc; | ||
2627 | } | ||
2628 | |||
2629 | /** | ||
2630 | * pmcraid_io_done - SCSI completion function | ||
2631 | * | ||
2632 | * @cmd: pointer to pmcraid command struct | ||
2633 | * | ||
2634 | * This function is invoked by tasklet/mid-layer error handler to completing | ||
2635 | * the SCSI ops sent from mid-layer. | ||
2636 | * | ||
2637 | * Return value | ||
2638 | * none | ||
2639 | */ | ||
2640 | |||
2641 | static void pmcraid_io_done(struct pmcraid_cmd *cmd) | ||
2642 | { | ||
2643 | u32 ioasc = le32_to_cpu(cmd->ioa_cb->ioasa.ioasc); | ||
2644 | u32 reslen = le32_to_cpu(cmd->ioa_cb->ioasa.residual_data_length); | ||
2645 | |||
2646 | if (_pmcraid_io_done(cmd, reslen, ioasc) == 0) | ||
2647 | pmcraid_return_cmd(cmd); | ||
2648 | } | ||
2649 | |||
2650 | /** | ||
2651 | * pmcraid_abort_cmd - Aborts a single IOARCB already submitted to IOA | ||
2652 | * | ||
2653 | * @cmd: command block of the command to be aborted | ||
2654 | * | ||
2655 | * Return Value: | ||
2656 | * returns pointer to command structure used as cancelling cmd | ||
2657 | */ | ||
2658 | static struct pmcraid_cmd *pmcraid_abort_cmd(struct pmcraid_cmd *cmd) | ||
2659 | { | ||
2660 | struct pmcraid_cmd *cancel_cmd; | ||
2661 | struct pmcraid_instance *pinstance; | ||
2662 | struct pmcraid_resource_entry *res; | ||
2663 | |||
2664 | pinstance = (struct pmcraid_instance *)cmd->drv_inst; | ||
2665 | res = cmd->scsi_cmd->device->hostdata; | ||
2666 | |||
2667 | cancel_cmd = pmcraid_get_free_cmd(pinstance); | ||
2668 | |||
2669 | if (cancel_cmd == NULL) { | ||
2670 | pmcraid_err("%s: no cmd blocks are available\n", __func__); | ||
2671 | return NULL; | ||
2672 | } | ||
2673 | |||
2674 | pmcraid_prepare_cancel_cmd(cancel_cmd, cmd); | ||
2675 | |||
2676 | pmcraid_info("aborting command CDB[0]= %x with index = %d\n", | ||
2677 | cmd->ioa_cb->ioarcb.cdb[0], | ||
2678 | cmd->ioa_cb->ioarcb.response_handle >> 2); | ||
2679 | |||
2680 | init_completion(&cancel_cmd->wait_for_completion); | ||
2681 | cancel_cmd->completion_req = 1; | ||
2682 | |||
2683 | pmcraid_info("command (%d) CDB[0] = %x for %x\n", | ||
2684 | le32_to_cpu(cancel_cmd->ioa_cb->ioarcb.response_handle) >> 2, | ||
2685 | cmd->ioa_cb->ioarcb.cdb[0], | ||
2686 | le32_to_cpu(cancel_cmd->ioa_cb->ioarcb.resource_handle)); | ||
2687 | |||
2688 | pmcraid_send_cmd(cancel_cmd, | ||
2689 | pmcraid_internal_done, | ||
2690 | PMCRAID_INTERNAL_TIMEOUT, | ||
2691 | pmcraid_timeout_handler); | ||
2692 | return cancel_cmd; | ||
2693 | } | ||
2694 | |||
2695 | /** | ||
2696 | * pmcraid_abort_complete - Waits for ABORT TASK completion | ||
2697 | * | ||
2698 | * @cancel_cmd: command block use as cancelling command | ||
2699 | * | ||
2700 | * Return Value: | ||
2701 | * returns SUCCESS if ABORT TASK has good completion | ||
2702 | * otherwise FAILED | ||
2703 | */ | ||
2704 | static int pmcraid_abort_complete(struct pmcraid_cmd *cancel_cmd) | ||
2705 | { | ||
2706 | struct pmcraid_resource_entry *res; | ||
2707 | u32 ioasc; | ||
2708 | |||
2709 | wait_for_completion(&cancel_cmd->wait_for_completion); | ||
2710 | res = cancel_cmd->u.res; | ||
2711 | cancel_cmd->u.res = NULL; | ||
2712 | ioasc = le32_to_cpu(cancel_cmd->ioa_cb->ioasa.ioasc); | ||
2713 | |||
2714 | /* If the abort task is not timed out we will get a Good completion | ||
2715 | * as sense_key, otherwise we may get one the following responses | ||
2716 | * due to subsquent bus reset or device reset. In case IOASC is | ||
2717 | * NR_SYNC_REQUIRED, set sync_reqd flag for the corresponding resource | ||
2718 | */ | ||
2719 | if (ioasc == PMCRAID_IOASC_UA_BUS_WAS_RESET || | ||
2720 | ioasc == PMCRAID_IOASC_NR_SYNC_REQUIRED) { | ||
2721 | if (ioasc == PMCRAID_IOASC_NR_SYNC_REQUIRED) | ||
2722 | res->sync_reqd = 1; | ||
2723 | ioasc = 0; | ||
2724 | } | ||
2725 | |||
2726 | /* complete the command here itself */ | ||
2727 | pmcraid_return_cmd(cancel_cmd); | ||
2728 | return PMCRAID_IOASC_SENSE_KEY(ioasc) ? FAILED : SUCCESS; | ||
2729 | } | ||
2730 | |||
2731 | /** | ||
2732 | * pmcraid_eh_abort_handler - entry point for aborting a single task on errors | ||
2733 | * | ||
2734 | * @scsi_cmd: scsi command struct given by mid-layer. When this is called | ||
2735 | * mid-layer ensures that no other commands are queued. This | ||
2736 | * never gets called under interrupt, but a separate eh thread. | ||
2737 | * | ||
2738 | * Return value: | ||
2739 | * SUCCESS / FAILED | ||
2740 | */ | ||
2741 | static int pmcraid_eh_abort_handler(struct scsi_cmnd *scsi_cmd) | ||
2742 | { | ||
2743 | struct pmcraid_instance *pinstance; | ||
2744 | struct pmcraid_cmd *cmd; | ||
2745 | struct pmcraid_resource_entry *res; | ||
2746 | unsigned long host_lock_flags; | ||
2747 | unsigned long pending_lock_flags; | ||
2748 | struct pmcraid_cmd *cancel_cmd = NULL; | ||
2749 | int cmd_found = 0; | ||
2750 | int rc = FAILED; | ||
2751 | |||
2752 | pinstance = | ||
2753 | (struct pmcraid_instance *)scsi_cmd->device->host->hostdata; | ||
2754 | |||
2755 | dev_err(&pinstance->pdev->dev, | ||
2756 | "I/O command timed out, aborting it.\n"); | ||
2757 | |||
2758 | res = scsi_cmd->device->hostdata; | ||
2759 | |||
2760 | if (res == NULL) | ||
2761 | return rc; | ||
2762 | |||
2763 | /* If we are currently going through reset/reload, return failed. | ||
2764 | * This will force the mid-layer to eventually call | ||
2765 | * pmcraid_eh_host_reset which will then go to sleep and wait for the | ||
2766 | * reset to complete | ||
2767 | */ | ||
2768 | spin_lock_irqsave(pinstance->host->host_lock, host_lock_flags); | ||
2769 | |||
2770 | if (pinstance->ioa_reset_in_progress || | ||
2771 | pinstance->ioa_state == IOA_STATE_DEAD) { | ||
2772 | spin_unlock_irqrestore(pinstance->host->host_lock, | ||
2773 | host_lock_flags); | ||
2774 | return rc; | ||
2775 | } | ||
2776 | |||
2777 | /* loop over pending cmd list to find cmd corresponding to this | ||
2778 | * scsi_cmd. Note that this command might not have been completed | ||
2779 | * already. locking: all pending commands are protected with | ||
2780 | * pending_pool_lock. | ||
2781 | */ | ||
2782 | spin_lock_irqsave(&pinstance->pending_pool_lock, pending_lock_flags); | ||
2783 | list_for_each_entry(cmd, &pinstance->pending_cmd_pool, free_list) { | ||
2784 | |||
2785 | if (cmd->scsi_cmd == scsi_cmd) { | ||
2786 | cmd_found = 1; | ||
2787 | break; | ||
2788 | } | ||
2789 | } | ||
2790 | |||
2791 | spin_unlock_irqrestore(&pinstance->pending_pool_lock, | ||
2792 | pending_lock_flags); | ||
2793 | |||
2794 | /* If the command to be aborted was given to IOA and still pending with | ||
2795 | * it, send ABORT_TASK to abort this and wait for its completion | ||
2796 | */ | ||
2797 | if (cmd_found) | ||
2798 | cancel_cmd = pmcraid_abort_cmd(cmd); | ||
2799 | |||
2800 | spin_unlock_irqrestore(pinstance->host->host_lock, | ||
2801 | host_lock_flags); | ||
2802 | |||
2803 | if (cancel_cmd) { | ||
2804 | cancel_cmd->u.res = cmd->scsi_cmd->device->hostdata; | ||
2805 | rc = pmcraid_abort_complete(cancel_cmd); | ||
2806 | } | ||
2807 | |||
2808 | return cmd_found ? rc : SUCCESS; | ||
2809 | } | ||
2810 | |||
2811 | /** | ||
2812 | * pmcraid_eh_xxxx_reset_handler - bus/target/device reset handler callbacks | ||
2813 | * | ||
2814 | * @scmd: pointer to scsi_cmd that was sent to the resource to be reset. | ||
2815 | * | ||
2816 | * All these routines invokve pmcraid_reset_device with appropriate parameters. | ||
2817 | * Since these are called from mid-layer EH thread, no other IO will be queued | ||
2818 | * to the resource being reset. However, control path (IOCTL) may be active so | ||
2819 | * it is necessary to synchronize IOARRIN writes which pmcraid_reset_device | ||
2820 | * takes care by locking/unlocking host_lock. | ||
2821 | * | ||
2822 | * Return value | ||
2823 | * SUCCESS or FAILED | ||
2824 | */ | ||
2825 | static int pmcraid_eh_device_reset_handler(struct scsi_cmnd *scmd) | ||
2826 | { | ||
2827 | pmcraid_err("Doing device reset due to an I/O command timeout.\n"); | ||
2828 | return pmcraid_reset_device(scmd, | ||
2829 | PMCRAID_INTERNAL_TIMEOUT, | ||
2830 | RESET_DEVICE_LUN); | ||
2831 | } | ||
2832 | |||
2833 | static int pmcraid_eh_bus_reset_handler(struct scsi_cmnd *scmd) | ||
2834 | { | ||
2835 | pmcraid_err("Doing bus reset due to an I/O command timeout.\n"); | ||
2836 | return pmcraid_reset_device(scmd, | ||
2837 | PMCRAID_RESET_BUS_TIMEOUT, | ||
2838 | RESET_DEVICE_BUS); | ||
2839 | } | ||
2840 | |||
2841 | static int pmcraid_eh_target_reset_handler(struct scsi_cmnd *scmd) | ||
2842 | { | ||
2843 | pmcraid_err("Doing target reset due to an I/O command timeout.\n"); | ||
2844 | return pmcraid_reset_device(scmd, | ||
2845 | PMCRAID_INTERNAL_TIMEOUT, | ||
2846 | RESET_DEVICE_TARGET); | ||
2847 | } | ||
2848 | |||
2849 | /** | ||
2850 | * pmcraid_eh_host_reset_handler - adapter reset handler callback | ||
2851 | * | ||
2852 | * @scmd: pointer to scsi_cmd that was sent to a resource of adapter | ||
2853 | * | ||
2854 | * Initiates adapter reset to bring it up to operational state | ||
2855 | * | ||
2856 | * Return value | ||
2857 | * SUCCESS or FAILED | ||
2858 | */ | ||
2859 | static int pmcraid_eh_host_reset_handler(struct scsi_cmnd *scmd) | ||
2860 | { | ||
2861 | unsigned long interval = 10000; /* 10 seconds interval */ | ||
2862 | int waits = jiffies_to_msecs(PMCRAID_RESET_HOST_TIMEOUT) / interval; | ||
2863 | struct pmcraid_instance *pinstance = | ||
2864 | (struct pmcraid_instance *)(scmd->device->host->hostdata); | ||
2865 | |||
2866 | |||
2867 | /* wait for an additional 150 seconds just in case firmware could come | ||
2868 | * up and if it could complete all the pending commands excluding the | ||
2869 | * two HCAM (CCN and LDN). | ||
2870 | */ | ||
2871 | while (waits--) { | ||
2872 | if (atomic_read(&pinstance->outstanding_cmds) <= | ||
2873 | PMCRAID_MAX_HCAM_CMD) | ||
2874 | return SUCCESS; | ||
2875 | msleep(interval); | ||
2876 | } | ||
2877 | |||
2878 | dev_err(&pinstance->pdev->dev, | ||
2879 | "Adapter being reset due to an I/O command timeout.\n"); | ||
2880 | return pmcraid_reset_bringup(pinstance) == 0 ? SUCCESS : FAILED; | ||
2881 | } | ||
2882 | |||
2883 | /** | ||
2884 | * pmcraid_task_attributes - Translate SPI Q-Tags to task attributes | ||
2885 | * @scsi_cmd: scsi command struct | ||
2886 | * | ||
2887 | * Return value | ||
2888 | * number of tags or 0 if the task is not tagged | ||
2889 | */ | ||
2890 | static u8 pmcraid_task_attributes(struct scsi_cmnd *scsi_cmd) | ||
2891 | { | ||
2892 | char tag[2]; | ||
2893 | u8 rc = 0; | ||
2894 | |||
2895 | if (scsi_populate_tag_msg(scsi_cmd, tag)) { | ||
2896 | switch (tag[0]) { | ||
2897 | case MSG_SIMPLE_TAG: | ||
2898 | rc = TASK_TAG_SIMPLE; | ||
2899 | break; | ||
2900 | case MSG_HEAD_TAG: | ||
2901 | rc = TASK_TAG_QUEUE_HEAD; | ||
2902 | break; | ||
2903 | case MSG_ORDERED_TAG: | ||
2904 | rc = TASK_TAG_ORDERED; | ||
2905 | break; | ||
2906 | }; | ||
2907 | } | ||
2908 | |||
2909 | return rc; | ||
2910 | } | ||
2911 | |||
2912 | |||
2913 | /** | ||
2914 | * pmcraid_init_ioadls - initializes IOADL related fields in IOARCB | ||
2915 | * @cmd: pmcraid command struct | ||
2916 | * @sgcount: count of scatter-gather elements | ||
2917 | * | ||
2918 | * Return value | ||
2919 | * returns pointer pmcraid_ioadl_desc, initialized to point to internal | ||
2920 | * or external IOADLs | ||
2921 | */ | ||
2922 | struct pmcraid_ioadl_desc * | ||
2923 | pmcraid_init_ioadls(struct pmcraid_cmd *cmd, int sgcount) | ||
2924 | { | ||
2925 | struct pmcraid_ioadl_desc *ioadl; | ||
2926 | struct pmcraid_ioarcb *ioarcb = &cmd->ioa_cb->ioarcb; | ||
2927 | int ioadl_count = 0; | ||
2928 | |||
2929 | if (ioarcb->add_cmd_param_length) | ||
2930 | ioadl_count = DIV_ROUND_UP(ioarcb->add_cmd_param_length, 16); | ||
2931 | ioarcb->ioadl_length = | ||
2932 | sizeof(struct pmcraid_ioadl_desc) * sgcount; | ||
2933 | |||
2934 | if ((sgcount + ioadl_count) > (ARRAY_SIZE(ioarcb->add_data.u.ioadl))) { | ||
2935 | /* external ioadls start at offset 0x80 from control_block | ||
2936 | * structure, re-using 24 out of 27 ioadls part of IOARCB. | ||
2937 | * It is necessary to indicate to firmware that driver is | ||
2938 | * using ioadls to be treated as external to IOARCB. | ||
2939 | */ | ||
2940 | ioarcb->ioarcb_bus_addr &= ~(0x1FULL); | ||
2941 | ioarcb->ioadl_bus_addr = | ||
2942 | cpu_to_le64((cmd->ioa_cb_bus_addr) + | ||
2943 | offsetof(struct pmcraid_ioarcb, | ||
2944 | add_data.u.ioadl[3])); | ||
2945 | ioadl = &ioarcb->add_data.u.ioadl[3]; | ||
2946 | } else { | ||
2947 | ioarcb->ioadl_bus_addr = | ||
2948 | cpu_to_le64((cmd->ioa_cb_bus_addr) + | ||
2949 | offsetof(struct pmcraid_ioarcb, | ||
2950 | add_data.u.ioadl[ioadl_count])); | ||
2951 | |||
2952 | ioadl = &ioarcb->add_data.u.ioadl[ioadl_count]; | ||
2953 | ioarcb->ioarcb_bus_addr |= | ||
2954 | DIV_ROUND_CLOSEST(sgcount + ioadl_count, 8); | ||
2955 | } | ||
2956 | |||
2957 | return ioadl; | ||
2958 | } | ||
2959 | |||
2960 | /** | ||
2961 | * pmcraid_build_ioadl - Build a scatter/gather list and map the buffer | ||
2962 | * @pinstance: pointer to adapter instance structure | ||
2963 | * @cmd: pmcraid command struct | ||
2964 | * | ||
2965 | * This function is invoked by queuecommand entry point while sending a command | ||
2966 | * to firmware. This builds ioadl descriptors and sets up ioarcb fields. | ||
2967 | * | ||
2968 | * Return value: | ||
2969 | * 0 on success or -1 on failure | ||
2970 | */ | ||
2971 | static int pmcraid_build_ioadl( | ||
2972 | struct pmcraid_instance *pinstance, | ||
2973 | struct pmcraid_cmd *cmd | ||
2974 | ) | ||
2975 | { | ||
2976 | int i, nseg; | ||
2977 | struct scatterlist *sglist; | ||
2978 | |||
2979 | struct scsi_cmnd *scsi_cmd = cmd->scsi_cmd; | ||
2980 | struct pmcraid_ioarcb *ioarcb = &(cmd->ioa_cb->ioarcb); | ||
2981 | struct pmcraid_ioadl_desc *ioadl = ioarcb->add_data.u.ioadl; | ||
2982 | |||
2983 | u32 length = scsi_bufflen(scsi_cmd); | ||
2984 | |||
2985 | if (!length) | ||
2986 | return 0; | ||
2987 | |||
2988 | nseg = scsi_dma_map(scsi_cmd); | ||
2989 | |||
2990 | if (nseg < 0) { | ||
2991 | dev_err(&pinstance->pdev->dev, "scsi_map_dma failed!\n"); | ||
2992 | return -1; | ||
2993 | } else if (nseg > PMCRAID_MAX_IOADLS) { | ||
2994 | scsi_dma_unmap(scsi_cmd); | ||
2995 | dev_err(&pinstance->pdev->dev, | ||
2996 | "sg count is (%d) more than allowed!\n", nseg); | ||
2997 | return -1; | ||
2998 | } | ||
2999 | |||
3000 | /* Initialize IOARCB data transfer length fields */ | ||
3001 | if (scsi_cmd->sc_data_direction == DMA_TO_DEVICE) | ||
3002 | ioarcb->request_flags0 |= TRANSFER_DIR_WRITE; | ||
3003 | |||
3004 | ioarcb->request_flags0 |= NO_LINK_DESCS; | ||
3005 | ioarcb->data_transfer_length = cpu_to_le32(length); | ||
3006 | ioadl = pmcraid_init_ioadls(cmd, nseg); | ||
3007 | |||
3008 | /* Initialize IOADL descriptor addresses */ | ||
3009 | scsi_for_each_sg(scsi_cmd, sglist, nseg, i) { | ||
3010 | ioadl[i].data_len = cpu_to_le32(sg_dma_len(sglist)); | ||
3011 | ioadl[i].address = cpu_to_le64(sg_dma_address(sglist)); | ||
3012 | ioadl[i].flags = 0; | ||
3013 | } | ||
3014 | /* setup last descriptor */ | ||
3015 | ioadl[i - 1].flags = cpu_to_le32(IOADL_FLAGS_LAST_DESC); | ||
3016 | |||
3017 | return 0; | ||
3018 | } | ||
3019 | |||
3020 | /** | ||
3021 | * pmcraid_free_sglist - Frees an allocated SG buffer list | ||
3022 | * @sglist: scatter/gather list pointer | ||
3023 | * | ||
3024 | * Free a DMA'able memory previously allocated with pmcraid_alloc_sglist | ||
3025 | * | ||
3026 | * Return value: | ||
3027 | * none | ||
3028 | */ | ||
3029 | static void pmcraid_free_sglist(struct pmcraid_sglist *sglist) | ||
3030 | { | ||
3031 | int i; | ||
3032 | |||
3033 | for (i = 0; i < sglist->num_sg; i++) | ||
3034 | __free_pages(sg_page(&(sglist->scatterlist[i])), | ||
3035 | sglist->order); | ||
3036 | |||
3037 | kfree(sglist); | ||
3038 | } | ||
3039 | |||
3040 | /** | ||
3041 | * pmcraid_alloc_sglist - Allocates memory for a SG list | ||
3042 | * @buflen: buffer length | ||
3043 | * | ||
3044 | * Allocates a DMA'able buffer in chunks and assembles a scatter/gather | ||
3045 | * list. | ||
3046 | * | ||
3047 | * Return value | ||
3048 | * pointer to sglist / NULL on failure | ||
3049 | */ | ||
3050 | static struct pmcraid_sglist *pmcraid_alloc_sglist(int buflen) | ||
3051 | { | ||
3052 | struct pmcraid_sglist *sglist; | ||
3053 | struct scatterlist *scatterlist; | ||
3054 | struct page *page; | ||
3055 | int num_elem, i, j; | ||
3056 | int sg_size; | ||
3057 | int order; | ||
3058 | int bsize_elem; | ||
3059 | |||
3060 | sg_size = buflen / (PMCRAID_MAX_IOADLS - 1); | ||
3061 | order = (sg_size > 0) ? get_order(sg_size) : 0; | ||
3062 | bsize_elem = PAGE_SIZE * (1 << order); | ||
3063 | |||
3064 | /* Determine the actual number of sg entries needed */ | ||
3065 | if (buflen % bsize_elem) | ||
3066 | num_elem = (buflen / bsize_elem) + 1; | ||
3067 | else | ||
3068 | num_elem = buflen / bsize_elem; | ||
3069 | |||
3070 | /* Allocate a scatter/gather list for the DMA */ | ||
3071 | sglist = kzalloc(sizeof(struct pmcraid_sglist) + | ||
3072 | (sizeof(struct scatterlist) * (num_elem - 1)), | ||
3073 | GFP_KERNEL); | ||
3074 | |||
3075 | if (sglist == NULL) | ||
3076 | return NULL; | ||
3077 | |||
3078 | scatterlist = sglist->scatterlist; | ||
3079 | sg_init_table(scatterlist, num_elem); | ||
3080 | sglist->order = order; | ||
3081 | sglist->num_sg = num_elem; | ||
3082 | sg_size = buflen; | ||
3083 | |||
3084 | for (i = 0; i < num_elem; i++) { | ||
3085 | page = alloc_pages(GFP_KERNEL|GFP_DMA, order); | ||
3086 | if (!page) { | ||
3087 | for (j = i - 1; j >= 0; j--) | ||
3088 | __free_pages(sg_page(&scatterlist[j]), order); | ||
3089 | kfree(sglist); | ||
3090 | return NULL; | ||
3091 | } | ||
3092 | |||
3093 | sg_set_page(&scatterlist[i], page, | ||
3094 | sg_size < bsize_elem ? sg_size : bsize_elem, 0); | ||
3095 | sg_size -= bsize_elem; | ||
3096 | } | ||
3097 | |||
3098 | return sglist; | ||
3099 | } | ||
3100 | |||
3101 | /** | ||
3102 | * pmcraid_copy_sglist - Copy user buffer to kernel buffer's SG list | ||
3103 | * @sglist: scatter/gather list pointer | ||
3104 | * @buffer: buffer pointer | ||
3105 | * @len: buffer length | ||
3106 | * @direction: data transfer direction | ||
3107 | * | ||
3108 | * Copy a user buffer into a buffer allocated by pmcraid_alloc_sglist | ||
3109 | * | ||
3110 | * Return value: | ||
3111 | * 0 on success / other on failure | ||
3112 | */ | ||
3113 | static int pmcraid_copy_sglist( | ||
3114 | struct pmcraid_sglist *sglist, | ||
3115 | unsigned long buffer, | ||
3116 | u32 len, | ||
3117 | int direction | ||
3118 | ) | ||
3119 | { | ||
3120 | struct scatterlist *scatterlist; | ||
3121 | void *kaddr; | ||
3122 | int bsize_elem; | ||
3123 | int i; | ||
3124 | int rc = 0; | ||
3125 | |||
3126 | /* Determine the actual number of bytes per element */ | ||
3127 | bsize_elem = PAGE_SIZE * (1 << sglist->order); | ||
3128 | |||
3129 | scatterlist = sglist->scatterlist; | ||
3130 | |||
3131 | for (i = 0; i < (len / bsize_elem); i++, buffer += bsize_elem) { | ||
3132 | struct page *page = sg_page(&scatterlist[i]); | ||
3133 | |||
3134 | kaddr = kmap(page); | ||
3135 | if (direction == DMA_TO_DEVICE) | ||
3136 | rc = __copy_from_user(kaddr, | ||
3137 | (void *)buffer, | ||
3138 | bsize_elem); | ||
3139 | else | ||
3140 | rc = __copy_to_user((void *)buffer, kaddr, bsize_elem); | ||
3141 | |||
3142 | kunmap(page); | ||
3143 | |||
3144 | if (rc) { | ||
3145 | pmcraid_err("failed to copy user data into sg list\n"); | ||
3146 | return -EFAULT; | ||
3147 | } | ||
3148 | |||
3149 | scatterlist[i].length = bsize_elem; | ||
3150 | } | ||
3151 | |||
3152 | if (len % bsize_elem) { | ||
3153 | struct page *page = sg_page(&scatterlist[i]); | ||
3154 | |||
3155 | kaddr = kmap(page); | ||
3156 | |||
3157 | if (direction == DMA_TO_DEVICE) | ||
3158 | rc = __copy_from_user(kaddr, | ||
3159 | (void *)buffer, | ||
3160 | len % bsize_elem); | ||
3161 | else | ||
3162 | rc = __copy_to_user((void *)buffer, | ||
3163 | kaddr, | ||
3164 | len % bsize_elem); | ||
3165 | |||
3166 | kunmap(page); | ||
3167 | |||
3168 | scatterlist[i].length = len % bsize_elem; | ||
3169 | } | ||
3170 | |||
3171 | if (rc) { | ||
3172 | pmcraid_err("failed to copy user data into sg list\n"); | ||
3173 | rc = -EFAULT; | ||
3174 | } | ||
3175 | |||
3176 | return rc; | ||
3177 | } | ||
3178 | |||
3179 | /** | ||
3180 | * pmcraid_queuecommand - Queue a mid-layer request | ||
3181 | * @scsi_cmd: scsi command struct | ||
3182 | * @done: done function | ||
3183 | * | ||
3184 | * This function queues a request generated by the mid-layer. Midlayer calls | ||
3185 | * this routine within host->lock. Some of the functions called by queuecommand | ||
3186 | * would use cmd block queue locks (free_pool_lock and pending_pool_lock) | ||
3187 | * | ||
3188 | * Return value: | ||
3189 | * 0 on success | ||
3190 | * SCSI_MLQUEUE_DEVICE_BUSY if device is busy | ||
3191 | * SCSI_MLQUEUE_HOST_BUSY if host is busy | ||
3192 | */ | ||
3193 | static int pmcraid_queuecommand( | ||
3194 | struct scsi_cmnd *scsi_cmd, | ||
3195 | void (*done) (struct scsi_cmnd *) | ||
3196 | ) | ||
3197 | { | ||
3198 | struct pmcraid_instance *pinstance; | ||
3199 | struct pmcraid_resource_entry *res; | ||
3200 | struct pmcraid_ioarcb *ioarcb; | ||
3201 | struct pmcraid_cmd *cmd; | ||
3202 | int rc = 0; | ||
3203 | |||
3204 | pinstance = | ||
3205 | (struct pmcraid_instance *)scsi_cmd->device->host->hostdata; | ||
3206 | |||
3207 | scsi_cmd->scsi_done = done; | ||
3208 | res = scsi_cmd->device->hostdata; | ||
3209 | scsi_cmd->result = (DID_OK << 16); | ||
3210 | |||
3211 | /* if adapter is marked as dead, set result to DID_NO_CONNECT complete | ||
3212 | * the command | ||
3213 | */ | ||
3214 | if (pinstance->ioa_state == IOA_STATE_DEAD) { | ||
3215 | pmcraid_info("IOA is dead, but queuecommand is scheduled\n"); | ||
3216 | scsi_cmd->result = (DID_NO_CONNECT << 16); | ||
3217 | scsi_cmd->scsi_done(scsi_cmd); | ||
3218 | return 0; | ||
3219 | } | ||
3220 | |||
3221 | /* If IOA reset is in progress, can't queue the commands */ | ||
3222 | if (pinstance->ioa_reset_in_progress) | ||
3223 | return SCSI_MLQUEUE_HOST_BUSY; | ||
3224 | |||
3225 | /* initialize the command and IOARCB to be sent to IOA */ | ||
3226 | cmd = pmcraid_get_free_cmd(pinstance); | ||
3227 | |||
3228 | if (cmd == NULL) { | ||
3229 | pmcraid_err("free command block is not available\n"); | ||
3230 | return SCSI_MLQUEUE_HOST_BUSY; | ||
3231 | } | ||
3232 | |||
3233 | cmd->scsi_cmd = scsi_cmd; | ||
3234 | ioarcb = &(cmd->ioa_cb->ioarcb); | ||
3235 | memcpy(ioarcb->cdb, scsi_cmd->cmnd, scsi_cmd->cmd_len); | ||
3236 | ioarcb->resource_handle = res->cfg_entry.resource_handle; | ||
3237 | ioarcb->request_type = REQ_TYPE_SCSI; | ||
3238 | |||
3239 | cmd->cmd_done = pmcraid_io_done; | ||
3240 | |||
3241 | if (RES_IS_GSCSI(res->cfg_entry) || RES_IS_VSET(res->cfg_entry)) { | ||
3242 | if (scsi_cmd->underflow == 0) | ||
3243 | ioarcb->request_flags0 |= INHIBIT_UL_CHECK; | ||
3244 | |||
3245 | if (res->sync_reqd) { | ||
3246 | ioarcb->request_flags0 |= SYNC_COMPLETE; | ||
3247 | res->sync_reqd = 0; | ||
3248 | } | ||
3249 | |||
3250 | ioarcb->request_flags0 |= NO_LINK_DESCS; | ||
3251 | ioarcb->request_flags1 |= pmcraid_task_attributes(scsi_cmd); | ||
3252 | |||
3253 | if (RES_IS_GSCSI(res->cfg_entry)) | ||
3254 | ioarcb->request_flags1 |= DELAY_AFTER_RESET; | ||
3255 | } | ||
3256 | |||
3257 | rc = pmcraid_build_ioadl(pinstance, cmd); | ||
3258 | |||
3259 | pmcraid_info("command (%d) CDB[0] = %x for %x:%x:%x:%x\n", | ||
3260 | le32_to_cpu(ioarcb->response_handle) >> 2, | ||
3261 | scsi_cmd->cmnd[0], pinstance->host->unique_id, | ||
3262 | RES_IS_VSET(res->cfg_entry) ? PMCRAID_VSET_BUS_ID : | ||
3263 | PMCRAID_PHYS_BUS_ID, | ||
3264 | RES_IS_VSET(res->cfg_entry) ? | ||
3265 | res->cfg_entry.unique_flags1 : | ||
3266 | RES_TARGET(res->cfg_entry.resource_address), | ||
3267 | RES_LUN(res->cfg_entry.resource_address)); | ||
3268 | |||
3269 | if (likely(rc == 0)) { | ||
3270 | _pmcraid_fire_command(cmd); | ||
3271 | } else { | ||
3272 | pmcraid_err("queuecommand could not build ioadl\n"); | ||
3273 | pmcraid_return_cmd(cmd); | ||
3274 | rc = SCSI_MLQUEUE_HOST_BUSY; | ||
3275 | } | ||
3276 | |||
3277 | return rc; | ||
3278 | } | ||
3279 | |||
3280 | /** | ||
3281 | * pmcraid_open -char node "open" entry, allowed only users with admin access | ||
3282 | */ | ||
3283 | static int pmcraid_chr_open(struct inode *inode, struct file *filep) | ||
3284 | { | ||
3285 | struct pmcraid_instance *pinstance; | ||
3286 | |||
3287 | if (!capable(CAP_SYS_ADMIN)) | ||
3288 | return -EACCES; | ||
3289 | |||
3290 | /* Populate adapter instance * pointer for use by ioctl */ | ||
3291 | pinstance = container_of(inode->i_cdev, struct pmcraid_instance, cdev); | ||
3292 | filep->private_data = pinstance; | ||
3293 | |||
3294 | return 0; | ||
3295 | } | ||
3296 | |||
3297 | /** | ||
3298 | * pmcraid_release - char node "release" entry point | ||
3299 | */ | ||
3300 | static int pmcraid_chr_release(struct inode *inode, struct file *filep) | ||
3301 | { | ||
3302 | struct pmcraid_instance *pinstance = | ||
3303 | ((struct pmcraid_instance *)filep->private_data); | ||
3304 | |||
3305 | filep->private_data = NULL; | ||
3306 | fasync_helper(-1, filep, 0, &pinstance->aen_queue); | ||
3307 | |||
3308 | return 0; | ||
3309 | } | ||
3310 | |||
3311 | /** | ||
3312 | * pmcraid_fasync - Async notifier registration from applications | ||
3313 | * | ||
3314 | * This function adds the calling process to a driver global queue. When an | ||
3315 | * event occurs, SIGIO will be sent to all processes in this queue. | ||
3316 | */ | ||
3317 | static int pmcraid_chr_fasync(int fd, struct file *filep, int mode) | ||
3318 | { | ||
3319 | struct pmcraid_instance *pinstance; | ||
3320 | int rc; | ||
3321 | |||
3322 | pinstance = (struct pmcraid_instance *)filep->private_data; | ||
3323 | mutex_lock(&pinstance->aen_queue_lock); | ||
3324 | rc = fasync_helper(fd, filep, mode, &pinstance->aen_queue); | ||
3325 | mutex_unlock(&pinstance->aen_queue_lock); | ||
3326 | |||
3327 | return rc; | ||
3328 | } | ||
3329 | |||
3330 | |||
3331 | /** | ||
3332 | * pmcraid_build_passthrough_ioadls - builds SG elements for passthrough | ||
3333 | * commands sent over IOCTL interface | ||
3334 | * | ||
3335 | * @cmd : pointer to struct pmcraid_cmd | ||
3336 | * @buflen : length of the request buffer | ||
3337 | * @direction : data transfer direction | ||
3338 | * | ||
3339 | * Return value | ||
3340 | * 0 on sucess, non-zero error code on failure | ||
3341 | */ | ||
3342 | static int pmcraid_build_passthrough_ioadls( | ||
3343 | struct pmcraid_cmd *cmd, | ||
3344 | int buflen, | ||
3345 | int direction | ||
3346 | ) | ||
3347 | { | ||
3348 | struct pmcraid_sglist *sglist = NULL; | ||
3349 | struct scatterlist *sg = NULL; | ||
3350 | struct pmcraid_ioarcb *ioarcb = &cmd->ioa_cb->ioarcb; | ||
3351 | struct pmcraid_ioadl_desc *ioadl; | ||
3352 | int i; | ||
3353 | |||
3354 | sglist = pmcraid_alloc_sglist(buflen); | ||
3355 | |||
3356 | if (!sglist) { | ||
3357 | pmcraid_err("can't allocate memory for passthrough SGls\n"); | ||
3358 | return -ENOMEM; | ||
3359 | } | ||
3360 | |||
3361 | sglist->num_dma_sg = pci_map_sg(cmd->drv_inst->pdev, | ||
3362 | sglist->scatterlist, | ||
3363 | sglist->num_sg, direction); | ||
3364 | |||
3365 | if (!sglist->num_dma_sg || sglist->num_dma_sg > PMCRAID_MAX_IOADLS) { | ||
3366 | dev_err(&cmd->drv_inst->pdev->dev, | ||
3367 | "Failed to map passthrough buffer!\n"); | ||
3368 | pmcraid_free_sglist(sglist); | ||
3369 | return -EIO; | ||
3370 | } | ||
3371 | |||
3372 | cmd->sglist = sglist; | ||
3373 | ioarcb->request_flags0 |= NO_LINK_DESCS; | ||
3374 | |||
3375 | ioadl = pmcraid_init_ioadls(cmd, sglist->num_dma_sg); | ||
3376 | |||
3377 | /* Initialize IOADL descriptor addresses */ | ||
3378 | for_each_sg(sglist->scatterlist, sg, sglist->num_dma_sg, i) { | ||
3379 | ioadl[i].data_len = cpu_to_le32(sg_dma_len(sg)); | ||
3380 | ioadl[i].address = cpu_to_le64(sg_dma_address(sg)); | ||
3381 | ioadl[i].flags = 0; | ||
3382 | } | ||
3383 | |||
3384 | /* setup the last descriptor */ | ||
3385 | ioadl[i - 1].flags = cpu_to_le32(IOADL_FLAGS_LAST_DESC); | ||
3386 | |||
3387 | return 0; | ||
3388 | } | ||
3389 | |||
3390 | |||
3391 | /** | ||
3392 | * pmcraid_release_passthrough_ioadls - release passthrough ioadls | ||
3393 | * | ||
3394 | * @cmd: pointer to struct pmcraid_cmd for which ioadls were allocated | ||
3395 | * @buflen: size of the request buffer | ||
3396 | * @direction: data transfer direction | ||
3397 | * | ||
3398 | * Return value | ||
3399 | * 0 on sucess, non-zero error code on failure | ||
3400 | */ | ||
3401 | static void pmcraid_release_passthrough_ioadls( | ||
3402 | struct pmcraid_cmd *cmd, | ||
3403 | int buflen, | ||
3404 | int direction | ||
3405 | ) | ||
3406 | { | ||
3407 | struct pmcraid_sglist *sglist = cmd->sglist; | ||
3408 | |||
3409 | if (buflen > 0) { | ||
3410 | pci_unmap_sg(cmd->drv_inst->pdev, | ||
3411 | sglist->scatterlist, | ||
3412 | sglist->num_sg, | ||
3413 | direction); | ||
3414 | pmcraid_free_sglist(sglist); | ||
3415 | cmd->sglist = NULL; | ||
3416 | } | ||
3417 | } | ||
3418 | |||
3419 | /** | ||
3420 | * pmcraid_ioctl_passthrough - handling passthrough IOCTL commands | ||
3421 | * | ||
3422 | * @pinstance: pointer to adapter instance structure | ||
3423 | * @cmd: ioctl code | ||
3424 | * @arg: pointer to pmcraid_passthrough_buffer user buffer | ||
3425 | * | ||
3426 | * Return value | ||
3427 | * 0 on sucess, non-zero error code on failure | ||
3428 | */ | ||
3429 | static long pmcraid_ioctl_passthrough( | ||
3430 | struct pmcraid_instance *pinstance, | ||
3431 | unsigned int ioctl_cmd, | ||
3432 | unsigned int buflen, | ||
3433 | unsigned long arg | ||
3434 | ) | ||
3435 | { | ||
3436 | struct pmcraid_passthrough_ioctl_buffer *buffer; | ||
3437 | struct pmcraid_ioarcb *ioarcb; | ||
3438 | struct pmcraid_cmd *cmd; | ||
3439 | struct pmcraid_cmd *cancel_cmd; | ||
3440 | unsigned long request_buffer; | ||
3441 | unsigned long request_offset; | ||
3442 | unsigned long lock_flags; | ||
3443 | int request_size; | ||
3444 | int buffer_size; | ||
3445 | u8 access, direction; | ||
3446 | int rc = 0; | ||
3447 | |||
3448 | /* If IOA reset is in progress, wait 10 secs for reset to complete */ | ||
3449 | if (pinstance->ioa_reset_in_progress) { | ||
3450 | rc = wait_event_interruptible_timeout( | ||
3451 | pinstance->reset_wait_q, | ||
3452 | !pinstance->ioa_reset_in_progress, | ||
3453 | msecs_to_jiffies(10000)); | ||
3454 | |||
3455 | if (!rc) | ||
3456 | return -ETIMEDOUT; | ||
3457 | else if (rc < 0) | ||
3458 | return -ERESTARTSYS; | ||
3459 | } | ||
3460 | |||
3461 | /* If adapter is not in operational state, return error */ | ||
3462 | if (pinstance->ioa_state != IOA_STATE_OPERATIONAL) { | ||
3463 | pmcraid_err("IOA is not operational\n"); | ||
3464 | return -ENOTTY; | ||
3465 | } | ||
3466 | |||
3467 | buffer_size = sizeof(struct pmcraid_passthrough_ioctl_buffer); | ||
3468 | buffer = kmalloc(buffer_size, GFP_KERNEL); | ||
3469 | |||
3470 | if (!buffer) { | ||
3471 | pmcraid_err("no memory for passthrough buffer\n"); | ||
3472 | return -ENOMEM; | ||
3473 | } | ||
3474 | |||
3475 | request_offset = | ||
3476 | offsetof(struct pmcraid_passthrough_ioctl_buffer, request_buffer); | ||
3477 | |||
3478 | request_buffer = arg + request_offset; | ||
3479 | |||
3480 | rc = __copy_from_user(buffer, | ||
3481 | (struct pmcraid_passthrough_ioctl_buffer *) arg, | ||
3482 | sizeof(struct pmcraid_passthrough_ioctl_buffer)); | ||
3483 | if (rc) { | ||
3484 | pmcraid_err("ioctl: can't copy passthrough buffer\n"); | ||
3485 | rc = -EFAULT; | ||
3486 | goto out_free_buffer; | ||
3487 | } | ||
3488 | |||
3489 | request_size = buffer->ioarcb.data_transfer_length; | ||
3490 | |||
3491 | if (buffer->ioarcb.request_flags0 & TRANSFER_DIR_WRITE) { | ||
3492 | access = VERIFY_READ; | ||
3493 | direction = DMA_TO_DEVICE; | ||
3494 | } else { | ||
3495 | access = VERIFY_WRITE; | ||
3496 | direction = DMA_FROM_DEVICE; | ||
3497 | } | ||
3498 | |||
3499 | if (request_size > 0) { | ||
3500 | rc = access_ok(access, arg, request_offset + request_size); | ||
3501 | |||
3502 | if (!rc) { | ||
3503 | rc = -EFAULT; | ||
3504 | goto out_free_buffer; | ||
3505 | } | ||
3506 | } | ||
3507 | |||
3508 | /* check if we have any additional command parameters */ | ||
3509 | if (buffer->ioarcb.add_cmd_param_length > PMCRAID_ADD_CMD_PARAM_LEN) { | ||
3510 | rc = -EINVAL; | ||
3511 | goto out_free_buffer; | ||
3512 | } | ||
3513 | |||
3514 | cmd = pmcraid_get_free_cmd(pinstance); | ||
3515 | |||
3516 | if (!cmd) { | ||
3517 | pmcraid_err("free command block is not available\n"); | ||
3518 | rc = -ENOMEM; | ||
3519 | goto out_free_buffer; | ||
3520 | } | ||
3521 | |||
3522 | cmd->scsi_cmd = NULL; | ||
3523 | ioarcb = &(cmd->ioa_cb->ioarcb); | ||
3524 | |||
3525 | /* Copy the user-provided IOARCB stuff field by field */ | ||
3526 | ioarcb->resource_handle = buffer->ioarcb.resource_handle; | ||
3527 | ioarcb->data_transfer_length = buffer->ioarcb.data_transfer_length; | ||
3528 | ioarcb->cmd_timeout = buffer->ioarcb.cmd_timeout; | ||
3529 | ioarcb->request_type = buffer->ioarcb.request_type; | ||
3530 | ioarcb->request_flags0 = buffer->ioarcb.request_flags0; | ||
3531 | ioarcb->request_flags1 = buffer->ioarcb.request_flags1; | ||
3532 | memcpy(ioarcb->cdb, buffer->ioarcb.cdb, PMCRAID_MAX_CDB_LEN); | ||
3533 | |||
3534 | if (buffer->ioarcb.add_cmd_param_length) { | ||
3535 | ioarcb->add_cmd_param_length = | ||
3536 | buffer->ioarcb.add_cmd_param_length; | ||
3537 | ioarcb->add_cmd_param_offset = | ||
3538 | buffer->ioarcb.add_cmd_param_offset; | ||
3539 | memcpy(ioarcb->add_data.u.add_cmd_params, | ||
3540 | buffer->ioarcb.add_data.u.add_cmd_params, | ||
3541 | buffer->ioarcb.add_cmd_param_length); | ||
3542 | } | ||
3543 | |||
3544 | if (request_size) { | ||
3545 | rc = pmcraid_build_passthrough_ioadls(cmd, | ||
3546 | request_size, | ||
3547 | direction); | ||
3548 | if (rc) { | ||
3549 | pmcraid_err("couldn't build passthrough ioadls\n"); | ||
3550 | goto out_free_buffer; | ||
3551 | } | ||
3552 | } | ||
3553 | |||
3554 | /* If data is being written into the device, copy the data from user | ||
3555 | * buffers | ||
3556 | */ | ||
3557 | if (direction == DMA_TO_DEVICE && request_size > 0) { | ||
3558 | rc = pmcraid_copy_sglist(cmd->sglist, | ||
3559 | request_buffer, | ||
3560 | request_size, | ||
3561 | direction); | ||
3562 | if (rc) { | ||
3563 | pmcraid_err("failed to copy user buffer\n"); | ||
3564 | goto out_free_sglist; | ||
3565 | } | ||
3566 | } | ||
3567 | |||
3568 | /* passthrough ioctl is a blocking command so, put the user to sleep | ||
3569 | * until timeout. Note that a timeout value of 0 means, do timeout. | ||
3570 | */ | ||
3571 | cmd->cmd_done = pmcraid_internal_done; | ||
3572 | init_completion(&cmd->wait_for_completion); | ||
3573 | cmd->completion_req = 1; | ||
3574 | |||
3575 | pmcraid_info("command(%d) (CDB[0] = %x) for %x\n", | ||
3576 | le32_to_cpu(cmd->ioa_cb->ioarcb.response_handle) >> 2, | ||
3577 | cmd->ioa_cb->ioarcb.cdb[0], | ||
3578 | le32_to_cpu(cmd->ioa_cb->ioarcb.resource_handle)); | ||
3579 | |||
3580 | spin_lock_irqsave(pinstance->host->host_lock, lock_flags); | ||
3581 | _pmcraid_fire_command(cmd); | ||
3582 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
3583 | |||
3584 | /* If command timeout is specified put caller to wait till that time, | ||
3585 | * otherwise it would be blocking wait. If command gets timed out, it | ||
3586 | * will be aborted. | ||
3587 | */ | ||
3588 | if (buffer->ioarcb.cmd_timeout == 0) { | ||
3589 | wait_for_completion(&cmd->wait_for_completion); | ||
3590 | } else if (!wait_for_completion_timeout( | ||
3591 | &cmd->wait_for_completion, | ||
3592 | msecs_to_jiffies(buffer->ioarcb.cmd_timeout * 1000))) { | ||
3593 | |||
3594 | pmcraid_info("aborting cmd %d (CDB[0] = %x) due to timeout\n", | ||
3595 | le32_to_cpu(cmd->ioa_cb->ioarcb.response_handle >> 2), | ||
3596 | cmd->ioa_cb->ioarcb.cdb[0]); | ||
3597 | |||
3598 | rc = -ETIMEDOUT; | ||
3599 | spin_lock_irqsave(pinstance->host->host_lock, lock_flags); | ||
3600 | cancel_cmd = pmcraid_abort_cmd(cmd); | ||
3601 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
3602 | |||
3603 | if (cancel_cmd) { | ||
3604 | wait_for_completion(&cancel_cmd->wait_for_completion); | ||
3605 | pmcraid_return_cmd(cancel_cmd); | ||
3606 | } | ||
3607 | |||
3608 | goto out_free_sglist; | ||
3609 | } | ||
3610 | |||
3611 | /* If the command failed for any reason, copy entire IOASA buffer and | ||
3612 | * return IOCTL success. If copying IOASA to user-buffer fails, return | ||
3613 | * EFAULT | ||
3614 | */ | ||
3615 | if (le32_to_cpu(cmd->ioa_cb->ioasa.ioasc)) { | ||
3616 | |||
3617 | void *ioasa = | ||
3618 | (void *)(arg + | ||
3619 | offsetof(struct pmcraid_passthrough_ioctl_buffer, ioasa)); | ||
3620 | |||
3621 | pmcraid_info("command failed with %x\n", | ||
3622 | le32_to_cpu(cmd->ioa_cb->ioasa.ioasc)); | ||
3623 | if (copy_to_user(ioasa, &cmd->ioa_cb->ioasa, | ||
3624 | sizeof(struct pmcraid_ioasa))) { | ||
3625 | pmcraid_err("failed to copy ioasa buffer to user\n"); | ||
3626 | rc = -EFAULT; | ||
3627 | } | ||
3628 | } | ||
3629 | /* If the data transfer was from device, copy the data onto user | ||
3630 | * buffers | ||
3631 | */ | ||
3632 | else if (direction == DMA_FROM_DEVICE && request_size > 0) { | ||
3633 | rc = pmcraid_copy_sglist(cmd->sglist, | ||
3634 | request_buffer, | ||
3635 | request_size, | ||
3636 | direction); | ||
3637 | if (rc) { | ||
3638 | pmcraid_err("failed to copy user buffer\n"); | ||
3639 | rc = -EFAULT; | ||
3640 | } | ||
3641 | } | ||
3642 | |||
3643 | out_free_sglist: | ||
3644 | pmcraid_release_passthrough_ioadls(cmd, request_size, direction); | ||
3645 | pmcraid_return_cmd(cmd); | ||
3646 | |||
3647 | out_free_buffer: | ||
3648 | kfree(buffer); | ||
3649 | |||
3650 | return rc; | ||
3651 | } | ||
3652 | |||
3653 | |||
3654 | |||
3655 | |||
3656 | /** | ||
3657 | * pmcraid_ioctl_driver - ioctl handler for commands handled by driver itself | ||
3658 | * | ||
3659 | * @pinstance: pointer to adapter instance structure | ||
3660 | * @cmd: ioctl command passed in | ||
3661 | * @buflen: length of user_buffer | ||
3662 | * @user_buffer: user buffer pointer | ||
3663 | * | ||
3664 | * Return Value | ||
3665 | * 0 in case of success, otherwise appropriate error code | ||
3666 | */ | ||
3667 | static long pmcraid_ioctl_driver( | ||
3668 | struct pmcraid_instance *pinstance, | ||
3669 | unsigned int cmd, | ||
3670 | unsigned int buflen, | ||
3671 | void __user *user_buffer | ||
3672 | ) | ||
3673 | { | ||
3674 | int rc = -ENOSYS; | ||
3675 | |||
3676 | if (!access_ok(VERIFY_READ, user_buffer, _IOC_SIZE(cmd))) { | ||
3677 | pmcraid_err("ioctl_driver: access fault in request buffer \n"); | ||
3678 | return -EFAULT; | ||
3679 | } | ||
3680 | |||
3681 | switch (cmd) { | ||
3682 | case PMCRAID_IOCTL_RESET_ADAPTER: | ||
3683 | pmcraid_reset_bringup(pinstance); | ||
3684 | rc = 0; | ||
3685 | break; | ||
3686 | |||
3687 | default: | ||
3688 | break; | ||
3689 | } | ||
3690 | |||
3691 | return rc; | ||
3692 | } | ||
3693 | |||
3694 | /** | ||
3695 | * pmcraid_check_ioctl_buffer - check for proper access to user buffer | ||
3696 | * | ||
3697 | * @cmd: ioctl command | ||
3698 | * @arg: user buffer | ||
3699 | * @hdr: pointer to kernel memory for pmcraid_ioctl_header | ||
3700 | * | ||
3701 | * Return Value | ||
3702 | * negetive error code if there are access issues, otherwise zero. | ||
3703 | * Upon success, returns ioctl header copied out of user buffer. | ||
3704 | */ | ||
3705 | |||
3706 | static int pmcraid_check_ioctl_buffer( | ||
3707 | int cmd, | ||
3708 | void __user *arg, | ||
3709 | struct pmcraid_ioctl_header *hdr | ||
3710 | ) | ||
3711 | { | ||
3712 | int rc = 0; | ||
3713 | int access = VERIFY_READ; | ||
3714 | |||
3715 | if (copy_from_user(hdr, arg, sizeof(struct pmcraid_ioctl_header))) { | ||
3716 | pmcraid_err("couldn't copy ioctl header from user buffer\n"); | ||
3717 | return -EFAULT; | ||
3718 | } | ||
3719 | |||
3720 | /* check for valid driver signature */ | ||
3721 | rc = memcmp(hdr->signature, | ||
3722 | PMCRAID_IOCTL_SIGNATURE, | ||
3723 | sizeof(hdr->signature)); | ||
3724 | if (rc) { | ||
3725 | pmcraid_err("signature verification failed\n"); | ||
3726 | return -EINVAL; | ||
3727 | } | ||
3728 | |||
3729 | /* buffer length can't be negetive */ | ||
3730 | if (hdr->buffer_length < 0) { | ||
3731 | pmcraid_err("ioctl: invalid buffer length specified\n"); | ||
3732 | return -EINVAL; | ||
3733 | } | ||
3734 | |||
3735 | /* check for appropriate buffer access */ | ||
3736 | if ((_IOC_DIR(cmd) & _IOC_READ) == _IOC_READ) | ||
3737 | access = VERIFY_WRITE; | ||
3738 | |||
3739 | rc = access_ok(access, | ||
3740 | (arg + sizeof(struct pmcraid_ioctl_header)), | ||
3741 | hdr->buffer_length); | ||
3742 | if (!rc) { | ||
3743 | pmcraid_err("access failed for user buffer of size %d\n", | ||
3744 | hdr->buffer_length); | ||
3745 | return -EFAULT; | ||
3746 | } | ||
3747 | |||
3748 | return 0; | ||
3749 | } | ||
3750 | |||
3751 | /** | ||
3752 | * pmcraid_ioctl - char node ioctl entry point | ||
3753 | */ | ||
3754 | static long pmcraid_chr_ioctl( | ||
3755 | struct file *filep, | ||
3756 | unsigned int cmd, | ||
3757 | unsigned long arg | ||
3758 | ) | ||
3759 | { | ||
3760 | struct pmcraid_instance *pinstance = NULL; | ||
3761 | struct pmcraid_ioctl_header *hdr = NULL; | ||
3762 | int retval = -ENOTTY; | ||
3763 | |||
3764 | hdr = kmalloc(GFP_KERNEL, sizeof(struct pmcraid_ioctl_header)); | ||
3765 | |||
3766 | if (!hdr) { | ||
3767 | pmcraid_err("faile to allocate memory for ioctl header\n"); | ||
3768 | return -ENOMEM; | ||
3769 | } | ||
3770 | |||
3771 | retval = pmcraid_check_ioctl_buffer(cmd, (void *)arg, hdr); | ||
3772 | |||
3773 | if (retval) { | ||
3774 | pmcraid_info("chr_ioctl: header check failed\n"); | ||
3775 | kfree(hdr); | ||
3776 | return retval; | ||
3777 | } | ||
3778 | |||
3779 | pinstance = (struct pmcraid_instance *)filep->private_data; | ||
3780 | |||
3781 | if (!pinstance) { | ||
3782 | pmcraid_info("adapter instance is not found\n"); | ||
3783 | kfree(hdr); | ||
3784 | return -ENOTTY; | ||
3785 | } | ||
3786 | |||
3787 | switch (_IOC_TYPE(cmd)) { | ||
3788 | |||
3789 | case PMCRAID_PASSTHROUGH_IOCTL: | ||
3790 | /* If ioctl code is to download microcode, we need to block | ||
3791 | * mid-layer requests. | ||
3792 | */ | ||
3793 | if (cmd == PMCRAID_IOCTL_DOWNLOAD_MICROCODE) | ||
3794 | scsi_block_requests(pinstance->host); | ||
3795 | |||
3796 | retval = pmcraid_ioctl_passthrough(pinstance, | ||
3797 | cmd, | ||
3798 | hdr->buffer_length, | ||
3799 | arg); | ||
3800 | |||
3801 | if (cmd == PMCRAID_IOCTL_DOWNLOAD_MICROCODE) | ||
3802 | scsi_unblock_requests(pinstance->host); | ||
3803 | break; | ||
3804 | |||
3805 | case PMCRAID_DRIVER_IOCTL: | ||
3806 | arg += sizeof(struct pmcraid_ioctl_header); | ||
3807 | retval = pmcraid_ioctl_driver(pinstance, | ||
3808 | cmd, | ||
3809 | hdr->buffer_length, | ||
3810 | (void __user *)arg); | ||
3811 | break; | ||
3812 | |||
3813 | default: | ||
3814 | retval = -ENOTTY; | ||
3815 | break; | ||
3816 | } | ||
3817 | |||
3818 | kfree(hdr); | ||
3819 | |||
3820 | return retval; | ||
3821 | } | ||
3822 | |||
3823 | /** | ||
3824 | * File operations structure for management interface | ||
3825 | */ | ||
3826 | static const struct file_operations pmcraid_fops = { | ||
3827 | .owner = THIS_MODULE, | ||
3828 | .open = pmcraid_chr_open, | ||
3829 | .release = pmcraid_chr_release, | ||
3830 | .fasync = pmcraid_chr_fasync, | ||
3831 | .unlocked_ioctl = pmcraid_chr_ioctl, | ||
3832 | #ifdef CONFIG_COMPAT | ||
3833 | .compat_ioctl = pmcraid_chr_ioctl, | ||
3834 | #endif | ||
3835 | }; | ||
3836 | |||
3837 | |||
3838 | |||
3839 | |||
3840 | /** | ||
3841 | * pmcraid_show_log_level - Display adapter's error logging level | ||
3842 | * @dev: class device struct | ||
3843 | * @buf: buffer | ||
3844 | * | ||
3845 | * Return value: | ||
3846 | * number of bytes printed to buffer | ||
3847 | */ | ||
3848 | static ssize_t pmcraid_show_log_level( | ||
3849 | struct device *dev, | ||
3850 | struct device_attribute *attr, | ||
3851 | char *buf) | ||
3852 | { | ||
3853 | struct Scsi_Host *shost = class_to_shost(dev); | ||
3854 | struct pmcraid_instance *pinstance = | ||
3855 | (struct pmcraid_instance *)shost->hostdata; | ||
3856 | return snprintf(buf, PAGE_SIZE, "%d\n", pinstance->current_log_level); | ||
3857 | } | ||
3858 | |||
3859 | /** | ||
3860 | * pmcraid_store_log_level - Change the adapter's error logging level | ||
3861 | * @dev: class device struct | ||
3862 | * @buf: buffer | ||
3863 | * @count: not used | ||
3864 | * | ||
3865 | * Return value: | ||
3866 | * number of bytes printed to buffer | ||
3867 | */ | ||
3868 | static ssize_t pmcraid_store_log_level( | ||
3869 | struct device *dev, | ||
3870 | struct device_attribute *attr, | ||
3871 | const char *buf, | ||
3872 | size_t count | ||
3873 | ) | ||
3874 | { | ||
3875 | struct Scsi_Host *shost; | ||
3876 | struct pmcraid_instance *pinstance; | ||
3877 | unsigned long val; | ||
3878 | |||
3879 | if (strict_strtoul(buf, 10, &val)) | ||
3880 | return -EINVAL; | ||
3881 | /* log-level should be from 0 to 2 */ | ||
3882 | if (val > 2) | ||
3883 | return -EINVAL; | ||
3884 | |||
3885 | shost = class_to_shost(dev); | ||
3886 | pinstance = (struct pmcraid_instance *)shost->hostdata; | ||
3887 | pinstance->current_log_level = val; | ||
3888 | |||
3889 | return strlen(buf); | ||
3890 | } | ||
3891 | |||
3892 | static struct device_attribute pmcraid_log_level_attr = { | ||
3893 | .attr = { | ||
3894 | .name = "log_level", | ||
3895 | .mode = S_IRUGO | S_IWUSR, | ||
3896 | }, | ||
3897 | .show = pmcraid_show_log_level, | ||
3898 | .store = pmcraid_store_log_level, | ||
3899 | }; | ||
3900 | |||
3901 | /** | ||
3902 | * pmcraid_show_drv_version - Display driver version | ||
3903 | * @dev: class device struct | ||
3904 | * @buf: buffer | ||
3905 | * | ||
3906 | * Return value: | ||
3907 | * number of bytes printed to buffer | ||
3908 | */ | ||
3909 | static ssize_t pmcraid_show_drv_version( | ||
3910 | struct device *dev, | ||
3911 | struct device_attribute *attr, | ||
3912 | char *buf | ||
3913 | ) | ||
3914 | { | ||
3915 | return snprintf(buf, PAGE_SIZE, "version: %s, build date: %s\n", | ||
3916 | PMCRAID_DRIVER_VERSION, PMCRAID_DRIVER_DATE); | ||
3917 | } | ||
3918 | |||
3919 | static struct device_attribute pmcraid_driver_version_attr = { | ||
3920 | .attr = { | ||
3921 | .name = "drv_version", | ||
3922 | .mode = S_IRUGO, | ||
3923 | }, | ||
3924 | .show = pmcraid_show_drv_version, | ||
3925 | }; | ||
3926 | |||
3927 | /** | ||
3928 | * pmcraid_show_io_adapter_id - Display driver assigned adapter id | ||
3929 | * @dev: class device struct | ||
3930 | * @buf: buffer | ||
3931 | * | ||
3932 | * Return value: | ||
3933 | * number of bytes printed to buffer | ||
3934 | */ | ||
3935 | static ssize_t pmcraid_show_adapter_id( | ||
3936 | struct device *dev, | ||
3937 | struct device_attribute *attr, | ||
3938 | char *buf | ||
3939 | ) | ||
3940 | { | ||
3941 | struct Scsi_Host *shost = class_to_shost(dev); | ||
3942 | struct pmcraid_instance *pinstance = | ||
3943 | (struct pmcraid_instance *)shost->hostdata; | ||
3944 | u32 adapter_id = (pinstance->pdev->bus->number << 8) | | ||
3945 | pinstance->pdev->devfn; | ||
3946 | u32 aen_group = pmcraid_event_family.id; | ||
3947 | |||
3948 | return snprintf(buf, PAGE_SIZE, | ||
3949 | "adapter id: %d\nminor: %d\naen group: %d\n", | ||
3950 | adapter_id, MINOR(pinstance->cdev.dev), aen_group); | ||
3951 | } | ||
3952 | |||
3953 | static struct device_attribute pmcraid_adapter_id_attr = { | ||
3954 | .attr = { | ||
3955 | .name = "adapter_id", | ||
3956 | .mode = S_IRUGO | S_IWUSR, | ||
3957 | }, | ||
3958 | .show = pmcraid_show_adapter_id, | ||
3959 | }; | ||
3960 | |||
3961 | static struct device_attribute *pmcraid_host_attrs[] = { | ||
3962 | &pmcraid_log_level_attr, | ||
3963 | &pmcraid_driver_version_attr, | ||
3964 | &pmcraid_adapter_id_attr, | ||
3965 | NULL, | ||
3966 | }; | ||
3967 | |||
3968 | |||
3969 | /* host template structure for pmcraid driver */ | ||
3970 | static struct scsi_host_template pmcraid_host_template = { | ||
3971 | .module = THIS_MODULE, | ||
3972 | .name = PMCRAID_DRIVER_NAME, | ||
3973 | .queuecommand = pmcraid_queuecommand, | ||
3974 | .eh_abort_handler = pmcraid_eh_abort_handler, | ||
3975 | .eh_bus_reset_handler = pmcraid_eh_bus_reset_handler, | ||
3976 | .eh_target_reset_handler = pmcraid_eh_target_reset_handler, | ||
3977 | .eh_device_reset_handler = pmcraid_eh_device_reset_handler, | ||
3978 | .eh_host_reset_handler = pmcraid_eh_host_reset_handler, | ||
3979 | |||
3980 | .slave_alloc = pmcraid_slave_alloc, | ||
3981 | .slave_configure = pmcraid_slave_configure, | ||
3982 | .slave_destroy = pmcraid_slave_destroy, | ||
3983 | .change_queue_depth = pmcraid_change_queue_depth, | ||
3984 | .change_queue_type = pmcraid_change_queue_type, | ||
3985 | .can_queue = PMCRAID_MAX_IO_CMD, | ||
3986 | .this_id = -1, | ||
3987 | .sg_tablesize = PMCRAID_MAX_IOADLS, | ||
3988 | .max_sectors = PMCRAID_IOA_MAX_SECTORS, | ||
3989 | .cmd_per_lun = PMCRAID_MAX_CMD_PER_LUN, | ||
3990 | .use_clustering = ENABLE_CLUSTERING, | ||
3991 | .shost_attrs = pmcraid_host_attrs, | ||
3992 | .proc_name = PMCRAID_DRIVER_NAME | ||
3993 | }; | ||
3994 | |||
3995 | /** | ||
3996 | * pmcraid_isr_common - Common interrupt handler routine | ||
3997 | * | ||
3998 | * @pinstance: pointer to adapter instance | ||
3999 | * @intrs: active interrupts (contents of ioa_host_interrupt register) | ||
4000 | * @hrrq_id: Host RRQ index | ||
4001 | * | ||
4002 | * Return Value | ||
4003 | * none | ||
4004 | */ | ||
4005 | static void pmcraid_isr_common( | ||
4006 | struct pmcraid_instance *pinstance, | ||
4007 | u32 intrs, | ||
4008 | int hrrq_id | ||
4009 | ) | ||
4010 | { | ||
4011 | u32 intrs_clear = | ||
4012 | (intrs & INTRS_CRITICAL_OP_IN_PROGRESS) ? intrs | ||
4013 | : INTRS_HRRQ_VALID; | ||
4014 | iowrite32(intrs_clear, | ||
4015 | pinstance->int_regs.ioa_host_interrupt_clr_reg); | ||
4016 | intrs = ioread32(pinstance->int_regs.ioa_host_interrupt_reg); | ||
4017 | |||
4018 | /* hrrq valid bit was set, schedule tasklet to handle the response */ | ||
4019 | if (intrs_clear == INTRS_HRRQ_VALID) | ||
4020 | tasklet_schedule(&(pinstance->isr_tasklet[hrrq_id])); | ||
4021 | } | ||
4022 | |||
4023 | /** | ||
4024 | * pmcraid_isr - implements interrupt handling routine | ||
4025 | * | ||
4026 | * @irq: interrupt vector number | ||
4027 | * @dev_id: pointer hrrq_vector | ||
4028 | * | ||
4029 | * Return Value | ||
4030 | * IRQ_HANDLED if interrupt is handled or IRQ_NONE if ignored | ||
4031 | */ | ||
4032 | static irqreturn_t pmcraid_isr(int irq, void *dev_id) | ||
4033 | { | ||
4034 | struct pmcraid_isr_param *hrrq_vector; | ||
4035 | struct pmcraid_instance *pinstance; | ||
4036 | unsigned long lock_flags; | ||
4037 | u32 intrs; | ||
4038 | |||
4039 | /* In case of legacy interrupt mode where interrupts are shared across | ||
4040 | * isrs, it may be possible that the current interrupt is not from IOA | ||
4041 | */ | ||
4042 | if (!dev_id) { | ||
4043 | printk(KERN_INFO "%s(): NULL host pointer\n", __func__); | ||
4044 | return IRQ_NONE; | ||
4045 | } | ||
4046 | |||
4047 | hrrq_vector = (struct pmcraid_isr_param *)dev_id; | ||
4048 | pinstance = hrrq_vector->drv_inst; | ||
4049 | |||
4050 | /* Acquire the lock (currently host_lock) while processing interrupts. | ||
4051 | * This interval is small as most of the response processing is done by | ||
4052 | * tasklet without the lock. | ||
4053 | */ | ||
4054 | spin_lock_irqsave(pinstance->host->host_lock, lock_flags); | ||
4055 | intrs = pmcraid_read_interrupts(pinstance); | ||
4056 | |||
4057 | if (unlikely((intrs & PMCRAID_PCI_INTERRUPTS) == 0)) { | ||
4058 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
4059 | return IRQ_NONE; | ||
4060 | } | ||
4061 | |||
4062 | /* Any error interrupts including unit_check, initiate IOA reset. | ||
4063 | * In case of unit check indicate to reset_sequence that IOA unit | ||
4064 | * checked and prepare for a dump during reset sequence | ||
4065 | */ | ||
4066 | if (intrs & PMCRAID_ERROR_INTERRUPTS) { | ||
4067 | |||
4068 | if (intrs & INTRS_IOA_UNIT_CHECK) | ||
4069 | pinstance->ioa_unit_check = 1; | ||
4070 | |||
4071 | iowrite32(intrs, | ||
4072 | pinstance->int_regs.ioa_host_interrupt_clr_reg); | ||
4073 | pmcraid_err("ISR: error interrupts: %x initiating reset\n", | ||
4074 | intrs); | ||
4075 | intrs = ioread32(pinstance->int_regs.ioa_host_interrupt_reg); | ||
4076 | pmcraid_initiate_reset(pinstance); | ||
4077 | } else { | ||
4078 | pmcraid_isr_common(pinstance, intrs, hrrq_vector->hrrq_id); | ||
4079 | } | ||
4080 | |||
4081 | spin_unlock_irqrestore(pinstance->host->host_lock, lock_flags); | ||
4082 | |||
4083 | return IRQ_HANDLED; | ||
4084 | } | ||
4085 | |||
4086 | |||
4087 | /** | ||
4088 | * pmcraid_worker_function - worker thread function | ||
4089 | * | ||
4090 | * @workp: pointer to struct work queue | ||
4091 | * | ||
4092 | * Return Value | ||
4093 | * None | ||
4094 | */ | ||
4095 | |||
4096 | static void pmcraid_worker_function(struct work_struct *workp) | ||
4097 | { | ||
4098 | struct pmcraid_instance *pinstance; | ||
4099 | struct pmcraid_resource_entry *res; | ||
4100 | struct pmcraid_resource_entry *temp; | ||
4101 | struct scsi_device *sdev; | ||
4102 | unsigned long lock_flags; | ||
4103 | unsigned long host_lock_flags; | ||
4104 | u8 bus, target, lun; | ||
4105 | |||
4106 | pinstance = container_of(workp, struct pmcraid_instance, worker_q); | ||
4107 | /* add resources only after host is added into system */ | ||
4108 | if (!atomic_read(&pinstance->expose_resources)) | ||
4109 | return; | ||
4110 | |||
4111 | spin_lock_irqsave(&pinstance->resource_lock, lock_flags); | ||
4112 | list_for_each_entry_safe(res, temp, &pinstance->used_res_q, queue) { | ||
4113 | |||
4114 | if (res->change_detected == RES_CHANGE_DEL && res->scsi_dev) { | ||
4115 | sdev = res->scsi_dev; | ||
4116 | |||
4117 | /* host_lock must be held before calling | ||
4118 | * scsi_device_get | ||
4119 | */ | ||
4120 | spin_lock_irqsave(pinstance->host->host_lock, | ||
4121 | host_lock_flags); | ||
4122 | if (!scsi_device_get(sdev)) { | ||
4123 | spin_unlock_irqrestore( | ||
4124 | pinstance->host->host_lock, | ||
4125 | host_lock_flags); | ||
4126 | pmcraid_info("deleting %x from midlayer\n", | ||
4127 | res->cfg_entry.resource_address); | ||
4128 | list_move_tail(&res->queue, | ||
4129 | &pinstance->free_res_q); | ||
4130 | spin_unlock_irqrestore( | ||
4131 | &pinstance->resource_lock, | ||
4132 | lock_flags); | ||
4133 | scsi_remove_device(sdev); | ||
4134 | scsi_device_put(sdev); | ||
4135 | spin_lock_irqsave(&pinstance->resource_lock, | ||
4136 | lock_flags); | ||
4137 | res->change_detected = 0; | ||
4138 | } else { | ||
4139 | spin_unlock_irqrestore( | ||
4140 | pinstance->host->host_lock, | ||
4141 | host_lock_flags); | ||
4142 | } | ||
4143 | } | ||
4144 | } | ||
4145 | |||
4146 | list_for_each_entry(res, &pinstance->used_res_q, queue) { | ||
4147 | |||
4148 | if (res->change_detected == RES_CHANGE_ADD) { | ||
4149 | |||
4150 | if (!pmcraid_expose_resource(&res->cfg_entry)) | ||
4151 | continue; | ||
4152 | |||
4153 | if (RES_IS_VSET(res->cfg_entry)) { | ||
4154 | bus = PMCRAID_VSET_BUS_ID; | ||
4155 | target = res->cfg_entry.unique_flags1; | ||
4156 | lun = PMCRAID_VSET_LUN_ID; | ||
4157 | } else { | ||
4158 | bus = PMCRAID_PHYS_BUS_ID; | ||
4159 | target = | ||
4160 | RES_TARGET( | ||
4161 | res->cfg_entry.resource_address); | ||
4162 | lun = RES_LUN(res->cfg_entry.resource_address); | ||
4163 | } | ||
4164 | |||
4165 | res->change_detected = 0; | ||
4166 | spin_unlock_irqrestore(&pinstance->resource_lock, | ||
4167 | lock_flags); | ||
4168 | scsi_add_device(pinstance->host, bus, target, lun); | ||
4169 | spin_lock_irqsave(&pinstance->resource_lock, | ||
4170 | lock_flags); | ||
4171 | } | ||
4172 | } | ||
4173 | |||
4174 | spin_unlock_irqrestore(&pinstance->resource_lock, lock_flags); | ||
4175 | } | ||
4176 | |||
4177 | /** | ||
4178 | * pmcraid_tasklet_function - Tasklet function | ||
4179 | * | ||
4180 | * @instance: pointer to msix param structure | ||
4181 | * | ||
4182 | * Return Value | ||
4183 | * None | ||
4184 | */ | ||
4185 | void pmcraid_tasklet_function(unsigned long instance) | ||
4186 | { | ||
4187 | struct pmcraid_isr_param *hrrq_vector; | ||
4188 | struct pmcraid_instance *pinstance; | ||
4189 | unsigned long hrrq_lock_flags; | ||
4190 | unsigned long pending_lock_flags; | ||
4191 | unsigned long host_lock_flags; | ||
4192 | spinlock_t *lockp; /* hrrq buffer lock */ | ||
4193 | int id; | ||
4194 | u32 intrs; | ||
4195 | __le32 resp; | ||
4196 | |||
4197 | hrrq_vector = (struct pmcraid_isr_param *)instance; | ||
4198 | pinstance = hrrq_vector->drv_inst; | ||
4199 | id = hrrq_vector->hrrq_id; | ||
4200 | lockp = &(pinstance->hrrq_lock[id]); | ||
4201 | intrs = pmcraid_read_interrupts(pinstance); | ||
4202 | |||
4203 | /* If interrupts was as part of the ioa initialization, clear and mask | ||
4204 | * it. Delete the timer and wakeup the reset engine to proceed with | ||
4205 | * reset sequence | ||
4206 | */ | ||
4207 | if (intrs & INTRS_TRANSITION_TO_OPERATIONAL) { | ||
4208 | iowrite32(INTRS_TRANSITION_TO_OPERATIONAL, | ||
4209 | pinstance->int_regs.ioa_host_interrupt_mask_reg); | ||
4210 | iowrite32(INTRS_TRANSITION_TO_OPERATIONAL, | ||
4211 | pinstance->int_regs.ioa_host_interrupt_clr_reg); | ||
4212 | |||
4213 | if (pinstance->reset_cmd != NULL) { | ||
4214 | del_timer(&pinstance->reset_cmd->timer); | ||
4215 | spin_lock_irqsave(pinstance->host->host_lock, | ||
4216 | host_lock_flags); | ||
4217 | pinstance->reset_cmd->cmd_done(pinstance->reset_cmd); | ||
4218 | spin_unlock_irqrestore(pinstance->host->host_lock, | ||
4219 | host_lock_flags); | ||
4220 | } | ||
4221 | return; | ||
4222 | } | ||
4223 | |||
4224 | /* loop through each of the commands responded by IOA. Each HRRQ buf is | ||
4225 | * protected by its own lock. Traversals must be done within this lock | ||
4226 | * as there may be multiple tasklets running on multiple CPUs. Note | ||
4227 | * that the lock is held just for picking up the response handle and | ||
4228 | * manipulating hrrq_curr/toggle_bit values. | ||
4229 | */ | ||
4230 | spin_lock_irqsave(lockp, hrrq_lock_flags); | ||
4231 | |||
4232 | resp = le32_to_cpu(*(pinstance->hrrq_curr[id])); | ||
4233 | |||
4234 | while ((resp & HRRQ_TOGGLE_BIT) == | ||
4235 | pinstance->host_toggle_bit[id]) { | ||
4236 | |||
4237 | int cmd_index = resp >> 2; | ||
4238 | struct pmcraid_cmd *cmd = NULL; | ||
4239 | |||
4240 | if (cmd_index < PMCRAID_MAX_CMD) { | ||
4241 | cmd = pinstance->cmd_list[cmd_index]; | ||
4242 | } else { | ||
4243 | /* In case of invalid response handle, initiate IOA | ||
4244 | * reset sequence. | ||
4245 | */ | ||
4246 | spin_unlock_irqrestore(lockp, hrrq_lock_flags); | ||
4247 | |||
4248 | pmcraid_err("Invalid response %d initiating reset\n", | ||
4249 | cmd_index); | ||
4250 | |||
4251 | spin_lock_irqsave(pinstance->host->host_lock, | ||
4252 | host_lock_flags); | ||
4253 | pmcraid_initiate_reset(pinstance); | ||
4254 | spin_unlock_irqrestore(pinstance->host->host_lock, | ||
4255 | host_lock_flags); | ||
4256 | |||
4257 | spin_lock_irqsave(lockp, hrrq_lock_flags); | ||
4258 | break; | ||
4259 | } | ||
4260 | |||
4261 | if (pinstance->hrrq_curr[id] < pinstance->hrrq_end[id]) { | ||
4262 | pinstance->hrrq_curr[id]++; | ||
4263 | } else { | ||
4264 | pinstance->hrrq_curr[id] = pinstance->hrrq_start[id]; | ||
4265 | pinstance->host_toggle_bit[id] ^= 1u; | ||
4266 | } | ||
4267 | |||
4268 | spin_unlock_irqrestore(lockp, hrrq_lock_flags); | ||
4269 | |||
4270 | spin_lock_irqsave(&pinstance->pending_pool_lock, | ||
4271 | pending_lock_flags); | ||
4272 | list_del(&cmd->free_list); | ||
4273 | spin_unlock_irqrestore(&pinstance->pending_pool_lock, | ||
4274 | pending_lock_flags); | ||
4275 | del_timer(&cmd->timer); | ||
4276 | atomic_dec(&pinstance->outstanding_cmds); | ||
4277 | |||
4278 | if (cmd->cmd_done == pmcraid_ioa_reset) { | ||
4279 | spin_lock_irqsave(pinstance->host->host_lock, | ||
4280 | host_lock_flags); | ||
4281 | cmd->cmd_done(cmd); | ||
4282 | spin_unlock_irqrestore(pinstance->host->host_lock, | ||
4283 | host_lock_flags); | ||
4284 | } else if (cmd->cmd_done != NULL) { | ||
4285 | cmd->cmd_done(cmd); | ||
4286 | } | ||
4287 | /* loop over until we are done with all responses */ | ||
4288 | spin_lock_irqsave(lockp, hrrq_lock_flags); | ||
4289 | resp = le32_to_cpu(*(pinstance->hrrq_curr[id])); | ||
4290 | } | ||
4291 | |||
4292 | spin_unlock_irqrestore(lockp, hrrq_lock_flags); | ||
4293 | } | ||
4294 | |||
4295 | /** | ||
4296 | * pmcraid_unregister_interrupt_handler - de-register interrupts handlers | ||
4297 | * @pinstance: pointer to adapter instance structure | ||
4298 | * | ||
4299 | * This routine un-registers registered interrupt handler and | ||
4300 | * also frees irqs/vectors. | ||
4301 | * | ||
4302 | * Retun Value | ||
4303 | * None | ||
4304 | */ | ||
4305 | static | ||
4306 | void pmcraid_unregister_interrupt_handler(struct pmcraid_instance *pinstance) | ||
4307 | { | ||
4308 | free_irq(pinstance->pdev->irq, &(pinstance->hrrq_vector[0])); | ||
4309 | } | ||
4310 | |||
4311 | /** | ||
4312 | * pmcraid_register_interrupt_handler - registers interrupt handler | ||
4313 | * @pinstance: pointer to per-adapter instance structure | ||
4314 | * | ||
4315 | * Return Value | ||
4316 | * 0 on success, non-zero error code otherwise. | ||
4317 | */ | ||
4318 | static int | ||
4319 | pmcraid_register_interrupt_handler(struct pmcraid_instance *pinstance) | ||
4320 | { | ||
4321 | struct pci_dev *pdev = pinstance->pdev; | ||
4322 | |||
4323 | pinstance->hrrq_vector[0].hrrq_id = 0; | ||
4324 | pinstance->hrrq_vector[0].drv_inst = pinstance; | ||
4325 | pinstance->hrrq_vector[0].vector = 0; | ||
4326 | pinstance->num_hrrq = 1; | ||
4327 | return request_irq(pdev->irq, pmcraid_isr, IRQF_SHARED, | ||
4328 | PMCRAID_DRIVER_NAME, &pinstance->hrrq_vector[0]); | ||
4329 | } | ||
4330 | |||
4331 | /** | ||
4332 | * pmcraid_release_cmd_blocks - release buufers allocated for command blocks | ||
4333 | * @pinstance: per adapter instance structure pointer | ||
4334 | * @max_index: number of buffer blocks to release | ||
4335 | * | ||
4336 | * Return Value | ||
4337 | * None | ||
4338 | */ | ||
4339 | static void | ||
4340 | pmcraid_release_cmd_blocks(struct pmcraid_instance *pinstance, int max_index) | ||
4341 | { | ||
4342 | int i; | ||
4343 | for (i = 0; i < max_index; i++) { | ||
4344 | kmem_cache_free(pinstance->cmd_cachep, pinstance->cmd_list[i]); | ||
4345 | pinstance->cmd_list[i] = NULL; | ||
4346 | } | ||
4347 | kmem_cache_destroy(pinstance->cmd_cachep); | ||
4348 | pinstance->cmd_cachep = NULL; | ||
4349 | } | ||
4350 | |||
4351 | /** | ||
4352 | * pmcraid_release_control_blocks - releases buffers alloced for control blocks | ||
4353 | * @pinstance: pointer to per adapter instance structure | ||
4354 | * @max_index: number of buffers (from 0 onwards) to release | ||
4355 | * | ||
4356 | * This function assumes that the command blocks for which control blocks are | ||
4357 | * linked are not released. | ||
4358 | * | ||
4359 | * Return Value | ||
4360 | * None | ||
4361 | */ | ||
4362 | static void | ||
4363 | pmcraid_release_control_blocks( | ||
4364 | struct pmcraid_instance *pinstance, | ||
4365 | int max_index | ||
4366 | ) | ||
4367 | { | ||
4368 | int i; | ||
4369 | |||
4370 | if (pinstance->control_pool == NULL) | ||
4371 | return; | ||
4372 | |||
4373 | for (i = 0; i < max_index; i++) { | ||
4374 | pci_pool_free(pinstance->control_pool, | ||
4375 | pinstance->cmd_list[i]->ioa_cb, | ||
4376 | pinstance->cmd_list[i]->ioa_cb_bus_addr); | ||
4377 | pinstance->cmd_list[i]->ioa_cb = NULL; | ||
4378 | pinstance->cmd_list[i]->ioa_cb_bus_addr = 0; | ||
4379 | } | ||
4380 | pci_pool_destroy(pinstance->control_pool); | ||
4381 | pinstance->control_pool = NULL; | ||
4382 | } | ||
4383 | |||
4384 | /** | ||
4385 | * pmcraid_allocate_cmd_blocks - allocate memory for cmd block structures | ||
4386 | * @pinstance - pointer to per adapter instance structure | ||
4387 | * | ||
4388 | * Allocates memory for command blocks using kernel slab allocator. | ||
4389 | * | ||
4390 | * Return Value | ||
4391 | * 0 in case of success; -ENOMEM in case of failure | ||
4392 | */ | ||
4393 | static int __devinit | ||
4394 | pmcraid_allocate_cmd_blocks(struct pmcraid_instance *pinstance) | ||
4395 | { | ||
4396 | int i; | ||
4397 | |||
4398 | sprintf(pinstance->cmd_pool_name, "pmcraid_cmd_pool_%d", | ||
4399 | pinstance->host->unique_id); | ||
4400 | |||
4401 | |||
4402 | pinstance->cmd_cachep = kmem_cache_create( | ||
4403 | pinstance->cmd_pool_name, | ||
4404 | sizeof(struct pmcraid_cmd), 0, | ||
4405 | SLAB_HWCACHE_ALIGN, NULL); | ||
4406 | if (!pinstance->cmd_cachep) | ||
4407 | return -ENOMEM; | ||
4408 | |||
4409 | for (i = 0; i < PMCRAID_MAX_CMD; i++) { | ||
4410 | pinstance->cmd_list[i] = | ||
4411 | kmem_cache_alloc(pinstance->cmd_cachep, GFP_KERNEL); | ||
4412 | if (!pinstance->cmd_list[i]) { | ||
4413 | pmcraid_release_cmd_blocks(pinstance, i); | ||
4414 | return -ENOMEM; | ||
4415 | } | ||
4416 | } | ||
4417 | return 0; | ||
4418 | } | ||
4419 | |||
4420 | /** | ||
4421 | * pmcraid_allocate_control_blocks - allocates memory control blocks | ||
4422 | * @pinstance : pointer to per adapter instance structure | ||
4423 | * | ||
4424 | * This function allocates PCI memory for DMAable buffers like IOARCB, IOADLs | ||
4425 | * and IOASAs. This is called after command blocks are already allocated. | ||
4426 | * | ||
4427 | * Return Value | ||
4428 | * 0 in case it can allocate all control blocks, otherwise -ENOMEM | ||
4429 | */ | ||
4430 | static int __devinit | ||
4431 | pmcraid_allocate_control_blocks(struct pmcraid_instance *pinstance) | ||
4432 | { | ||
4433 | int i; | ||
4434 | |||
4435 | sprintf(pinstance->ctl_pool_name, "pmcraid_control_pool_%d", | ||
4436 | pinstance->host->unique_id); | ||
4437 | |||
4438 | pinstance->control_pool = | ||
4439 | pci_pool_create(pinstance->ctl_pool_name, | ||
4440 | pinstance->pdev, | ||
4441 | sizeof(struct pmcraid_control_block), | ||
4442 | PMCRAID_IOARCB_ALIGNMENT, 0); | ||
4443 | |||
4444 | if (!pinstance->control_pool) | ||
4445 | return -ENOMEM; | ||
4446 | |||
4447 | for (i = 0; i < PMCRAID_MAX_CMD; i++) { | ||
4448 | pinstance->cmd_list[i]->ioa_cb = | ||
4449 | pci_pool_alloc( | ||
4450 | pinstance->control_pool, | ||
4451 | GFP_KERNEL, | ||
4452 | &(pinstance->cmd_list[i]->ioa_cb_bus_addr)); | ||
4453 | |||
4454 | if (!pinstance->cmd_list[i]->ioa_cb) { | ||
4455 | pmcraid_release_control_blocks(pinstance, i); | ||
4456 | return -ENOMEM; | ||
4457 | } | ||
4458 | memset(pinstance->cmd_list[i]->ioa_cb, 0, | ||
4459 | sizeof(struct pmcraid_control_block)); | ||
4460 | } | ||
4461 | return 0; | ||
4462 | } | ||
4463 | |||
4464 | /** | ||
4465 | * pmcraid_release_host_rrqs - release memory allocated for hrrq buffer(s) | ||
4466 | * @pinstance: pointer to per adapter instance structure | ||
4467 | * @maxindex: size of hrrq buffer pointer array | ||
4468 | * | ||
4469 | * Return Value | ||
4470 | * None | ||
4471 | */ | ||
4472 | static void | ||
4473 | pmcraid_release_host_rrqs(struct pmcraid_instance *pinstance, int maxindex) | ||
4474 | { | ||
4475 | int i; | ||
4476 | for (i = 0; i < maxindex; i++) { | ||
4477 | |||
4478 | pci_free_consistent(pinstance->pdev, | ||
4479 | HRRQ_ENTRY_SIZE * PMCRAID_MAX_CMD, | ||
4480 | pinstance->hrrq_start[i], | ||
4481 | pinstance->hrrq_start_bus_addr[i]); | ||
4482 | |||
4483 | /* reset pointers and toggle bit to zeros */ | ||
4484 | pinstance->hrrq_start[i] = NULL; | ||
4485 | pinstance->hrrq_start_bus_addr[i] = 0; | ||
4486 | pinstance->host_toggle_bit[i] = 0; | ||
4487 | } | ||
4488 | } | ||
4489 | |||
4490 | /** | ||
4491 | * pmcraid_allocate_host_rrqs - Allocate and initialize host RRQ buffers | ||
4492 | * @pinstance: pointer to per adapter instance structure | ||
4493 | * | ||
4494 | * Return value | ||
4495 | * 0 hrrq buffers are allocated, -ENOMEM otherwise. | ||
4496 | */ | ||
4497 | static int __devinit | ||
4498 | pmcraid_allocate_host_rrqs(struct pmcraid_instance *pinstance) | ||
4499 | { | ||
4500 | int i; | ||
4501 | int buf_count = PMCRAID_MAX_CMD / pinstance->num_hrrq; | ||
4502 | |||
4503 | for (i = 0; i < pinstance->num_hrrq; i++) { | ||
4504 | int buffer_size = HRRQ_ENTRY_SIZE * buf_count; | ||
4505 | |||
4506 | pinstance->hrrq_start[i] = | ||
4507 | pci_alloc_consistent( | ||
4508 | pinstance->pdev, | ||
4509 | buffer_size, | ||
4510 | &(pinstance->hrrq_start_bus_addr[i])); | ||
4511 | |||
4512 | if (pinstance->hrrq_start[i] == 0) { | ||
4513 | pmcraid_err("could not allocate host rrq: %d\n", i); | ||
4514 | pmcraid_release_host_rrqs(pinstance, i); | ||
4515 | return -ENOMEM; | ||
4516 | } | ||
4517 | |||
4518 | memset(pinstance->hrrq_start[i], 0, buffer_size); | ||
4519 | pinstance->hrrq_curr[i] = pinstance->hrrq_start[i]; | ||
4520 | pinstance->hrrq_end[i] = | ||
4521 | pinstance->hrrq_start[i] + buf_count - 1; | ||
4522 | pinstance->host_toggle_bit[i] = 1; | ||
4523 | spin_lock_init(&pinstance->hrrq_lock[i]); | ||
4524 | } | ||
4525 | return 0; | ||
4526 | } | ||
4527 | |||
4528 | /** | ||
4529 | * pmcraid_release_hcams - release HCAM buffers | ||
4530 | * | ||
4531 | * @pinstance: pointer to per adapter instance structure | ||
4532 | * | ||
4533 | * Return value | ||
4534 | * none | ||
4535 | */ | ||
4536 | static void pmcraid_release_hcams(struct pmcraid_instance *pinstance) | ||
4537 | { | ||
4538 | if (pinstance->ccn.msg != NULL) { | ||
4539 | pci_free_consistent(pinstance->pdev, | ||
4540 | PMCRAID_AEN_HDR_SIZE + | ||
4541 | sizeof(struct pmcraid_hcam_ccn), | ||
4542 | pinstance->ccn.msg, | ||
4543 | pinstance->ccn.baddr); | ||
4544 | |||
4545 | pinstance->ccn.msg = NULL; | ||
4546 | pinstance->ccn.hcam = NULL; | ||
4547 | pinstance->ccn.baddr = 0; | ||
4548 | } | ||
4549 | |||
4550 | if (pinstance->ldn.msg != NULL) { | ||
4551 | pci_free_consistent(pinstance->pdev, | ||
4552 | PMCRAID_AEN_HDR_SIZE + | ||
4553 | sizeof(struct pmcraid_hcam_ldn), | ||
4554 | pinstance->ldn.msg, | ||
4555 | pinstance->ldn.baddr); | ||
4556 | |||
4557 | pinstance->ldn.msg = NULL; | ||
4558 | pinstance->ldn.hcam = NULL; | ||
4559 | pinstance->ldn.baddr = 0; | ||
4560 | } | ||
4561 | } | ||
4562 | |||
4563 | /** | ||
4564 | * pmcraid_allocate_hcams - allocates HCAM buffers | ||
4565 | * @pinstance : pointer to per adapter instance structure | ||
4566 | * | ||
4567 | * Return Value: | ||
4568 | * 0 in case of successful allocation, non-zero otherwise | ||
4569 | */ | ||
4570 | static int pmcraid_allocate_hcams(struct pmcraid_instance *pinstance) | ||
4571 | { | ||
4572 | pinstance->ccn.msg = pci_alloc_consistent( | ||
4573 | pinstance->pdev, | ||
4574 | PMCRAID_AEN_HDR_SIZE + | ||
4575 | sizeof(struct pmcraid_hcam_ccn), | ||
4576 | &(pinstance->ccn.baddr)); | ||
4577 | |||
4578 | pinstance->ldn.msg = pci_alloc_consistent( | ||
4579 | pinstance->pdev, | ||
4580 | PMCRAID_AEN_HDR_SIZE + | ||
4581 | sizeof(struct pmcraid_hcam_ldn), | ||
4582 | &(pinstance->ldn.baddr)); | ||
4583 | |||
4584 | if (pinstance->ldn.msg == NULL || pinstance->ccn.msg == NULL) { | ||
4585 | pmcraid_release_hcams(pinstance); | ||
4586 | } else { | ||
4587 | pinstance->ccn.hcam = | ||
4588 | (void *)pinstance->ccn.msg + PMCRAID_AEN_HDR_SIZE; | ||
4589 | pinstance->ldn.hcam = | ||
4590 | (void *)pinstance->ldn.msg + PMCRAID_AEN_HDR_SIZE; | ||
4591 | |||
4592 | atomic_set(&pinstance->ccn.ignore, 0); | ||
4593 | atomic_set(&pinstance->ldn.ignore, 0); | ||
4594 | } | ||
4595 | |||
4596 | return (pinstance->ldn.msg == NULL) ? -ENOMEM : 0; | ||
4597 | } | ||
4598 | |||
4599 | /** | ||
4600 | * pmcraid_release_config_buffers - release config.table buffers | ||
4601 | * @pinstance: pointer to per adapter instance structure | ||
4602 | * | ||
4603 | * Return Value | ||
4604 | * none | ||
4605 | */ | ||
4606 | static void pmcraid_release_config_buffers(struct pmcraid_instance *pinstance) | ||
4607 | { | ||
4608 | if (pinstance->cfg_table != NULL && | ||
4609 | pinstance->cfg_table_bus_addr != 0) { | ||
4610 | pci_free_consistent(pinstance->pdev, | ||
4611 | sizeof(struct pmcraid_config_table), | ||
4612 | pinstance->cfg_table, | ||
4613 | pinstance->cfg_table_bus_addr); | ||
4614 | pinstance->cfg_table = NULL; | ||
4615 | pinstance->cfg_table_bus_addr = 0; | ||
4616 | } | ||
4617 | |||
4618 | if (pinstance->res_entries != NULL) { | ||
4619 | int i; | ||
4620 | |||
4621 | for (i = 0; i < PMCRAID_MAX_RESOURCES; i++) | ||
4622 | list_del(&pinstance->res_entries[i].queue); | ||
4623 | kfree(pinstance->res_entries); | ||
4624 | pinstance->res_entries = NULL; | ||
4625 | } | ||
4626 | |||
4627 | pmcraid_release_hcams(pinstance); | ||
4628 | } | ||
4629 | |||
4630 | /** | ||
4631 | * pmcraid_allocate_config_buffers - allocates DMAable memory for config table | ||
4632 | * @pinstance : pointer to per adapter instance structure | ||
4633 | * | ||
4634 | * Return Value | ||
4635 | * 0 for successful allocation, -ENOMEM for any failure | ||
4636 | */ | ||
4637 | static int __devinit | ||
4638 | pmcraid_allocate_config_buffers(struct pmcraid_instance *pinstance) | ||
4639 | { | ||
4640 | int i; | ||
4641 | |||
4642 | pinstance->res_entries = | ||
4643 | kzalloc(sizeof(struct pmcraid_resource_entry) * | ||
4644 | PMCRAID_MAX_RESOURCES, GFP_KERNEL); | ||
4645 | |||
4646 | if (NULL == pinstance->res_entries) { | ||
4647 | pmcraid_err("failed to allocate memory for resource table\n"); | ||
4648 | return -ENOMEM; | ||
4649 | } | ||
4650 | |||
4651 | for (i = 0; i < PMCRAID_MAX_RESOURCES; i++) | ||
4652 | list_add_tail(&pinstance->res_entries[i].queue, | ||
4653 | &pinstance->free_res_q); | ||
4654 | |||
4655 | pinstance->cfg_table = | ||
4656 | pci_alloc_consistent(pinstance->pdev, | ||
4657 | sizeof(struct pmcraid_config_table), | ||
4658 | &pinstance->cfg_table_bus_addr); | ||
4659 | |||
4660 | if (NULL == pinstance->cfg_table) { | ||
4661 | pmcraid_err("couldn't alloc DMA memory for config table\n"); | ||
4662 | pmcraid_release_config_buffers(pinstance); | ||
4663 | return -ENOMEM; | ||
4664 | } | ||
4665 | |||
4666 | if (pmcraid_allocate_hcams(pinstance)) { | ||
4667 | pmcraid_err("could not alloc DMA memory for HCAMS\n"); | ||
4668 | pmcraid_release_config_buffers(pinstance); | ||
4669 | return -ENOMEM; | ||
4670 | } | ||
4671 | |||
4672 | return 0; | ||
4673 | } | ||
4674 | |||
4675 | /** | ||
4676 | * pmcraid_init_tasklets - registers tasklets for response handling | ||
4677 | * | ||
4678 | * @pinstance: pointer adapter instance structure | ||
4679 | * | ||
4680 | * Return value | ||
4681 | * none | ||
4682 | */ | ||
4683 | static void pmcraid_init_tasklets(struct pmcraid_instance *pinstance) | ||
4684 | { | ||
4685 | int i; | ||
4686 | for (i = 0; i < pinstance->num_hrrq; i++) | ||
4687 | tasklet_init(&pinstance->isr_tasklet[i], | ||
4688 | pmcraid_tasklet_function, | ||
4689 | (unsigned long)&pinstance->hrrq_vector[i]); | ||
4690 | } | ||
4691 | |||
4692 | /** | ||
4693 | * pmcraid_kill_tasklets - destroys tasklets registered for response handling | ||
4694 | * | ||
4695 | * @pinstance: pointer to adapter instance structure | ||
4696 | * | ||
4697 | * Return value | ||
4698 | * none | ||
4699 | */ | ||
4700 | static void pmcraid_kill_tasklets(struct pmcraid_instance *pinstance) | ||
4701 | { | ||
4702 | int i; | ||
4703 | for (i = 0; i < pinstance->num_hrrq; i++) | ||
4704 | tasklet_kill(&pinstance->isr_tasklet[i]); | ||
4705 | } | ||
4706 | |||
4707 | /** | ||
4708 | * pmcraid_init_buffers - allocates memory and initializes various structures | ||
4709 | * @pinstance: pointer to per adapter instance structure | ||
4710 | * | ||
4711 | * This routine pre-allocates memory based on the type of block as below: | ||
4712 | * cmdblocks(PMCRAID_MAX_CMD): kernel memory using kernel's slab_allocator, | ||
4713 | * IOARCBs(PMCRAID_MAX_CMD) : DMAable memory, using pci pool allocator | ||
4714 | * config-table entries : DMAable memory using pci_alloc_consistent | ||
4715 | * HostRRQs : DMAable memory, using pci_alloc_consistent | ||
4716 | * | ||
4717 | * Return Value | ||
4718 | * 0 in case all of the blocks are allocated, -ENOMEM otherwise. | ||
4719 | */ | ||
4720 | static int __devinit pmcraid_init_buffers(struct pmcraid_instance *pinstance) | ||
4721 | { | ||
4722 | int i; | ||
4723 | |||
4724 | if (pmcraid_allocate_host_rrqs(pinstance)) { | ||
4725 | pmcraid_err("couldn't allocate memory for %d host rrqs\n", | ||
4726 | pinstance->num_hrrq); | ||
4727 | return -ENOMEM; | ||
4728 | } | ||
4729 | |||
4730 | if (pmcraid_allocate_config_buffers(pinstance)) { | ||
4731 | pmcraid_err("couldn't allocate memory for config buffers\n"); | ||
4732 | pmcraid_release_host_rrqs(pinstance, pinstance->num_hrrq); | ||
4733 | return -ENOMEM; | ||
4734 | } | ||
4735 | |||
4736 | if (pmcraid_allocate_cmd_blocks(pinstance)) { | ||
4737 | pmcraid_err("couldn't allocate memory for cmd blocks \n"); | ||
4738 | pmcraid_release_config_buffers(pinstance); | ||
4739 | pmcraid_release_host_rrqs(pinstance, pinstance->num_hrrq); | ||
4740 | return -ENOMEM; | ||
4741 | } | ||
4742 | |||
4743 | if (pmcraid_allocate_control_blocks(pinstance)) { | ||
4744 | pmcraid_err("couldn't allocate memory control blocks \n"); | ||
4745 | pmcraid_release_config_buffers(pinstance); | ||
4746 | pmcraid_release_cmd_blocks(pinstance, PMCRAID_MAX_CMD); | ||
4747 | pmcraid_release_host_rrqs(pinstance, pinstance->num_hrrq); | ||
4748 | return -ENOMEM; | ||
4749 | } | ||
4750 | |||
4751 | /* Initialize all the command blocks and add them to free pool. No | ||
4752 | * need to lock (free_pool_lock) as this is done in initialization | ||
4753 | * itself | ||
4754 | */ | ||
4755 | for (i = 0; i < PMCRAID_MAX_CMD; i++) { | ||
4756 | struct pmcraid_cmd *cmdp = pinstance->cmd_list[i]; | ||
4757 | pmcraid_init_cmdblk(cmdp, i); | ||
4758 | cmdp->drv_inst = pinstance; | ||
4759 | list_add_tail(&cmdp->free_list, &pinstance->free_cmd_pool); | ||
4760 | } | ||
4761 | |||
4762 | return 0; | ||
4763 | } | ||
4764 | |||
4765 | /** | ||
4766 | * pmcraid_reinit_buffers - resets various buffer pointers | ||
4767 | * @pinstance: pointer to adapter instance | ||
4768 | * Return value | ||
4769 | * none | ||
4770 | */ | ||
4771 | static void pmcraid_reinit_buffers(struct pmcraid_instance *pinstance) | ||
4772 | { | ||
4773 | int i; | ||
4774 | int buffer_size = HRRQ_ENTRY_SIZE * PMCRAID_MAX_CMD; | ||
4775 | |||
4776 | for (i = 0; i < pinstance->num_hrrq; i++) { | ||
4777 | memset(pinstance->hrrq_start[i], 0, buffer_size); | ||
4778 | pinstance->hrrq_curr[i] = pinstance->hrrq_start[i]; | ||
4779 | pinstance->hrrq_end[i] = | ||
4780 | pinstance->hrrq_start[i] + PMCRAID_MAX_CMD - 1; | ||
4781 | pinstance->host_toggle_bit[i] = 1; | ||
4782 | } | ||
4783 | } | ||
4784 | |||
4785 | /** | ||
4786 | * pmcraid_init_instance - initialize per instance data structure | ||
4787 | * @pdev: pointer to pci device structure | ||
4788 | * @host: pointer to Scsi_Host structure | ||
4789 | * @mapped_pci_addr: memory mapped IOA configuration registers | ||
4790 | * | ||
4791 | * Return Value | ||
4792 | * 0 on success, non-zero in case of any failure | ||
4793 | */ | ||
4794 | static int __devinit pmcraid_init_instance( | ||
4795 | struct pci_dev *pdev, | ||
4796 | struct Scsi_Host *host, | ||
4797 | void __iomem *mapped_pci_addr | ||
4798 | ) | ||
4799 | { | ||
4800 | struct pmcraid_instance *pinstance = | ||
4801 | (struct pmcraid_instance *)host->hostdata; | ||
4802 | |||
4803 | pinstance->host = host; | ||
4804 | pinstance->pdev = pdev; | ||
4805 | |||
4806 | /* Initialize register addresses */ | ||
4807 | pinstance->mapped_dma_addr = mapped_pci_addr; | ||
4808 | |||
4809 | /* Initialize chip-specific details */ | ||
4810 | { | ||
4811 | struct pmcraid_chip_details *chip_cfg = pinstance->chip_cfg; | ||
4812 | struct pmcraid_interrupts *pint_regs = &pinstance->int_regs; | ||
4813 | |||
4814 | pinstance->ioarrin = mapped_pci_addr + chip_cfg->ioarrin; | ||
4815 | |||
4816 | pint_regs->ioa_host_interrupt_reg = | ||
4817 | mapped_pci_addr + chip_cfg->ioa_host_intr; | ||
4818 | pint_regs->ioa_host_interrupt_clr_reg = | ||
4819 | mapped_pci_addr + chip_cfg->ioa_host_intr_clr; | ||
4820 | pint_regs->host_ioa_interrupt_reg = | ||
4821 | mapped_pci_addr + chip_cfg->host_ioa_intr; | ||
4822 | pint_regs->host_ioa_interrupt_clr_reg = | ||
4823 | mapped_pci_addr + chip_cfg->host_ioa_intr_clr; | ||
4824 | |||
4825 | /* Current version of firmware exposes interrupt mask set | ||
4826 | * and mask clr registers through memory mapped bar0. | ||
4827 | */ | ||
4828 | pinstance->mailbox = mapped_pci_addr + chip_cfg->mailbox; | ||
4829 | pinstance->ioa_status = mapped_pci_addr + chip_cfg->ioastatus; | ||
4830 | pint_regs->ioa_host_interrupt_mask_reg = | ||
4831 | mapped_pci_addr + chip_cfg->ioa_host_mask; | ||
4832 | pint_regs->ioa_host_interrupt_mask_clr_reg = | ||
4833 | mapped_pci_addr + chip_cfg->ioa_host_mask_clr; | ||
4834 | pint_regs->global_interrupt_mask_reg = | ||
4835 | mapped_pci_addr + chip_cfg->global_intr_mask; | ||
4836 | }; | ||
4837 | |||
4838 | pinstance->ioa_reset_attempts = 0; | ||
4839 | init_waitqueue_head(&pinstance->reset_wait_q); | ||
4840 | |||
4841 | atomic_set(&pinstance->outstanding_cmds, 0); | ||
4842 | atomic_set(&pinstance->expose_resources, 0); | ||
4843 | |||
4844 | INIT_LIST_HEAD(&pinstance->free_res_q); | ||
4845 | INIT_LIST_HEAD(&pinstance->used_res_q); | ||
4846 | INIT_LIST_HEAD(&pinstance->free_cmd_pool); | ||
4847 | INIT_LIST_HEAD(&pinstance->pending_cmd_pool); | ||
4848 | |||
4849 | spin_lock_init(&pinstance->free_pool_lock); | ||
4850 | spin_lock_init(&pinstance->pending_pool_lock); | ||
4851 | spin_lock_init(&pinstance->resource_lock); | ||
4852 | mutex_init(&pinstance->aen_queue_lock); | ||
4853 | |||
4854 | /* Work-queue (Shared) for deferred processing error handling */ | ||
4855 | INIT_WORK(&pinstance->worker_q, pmcraid_worker_function); | ||
4856 | |||
4857 | /* Initialize the default log_level */ | ||
4858 | pinstance->current_log_level = pmcraid_log_level; | ||
4859 | |||
4860 | /* Setup variables required for reset engine */ | ||
4861 | pinstance->ioa_state = IOA_STATE_UNKNOWN; | ||
4862 | pinstance->reset_cmd = NULL; | ||
4863 | return 0; | ||
4864 | } | ||
4865 | |||
4866 | /** | ||
4867 | * pmcraid_release_buffers - release per-adapter buffers allocated | ||
4868 | * | ||
4869 | * @pinstance: pointer to adapter soft state | ||
4870 | * | ||
4871 | * Return Value | ||
4872 | * none | ||
4873 | */ | ||
4874 | static void pmcraid_release_buffers(struct pmcraid_instance *pinstance) | ||
4875 | { | ||
4876 | pmcraid_release_config_buffers(pinstance); | ||
4877 | pmcraid_release_control_blocks(pinstance, PMCRAID_MAX_CMD); | ||
4878 | pmcraid_release_cmd_blocks(pinstance, PMCRAID_MAX_CMD); | ||
4879 | pmcraid_release_host_rrqs(pinstance, pinstance->num_hrrq); | ||
4880 | |||
4881 | } | ||
4882 | |||
4883 | /** | ||
4884 | * pmcraid_shutdown - shutdown adapter controller. | ||
4885 | * @pdev: pci device struct | ||
4886 | * | ||
4887 | * Issues an adapter shutdown to the card waits for its completion | ||
4888 | * | ||
4889 | * Return value | ||
4890 | * none | ||
4891 | */ | ||
4892 | static void pmcraid_shutdown(struct pci_dev *pdev) | ||
4893 | { | ||
4894 | struct pmcraid_instance *pinstance = pci_get_drvdata(pdev); | ||
4895 | pmcraid_reset_bringdown(pinstance); | ||
4896 | } | ||
4897 | |||
4898 | |||
4899 | /** | ||
4900 | * pmcraid_get_minor - returns unused minor number from minor number bitmap | ||
4901 | */ | ||
4902 | static unsigned short pmcraid_get_minor(void) | ||
4903 | { | ||
4904 | int minor; | ||
4905 | |||
4906 | minor = find_first_zero_bit(pmcraid_minor, sizeof(pmcraid_minor)); | ||
4907 | __set_bit(minor, pmcraid_minor); | ||
4908 | return minor; | ||
4909 | } | ||
4910 | |||
4911 | /** | ||
4912 | * pmcraid_release_minor - releases given minor back to minor number bitmap | ||
4913 | */ | ||
4914 | static void pmcraid_release_minor(unsigned short minor) | ||
4915 | { | ||
4916 | __clear_bit(minor, pmcraid_minor); | ||
4917 | } | ||
4918 | |||
4919 | /** | ||
4920 | * pmcraid_setup_chrdev - allocates a minor number and registers a char device | ||
4921 | * | ||
4922 | * @pinstance: pointer to adapter instance for which to register device | ||
4923 | * | ||
4924 | * Return value | ||
4925 | * 0 in case of success, otherwise non-zero | ||
4926 | */ | ||
4927 | static int pmcraid_setup_chrdev(struct pmcraid_instance *pinstance) | ||
4928 | { | ||
4929 | int minor; | ||
4930 | int error; | ||
4931 | |||
4932 | minor = pmcraid_get_minor(); | ||
4933 | cdev_init(&pinstance->cdev, &pmcraid_fops); | ||
4934 | pinstance->cdev.owner = THIS_MODULE; | ||
4935 | |||
4936 | error = cdev_add(&pinstance->cdev, MKDEV(pmcraid_major, minor), 1); | ||
4937 | |||
4938 | if (error) | ||
4939 | pmcraid_release_minor(minor); | ||
4940 | else | ||
4941 | device_create(pmcraid_class, NULL, MKDEV(pmcraid_major, minor), | ||
4942 | NULL, "pmcsas%u", minor); | ||
4943 | return error; | ||
4944 | } | ||
4945 | |||
4946 | /** | ||
4947 | * pmcraid_release_chrdev - unregisters per-adapter management interface | ||
4948 | * | ||
4949 | * @pinstance: pointer to adapter instance structure | ||
4950 | * | ||
4951 | * Return value | ||
4952 | * none | ||
4953 | */ | ||
4954 | static void pmcraid_release_chrdev(struct pmcraid_instance *pinstance) | ||
4955 | { | ||
4956 | pmcraid_release_minor(MINOR(pinstance->cdev.dev)); | ||
4957 | device_destroy(pmcraid_class, | ||
4958 | MKDEV(pmcraid_major, MINOR(pinstance->cdev.dev))); | ||
4959 | cdev_del(&pinstance->cdev); | ||
4960 | } | ||
4961 | |||
4962 | /** | ||
4963 | * pmcraid_remove - IOA hot plug remove entry point | ||
4964 | * @pdev: pci device struct | ||
4965 | * | ||
4966 | * Return value | ||
4967 | * none | ||
4968 | */ | ||
4969 | static void __devexit pmcraid_remove(struct pci_dev *pdev) | ||
4970 | { | ||
4971 | struct pmcraid_instance *pinstance = pci_get_drvdata(pdev); | ||
4972 | |||
4973 | /* remove the management interface (/dev file) for this device */ | ||
4974 | pmcraid_release_chrdev(pinstance); | ||
4975 | |||
4976 | /* remove host template from scsi midlayer */ | ||
4977 | scsi_remove_host(pinstance->host); | ||
4978 | |||
4979 | /* block requests from mid-layer */ | ||
4980 | scsi_block_requests(pinstance->host); | ||
4981 | |||
4982 | /* initiate shutdown adapter */ | ||
4983 | pmcraid_shutdown(pdev); | ||
4984 | |||
4985 | pmcraid_disable_interrupts(pinstance, ~0); | ||
4986 | flush_scheduled_work(); | ||
4987 | |||
4988 | pmcraid_kill_tasklets(pinstance); | ||
4989 | pmcraid_unregister_interrupt_handler(pinstance); | ||
4990 | pmcraid_release_buffers(pinstance); | ||
4991 | iounmap(pinstance->mapped_dma_addr); | ||
4992 | pci_release_regions(pdev); | ||
4993 | scsi_host_put(pinstance->host); | ||
4994 | pci_disable_device(pdev); | ||
4995 | |||
4996 | return; | ||
4997 | } | ||
4998 | |||
4999 | #ifdef CONFIG_PM | ||
5000 | /** | ||
5001 | * pmcraid_suspend - driver suspend entry point for power management | ||
5002 | * @pdev: PCI device structure | ||
5003 | * @state: PCI power state to suspend routine | ||
5004 | * | ||
5005 | * Return Value - 0 always | ||
5006 | */ | ||
5007 | static int pmcraid_suspend(struct pci_dev *pdev, pm_message_t state) | ||
5008 | { | ||
5009 | struct pmcraid_instance *pinstance = pci_get_drvdata(pdev); | ||
5010 | |||
5011 | pmcraid_shutdown(pdev); | ||
5012 | pmcraid_disable_interrupts(pinstance, ~0); | ||
5013 | pmcraid_kill_tasklets(pinstance); | ||
5014 | pci_set_drvdata(pinstance->pdev, pinstance); | ||
5015 | pmcraid_unregister_interrupt_handler(pinstance); | ||
5016 | pci_save_state(pdev); | ||
5017 | pci_disable_device(pdev); | ||
5018 | pci_set_power_state(pdev, pci_choose_state(pdev, state)); | ||
5019 | |||
5020 | return 0; | ||
5021 | } | ||
5022 | |||
5023 | /** | ||
5024 | * pmcraid_resume - driver resume entry point PCI power management | ||
5025 | * @pdev: PCI device structure | ||
5026 | * | ||
5027 | * Return Value - 0 in case of success. Error code in case of any failure | ||
5028 | */ | ||
5029 | static int pmcraid_resume(struct pci_dev *pdev) | ||
5030 | { | ||
5031 | struct pmcraid_instance *pinstance = pci_get_drvdata(pdev); | ||
5032 | struct Scsi_Host *host = pinstance->host; | ||
5033 | int rc; | ||
5034 | int hrrqs; | ||
5035 | |||
5036 | pci_set_power_state(pdev, PCI_D0); | ||
5037 | pci_enable_wake(pdev, PCI_D0, 0); | ||
5038 | pci_restore_state(pdev); | ||
5039 | |||
5040 | rc = pci_enable_device(pdev); | ||
5041 | |||
5042 | if (rc) { | ||
5043 | pmcraid_err("pmcraid: Enable device failed\n"); | ||
5044 | return rc; | ||
5045 | } | ||
5046 | |||
5047 | pci_set_master(pdev); | ||
5048 | |||
5049 | if ((sizeof(dma_addr_t) == 4) || | ||
5050 | pci_set_dma_mask(pdev, DMA_BIT_MASK(64))) | ||
5051 | rc = pci_set_dma_mask(pdev, DMA_BIT_MASK(32)); | ||
5052 | |||
5053 | if (rc == 0) | ||
5054 | rc = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32)); | ||
5055 | |||
5056 | if (rc != 0) { | ||
5057 | dev_err(&pdev->dev, "Failed to set PCI DMA mask\n"); | ||
5058 | goto disable_device; | ||
5059 | } | ||
5060 | |||
5061 | atomic_set(&pinstance->outstanding_cmds, 0); | ||
5062 | hrrqs = pinstance->num_hrrq; | ||
5063 | rc = pmcraid_register_interrupt_handler(pinstance); | ||
5064 | |||
5065 | if (rc) { | ||
5066 | pmcraid_err("resume: couldn't register interrupt handlers\n"); | ||
5067 | rc = -ENODEV; | ||
5068 | goto release_host; | ||
5069 | } | ||
5070 | |||
5071 | pmcraid_init_tasklets(pinstance); | ||
5072 | pmcraid_enable_interrupts(pinstance, PMCRAID_PCI_INTERRUPTS); | ||
5073 | |||
5074 | /* Start with hard reset sequence which brings up IOA to operational | ||
5075 | * state as well as completes the reset sequence. | ||
5076 | */ | ||
5077 | pinstance->ioa_hard_reset = 1; | ||
5078 | |||
5079 | /* Start IOA firmware initialization and bring card to Operational | ||
5080 | * state. | ||
5081 | */ | ||
5082 | if (pmcraid_reset_bringup(pinstance)) { | ||
5083 | pmcraid_err("couldn't initialize IOA \n"); | ||
5084 | rc = -ENODEV; | ||
5085 | goto release_tasklets; | ||
5086 | } | ||
5087 | |||
5088 | return 0; | ||
5089 | |||
5090 | release_tasklets: | ||
5091 | pmcraid_kill_tasklets(pinstance); | ||
5092 | pmcraid_unregister_interrupt_handler(pinstance); | ||
5093 | |||
5094 | release_host: | ||
5095 | scsi_host_put(host); | ||
5096 | |||
5097 | disable_device: | ||
5098 | pci_disable_device(pdev); | ||
5099 | |||
5100 | return rc; | ||
5101 | } | ||
5102 | |||
5103 | #else | ||
5104 | |||
5105 | #define pmcraid_suspend NULL | ||
5106 | #define pmcraid_resume NULL | ||
5107 | |||
5108 | #endif /* CONFIG_PM */ | ||
5109 | |||
5110 | /** | ||
5111 | * pmcraid_complete_ioa_reset - Called by either timer or tasklet during | ||
5112 | * completion of the ioa reset | ||
5113 | * @cmd: pointer to reset command block | ||
5114 | */ | ||
5115 | static void pmcraid_complete_ioa_reset(struct pmcraid_cmd *cmd) | ||
5116 | { | ||
5117 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
5118 | unsigned long flags; | ||
5119 | |||
5120 | spin_lock_irqsave(pinstance->host->host_lock, flags); | ||
5121 | pmcraid_ioa_reset(cmd); | ||
5122 | spin_unlock_irqrestore(pinstance->host->host_lock, flags); | ||
5123 | scsi_unblock_requests(pinstance->host); | ||
5124 | schedule_work(&pinstance->worker_q); | ||
5125 | } | ||
5126 | |||
5127 | /** | ||
5128 | * pmcraid_set_supported_devs - sends SET SUPPORTED DEVICES to IOAFP | ||
5129 | * | ||
5130 | * @cmd: pointer to pmcraid_cmd structure | ||
5131 | * | ||
5132 | * Return Value | ||
5133 | * 0 for success or non-zero for failure cases | ||
5134 | */ | ||
5135 | static void pmcraid_set_supported_devs(struct pmcraid_cmd *cmd) | ||
5136 | { | ||
5137 | struct pmcraid_ioarcb *ioarcb = &cmd->ioa_cb->ioarcb; | ||
5138 | void (*cmd_done) (struct pmcraid_cmd *) = pmcraid_complete_ioa_reset; | ||
5139 | |||
5140 | pmcraid_reinit_cmdblk(cmd); | ||
5141 | |||
5142 | ioarcb->resource_handle = cpu_to_le32(PMCRAID_IOA_RES_HANDLE); | ||
5143 | ioarcb->request_type = REQ_TYPE_IOACMD; | ||
5144 | ioarcb->cdb[0] = PMCRAID_SET_SUPPORTED_DEVICES; | ||
5145 | ioarcb->cdb[1] = ALL_DEVICES_SUPPORTED; | ||
5146 | |||
5147 | /* If this was called as part of resource table reinitialization due to | ||
5148 | * lost CCN, it is enough to return the command block back to free pool | ||
5149 | * as part of set_supported_devs completion function. | ||
5150 | */ | ||
5151 | if (cmd->drv_inst->reinit_cfg_table) { | ||
5152 | cmd->drv_inst->reinit_cfg_table = 0; | ||
5153 | cmd->release = 1; | ||
5154 | cmd_done = pmcraid_reinit_cfgtable_done; | ||
5155 | } | ||
5156 | |||
5157 | /* we will be done with the reset sequence after set supported devices, | ||
5158 | * setup the done function to return the command block back to free | ||
5159 | * pool | ||
5160 | */ | ||
5161 | pmcraid_send_cmd(cmd, | ||
5162 | cmd_done, | ||
5163 | PMCRAID_SET_SUP_DEV_TIMEOUT, | ||
5164 | pmcraid_timeout_handler); | ||
5165 | return; | ||
5166 | } | ||
5167 | |||
5168 | /** | ||
5169 | * pmcraid_init_res_table - Initialize the resource table | ||
5170 | * @cmd: pointer to pmcraid command struct | ||
5171 | * | ||
5172 | * This function looks through the existing resource table, comparing | ||
5173 | * it with the config table. This function will take care of old/new | ||
5174 | * devices and schedule adding/removing them from the mid-layer | ||
5175 | * as appropriate. | ||
5176 | * | ||
5177 | * Return value | ||
5178 | * None | ||
5179 | */ | ||
5180 | static void pmcraid_init_res_table(struct pmcraid_cmd *cmd) | ||
5181 | { | ||
5182 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
5183 | struct pmcraid_resource_entry *res, *temp; | ||
5184 | struct pmcraid_config_table_entry *cfgte; | ||
5185 | unsigned long lock_flags; | ||
5186 | int found, rc, i; | ||
5187 | LIST_HEAD(old_res); | ||
5188 | |||
5189 | if (pinstance->cfg_table->flags & MICROCODE_UPDATE_REQUIRED) | ||
5190 | dev_err(&pinstance->pdev->dev, "Require microcode download\n"); | ||
5191 | |||
5192 | /* resource list is protected by pinstance->resource_lock. | ||
5193 | * init_res_table can be called from probe (user-thread) or runtime | ||
5194 | * reset (timer/tasklet) | ||
5195 | */ | ||
5196 | spin_lock_irqsave(&pinstance->resource_lock, lock_flags); | ||
5197 | |||
5198 | list_for_each_entry_safe(res, temp, &pinstance->used_res_q, queue) | ||
5199 | list_move_tail(&res->queue, &old_res); | ||
5200 | |||
5201 | for (i = 0; i < pinstance->cfg_table->num_entries; i++) { | ||
5202 | cfgte = &pinstance->cfg_table->entries[i]; | ||
5203 | |||
5204 | if (!pmcraid_expose_resource(cfgte)) | ||
5205 | continue; | ||
5206 | |||
5207 | found = 0; | ||
5208 | |||
5209 | /* If this entry was already detected and initialized */ | ||
5210 | list_for_each_entry_safe(res, temp, &old_res, queue) { | ||
5211 | |||
5212 | rc = memcmp(&res->cfg_entry.resource_address, | ||
5213 | &cfgte->resource_address, | ||
5214 | sizeof(cfgte->resource_address)); | ||
5215 | if (!rc) { | ||
5216 | list_move_tail(&res->queue, | ||
5217 | &pinstance->used_res_q); | ||
5218 | found = 1; | ||
5219 | break; | ||
5220 | } | ||
5221 | } | ||
5222 | |||
5223 | /* If this is new entry, initialize it and add it the queue */ | ||
5224 | if (!found) { | ||
5225 | |||
5226 | if (list_empty(&pinstance->free_res_q)) { | ||
5227 | dev_err(&pinstance->pdev->dev, | ||
5228 | "Too many devices attached\n"); | ||
5229 | break; | ||
5230 | } | ||
5231 | |||
5232 | found = 1; | ||
5233 | res = list_entry(pinstance->free_res_q.next, | ||
5234 | struct pmcraid_resource_entry, queue); | ||
5235 | |||
5236 | res->scsi_dev = NULL; | ||
5237 | res->change_detected = RES_CHANGE_ADD; | ||
5238 | res->reset_progress = 0; | ||
5239 | list_move_tail(&res->queue, &pinstance->used_res_q); | ||
5240 | } | ||
5241 | |||
5242 | /* copy new configuration table entry details into driver | ||
5243 | * maintained resource entry | ||
5244 | */ | ||
5245 | if (found) { | ||
5246 | memcpy(&res->cfg_entry, cfgte, | ||
5247 | sizeof(struct pmcraid_config_table_entry)); | ||
5248 | pmcraid_info("New res type:%x, vset:%x, addr:%x:\n", | ||
5249 | res->cfg_entry.resource_type, | ||
5250 | res->cfg_entry.unique_flags1, | ||
5251 | le32_to_cpu(res->cfg_entry.resource_address)); | ||
5252 | } | ||
5253 | } | ||
5254 | |||
5255 | /* Detect any deleted entries, mark them for deletion from mid-layer */ | ||
5256 | list_for_each_entry_safe(res, temp, &old_res, queue) { | ||
5257 | |||
5258 | if (res->scsi_dev) { | ||
5259 | res->change_detected = RES_CHANGE_DEL; | ||
5260 | res->cfg_entry.resource_handle = | ||
5261 | PMCRAID_INVALID_RES_HANDLE; | ||
5262 | list_move_tail(&res->queue, &pinstance->used_res_q); | ||
5263 | } else { | ||
5264 | list_move_tail(&res->queue, &pinstance->free_res_q); | ||
5265 | } | ||
5266 | } | ||
5267 | |||
5268 | /* release the resource list lock */ | ||
5269 | spin_unlock_irqrestore(&pinstance->resource_lock, lock_flags); | ||
5270 | pmcraid_set_supported_devs(cmd); | ||
5271 | } | ||
5272 | |||
5273 | /** | ||
5274 | * pmcraid_querycfg - Send a Query IOA Config to the adapter. | ||
5275 | * @cmd: pointer pmcraid_cmd struct | ||
5276 | * | ||
5277 | * This function sends a Query IOA Configuration command to the adapter to | ||
5278 | * retrieve the IOA configuration table. | ||
5279 | * | ||
5280 | * Return value: | ||
5281 | * none | ||
5282 | */ | ||
5283 | static void pmcraid_querycfg(struct pmcraid_cmd *cmd) | ||
5284 | { | ||
5285 | struct pmcraid_ioarcb *ioarcb = &cmd->ioa_cb->ioarcb; | ||
5286 | struct pmcraid_ioadl_desc *ioadl = ioarcb->add_data.u.ioadl; | ||
5287 | struct pmcraid_instance *pinstance = cmd->drv_inst; | ||
5288 | int cfg_table_size = cpu_to_be32(sizeof(struct pmcraid_config_table)); | ||
5289 | |||
5290 | ioarcb->request_type = REQ_TYPE_IOACMD; | ||
5291 | ioarcb->resource_handle = cpu_to_le32(PMCRAID_IOA_RES_HANDLE); | ||
5292 | |||
5293 | ioarcb->cdb[0] = PMCRAID_QUERY_IOA_CONFIG; | ||
5294 | |||
5295 | /* firmware requires 4-byte length field, specified in B.E format */ | ||
5296 | memcpy(&(ioarcb->cdb[10]), &cfg_table_size, sizeof(cfg_table_size)); | ||
5297 | |||
5298 | /* Since entire config table can be described by single IOADL, it can | ||
5299 | * be part of IOARCB itself | ||
5300 | */ | ||
5301 | ioarcb->ioadl_bus_addr = cpu_to_le64((cmd->ioa_cb_bus_addr) + | ||
5302 | offsetof(struct pmcraid_ioarcb, | ||
5303 | add_data.u.ioadl[0])); | ||
5304 | ioarcb->ioadl_length = cpu_to_le32(sizeof(struct pmcraid_ioadl_desc)); | ||
5305 | ioarcb->ioarcb_bus_addr &= ~(0x1FULL); | ||
5306 | |||
5307 | ioarcb->request_flags0 |= NO_LINK_DESCS; | ||
5308 | ioarcb->data_transfer_length = | ||
5309 | cpu_to_le32(sizeof(struct pmcraid_config_table)); | ||
5310 | |||
5311 | ioadl = &(ioarcb->add_data.u.ioadl[0]); | ||
5312 | ioadl->flags = cpu_to_le32(IOADL_FLAGS_LAST_DESC); | ||
5313 | ioadl->address = cpu_to_le64(pinstance->cfg_table_bus_addr); | ||
5314 | ioadl->data_len = cpu_to_le32(sizeof(struct pmcraid_config_table)); | ||
5315 | |||
5316 | pmcraid_send_cmd(cmd, pmcraid_init_res_table, | ||
5317 | PMCRAID_INTERNAL_TIMEOUT, pmcraid_timeout_handler); | ||
5318 | } | ||
5319 | |||
5320 | |||
5321 | /** | ||
5322 | * pmcraid_probe - PCI probe entry pointer for PMC MaxRaid controller driver | ||
5323 | * @pdev: pointer to pci device structure | ||
5324 | * @dev_id: pointer to device ids structure | ||
5325 | * | ||
5326 | * Return Value | ||
5327 | * returns 0 if the device is claimed and successfully configured. | ||
5328 | * returns non-zero error code in case of any failure | ||
5329 | */ | ||
5330 | static int __devinit pmcraid_probe( | ||
5331 | struct pci_dev *pdev, | ||
5332 | const struct pci_device_id *dev_id | ||
5333 | ) | ||
5334 | { | ||
5335 | struct pmcraid_instance *pinstance; | ||
5336 | struct Scsi_Host *host; | ||
5337 | void __iomem *mapped_pci_addr; | ||
5338 | int rc = PCIBIOS_SUCCESSFUL; | ||
5339 | |||
5340 | if (atomic_read(&pmcraid_adapter_count) >= PMCRAID_MAX_ADAPTERS) { | ||
5341 | pmcraid_err | ||
5342 | ("maximum number(%d) of supported adapters reached\n", | ||
5343 | atomic_read(&pmcraid_adapter_count)); | ||
5344 | return -ENOMEM; | ||
5345 | } | ||
5346 | |||
5347 | atomic_inc(&pmcraid_adapter_count); | ||
5348 | rc = pci_enable_device(pdev); | ||
5349 | |||
5350 | if (rc) { | ||
5351 | dev_err(&pdev->dev, "Cannot enable adapter\n"); | ||
5352 | atomic_dec(&pmcraid_adapter_count); | ||
5353 | return rc; | ||
5354 | } | ||
5355 | |||
5356 | dev_info(&pdev->dev, | ||
5357 | "Found new IOA(%x:%x), Total IOA count: %d\n", | ||
5358 | pdev->vendor, pdev->device, | ||
5359 | atomic_read(&pmcraid_adapter_count)); | ||
5360 | |||
5361 | rc = pci_request_regions(pdev, PMCRAID_DRIVER_NAME); | ||
5362 | |||
5363 | if (rc < 0) { | ||
5364 | dev_err(&pdev->dev, | ||
5365 | "Couldn't register memory range of registers\n"); | ||
5366 | goto out_disable_device; | ||
5367 | } | ||
5368 | |||
5369 | mapped_pci_addr = pci_iomap(pdev, 0, 0); | ||
5370 | |||
5371 | if (!mapped_pci_addr) { | ||
5372 | dev_err(&pdev->dev, "Couldn't map PCI registers memory\n"); | ||
5373 | rc = -ENOMEM; | ||
5374 | goto out_release_regions; | ||
5375 | } | ||
5376 | |||
5377 | pci_set_master(pdev); | ||
5378 | |||
5379 | /* Firmware requires the system bus address of IOARCB to be within | ||
5380 | * 32-bit addressable range though it has 64-bit IOARRIN register. | ||
5381 | * However, firmware supports 64-bit streaming DMA buffers, whereas | ||
5382 | * coherent buffers are to be 32-bit. Since pci_alloc_consistent always | ||
5383 | * returns memory within 4GB (if not, change this logic), coherent | ||
5384 | * buffers are within firmware acceptible address ranges. | ||
5385 | */ | ||
5386 | if ((sizeof(dma_addr_t) == 4) || | ||
5387 | pci_set_dma_mask(pdev, DMA_BIT_MASK(64))) | ||
5388 | rc = pci_set_dma_mask(pdev, DMA_BIT_MASK(32)); | ||
5389 | |||
5390 | /* firmware expects 32-bit DMA addresses for IOARRIN register; set 32 | ||
5391 | * bit mask for pci_alloc_consistent to return addresses within 4GB | ||
5392 | */ | ||
5393 | if (rc == 0) | ||
5394 | rc = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32)); | ||
5395 | |||
5396 | if (rc != 0) { | ||
5397 | dev_err(&pdev->dev, "Failed to set PCI DMA mask\n"); | ||
5398 | goto cleanup_nomem; | ||
5399 | } | ||
5400 | |||
5401 | host = scsi_host_alloc(&pmcraid_host_template, | ||
5402 | sizeof(struct pmcraid_instance)); | ||
5403 | |||
5404 | if (!host) { | ||
5405 | dev_err(&pdev->dev, "scsi_host_alloc failed!\n"); | ||
5406 | rc = -ENOMEM; | ||
5407 | goto cleanup_nomem; | ||
5408 | } | ||
5409 | |||
5410 | host->max_id = PMCRAID_MAX_NUM_TARGETS_PER_BUS; | ||
5411 | host->max_lun = PMCRAID_MAX_NUM_LUNS_PER_TARGET; | ||
5412 | host->unique_id = host->host_no; | ||
5413 | host->max_channel = PMCRAID_MAX_BUS_TO_SCAN; | ||
5414 | host->max_cmd_len = PMCRAID_MAX_CDB_LEN; | ||
5415 | |||
5416 | /* zero out entire instance structure */ | ||
5417 | pinstance = (struct pmcraid_instance *)host->hostdata; | ||
5418 | memset(pinstance, 0, sizeof(*pinstance)); | ||
5419 | |||
5420 | pinstance->chip_cfg = | ||
5421 | (struct pmcraid_chip_details *)(dev_id->driver_data); | ||
5422 | |||
5423 | rc = pmcraid_init_instance(pdev, host, mapped_pci_addr); | ||
5424 | |||
5425 | if (rc < 0) { | ||
5426 | dev_err(&pdev->dev, "failed to initialize adapter instance\n"); | ||
5427 | goto out_scsi_host_put; | ||
5428 | } | ||
5429 | |||
5430 | pci_set_drvdata(pdev, pinstance); | ||
5431 | |||
5432 | /* Save PCI config-space for use following the reset */ | ||
5433 | rc = pci_save_state(pinstance->pdev); | ||
5434 | |||
5435 | if (rc != 0) { | ||
5436 | dev_err(&pdev->dev, "Failed to save PCI config space\n"); | ||
5437 | goto out_scsi_host_put; | ||
5438 | } | ||
5439 | |||
5440 | pmcraid_disable_interrupts(pinstance, ~0); | ||
5441 | |||
5442 | rc = pmcraid_register_interrupt_handler(pinstance); | ||
5443 | |||
5444 | if (rc) { | ||
5445 | pmcraid_err("couldn't register interrupt handler\n"); | ||
5446 | goto out_scsi_host_put; | ||
5447 | } | ||
5448 | |||
5449 | pmcraid_init_tasklets(pinstance); | ||
5450 | |||
5451 | /* allocate verious buffers used by LLD.*/ | ||
5452 | rc = pmcraid_init_buffers(pinstance); | ||
5453 | |||
5454 | if (rc) { | ||
5455 | pmcraid_err("couldn't allocate memory blocks\n"); | ||
5456 | goto out_unregister_isr; | ||
5457 | } | ||
5458 | |||
5459 | /* check the reset type required */ | ||
5460 | pmcraid_reset_type(pinstance); | ||
5461 | |||
5462 | pmcraid_enable_interrupts(pinstance, PMCRAID_PCI_INTERRUPTS); | ||
5463 | |||
5464 | /* Start IOA firmware initialization and bring card to Operational | ||
5465 | * state. | ||
5466 | */ | ||
5467 | pmcraid_info("starting IOA initialization sequence\n"); | ||
5468 | if (pmcraid_reset_bringup(pinstance)) { | ||
5469 | pmcraid_err("couldn't initialize IOA \n"); | ||
5470 | rc = 1; | ||
5471 | goto out_release_bufs; | ||
5472 | } | ||
5473 | |||
5474 | /* Add adapter instance into mid-layer list */ | ||
5475 | rc = scsi_add_host(pinstance->host, &pdev->dev); | ||
5476 | if (rc != 0) { | ||
5477 | pmcraid_err("couldn't add host into mid-layer: %d\n", rc); | ||
5478 | goto out_release_bufs; | ||
5479 | } | ||
5480 | |||
5481 | scsi_scan_host(pinstance->host); | ||
5482 | |||
5483 | rc = pmcraid_setup_chrdev(pinstance); | ||
5484 | |||
5485 | if (rc != 0) { | ||
5486 | pmcraid_err("couldn't create mgmt interface, error: %x\n", | ||
5487 | rc); | ||
5488 | goto out_remove_host; | ||
5489 | } | ||
5490 | |||
5491 | /* Schedule worker thread to handle CCN and take care of adding and | ||
5492 | * removing devices to OS | ||
5493 | */ | ||
5494 | atomic_set(&pinstance->expose_resources, 1); | ||
5495 | schedule_work(&pinstance->worker_q); | ||
5496 | return rc; | ||
5497 | |||
5498 | out_remove_host: | ||
5499 | scsi_remove_host(host); | ||
5500 | |||
5501 | out_release_bufs: | ||
5502 | pmcraid_release_buffers(pinstance); | ||
5503 | |||
5504 | out_unregister_isr: | ||
5505 | pmcraid_kill_tasklets(pinstance); | ||
5506 | pmcraid_unregister_interrupt_handler(pinstance); | ||
5507 | |||
5508 | out_scsi_host_put: | ||
5509 | scsi_host_put(host); | ||
5510 | |||
5511 | cleanup_nomem: | ||
5512 | iounmap(mapped_pci_addr); | ||
5513 | |||
5514 | out_release_regions: | ||
5515 | pci_release_regions(pdev); | ||
5516 | |||
5517 | out_disable_device: | ||
5518 | atomic_dec(&pmcraid_adapter_count); | ||
5519 | pci_set_drvdata(pdev, NULL); | ||
5520 | pci_disable_device(pdev); | ||
5521 | return -ENODEV; | ||
5522 | } | ||
5523 | |||
5524 | /* | ||
5525 | * PCI driver structure of pcmraid driver | ||
5526 | */ | ||
5527 | static struct pci_driver pmcraid_driver = { | ||
5528 | .name = PMCRAID_DRIVER_NAME, | ||
5529 | .id_table = pmcraid_pci_table, | ||
5530 | .probe = pmcraid_probe, | ||
5531 | .remove = pmcraid_remove, | ||
5532 | .suspend = pmcraid_suspend, | ||
5533 | .resume = pmcraid_resume, | ||
5534 | .shutdown = pmcraid_shutdown | ||
5535 | }; | ||
5536 | |||
5537 | |||
5538 | /** | ||
5539 | * pmcraid_init - module load entry point | ||
5540 | */ | ||
5541 | static int __init pmcraid_init(void) | ||
5542 | { | ||
5543 | dev_t dev; | ||
5544 | int error; | ||
5545 | |||
5546 | pmcraid_info("%s Device Driver version: %s %s\n", | ||
5547 | PMCRAID_DRIVER_NAME, | ||
5548 | PMCRAID_DRIVER_VERSION, PMCRAID_DRIVER_DATE); | ||
5549 | |||
5550 | error = alloc_chrdev_region(&dev, 0, | ||
5551 | PMCRAID_MAX_ADAPTERS, | ||
5552 | PMCRAID_DEVFILE); | ||
5553 | |||
5554 | if (error) { | ||
5555 | pmcraid_err("failed to get a major number for adapters\n"); | ||
5556 | goto out_init; | ||
5557 | } | ||
5558 | |||
5559 | pmcraid_major = MAJOR(dev); | ||
5560 | pmcraid_class = class_create(THIS_MODULE, PMCRAID_DEVFILE); | ||
5561 | |||
5562 | if (IS_ERR(pmcraid_class)) { | ||
5563 | error = PTR_ERR(pmcraid_class); | ||
5564 | pmcraid_err("failed to register with with sysfs, error = %x\n", | ||
5565 | error); | ||
5566 | goto out_unreg_chrdev; | ||
5567 | } | ||
5568 | |||
5569 | |||
5570 | error = pmcraid_netlink_init(); | ||
5571 | |||
5572 | if (error) | ||
5573 | goto out_unreg_chrdev; | ||
5574 | |||
5575 | error = pci_register_driver(&pmcraid_driver); | ||
5576 | |||
5577 | if (error == 0) | ||
5578 | goto out_init; | ||
5579 | |||
5580 | pmcraid_err("failed to register pmcraid driver, error = %x\n", | ||
5581 | error); | ||
5582 | class_destroy(pmcraid_class); | ||
5583 | pmcraid_netlink_release(); | ||
5584 | |||
5585 | out_unreg_chrdev: | ||
5586 | unregister_chrdev_region(MKDEV(pmcraid_major, 0), PMCRAID_MAX_ADAPTERS); | ||
5587 | out_init: | ||
5588 | return error; | ||
5589 | } | ||
5590 | |||
5591 | /** | ||
5592 | * pmcraid_exit - module unload entry point | ||
5593 | */ | ||
5594 | static void __exit pmcraid_exit(void) | ||
5595 | { | ||
5596 | pmcraid_netlink_release(); | ||
5597 | class_destroy(pmcraid_class); | ||
5598 | unregister_chrdev_region(MKDEV(pmcraid_major, 0), | ||
5599 | PMCRAID_MAX_ADAPTERS); | ||
5600 | pci_unregister_driver(&pmcraid_driver); | ||
5601 | } | ||
5602 | |||
5603 | module_init(pmcraid_init); | ||
5604 | module_exit(pmcraid_exit); | ||
diff --git a/drivers/scsi/pmcraid.h b/drivers/scsi/pmcraid.h new file mode 100644 index 000000000000..614b3a764fed --- /dev/null +++ b/drivers/scsi/pmcraid.h | |||
@@ -0,0 +1,1029 @@ | |||
1 | /* | ||
2 | * pmcraid.h -- PMC Sierra MaxRAID controller driver header file | ||
3 | * | ||
4 | * Copyright (C) 2008, 2009 PMC Sierra Inc. | ||
5 | * | ||
6 | * This program is free software; you can redistribute it and/or modify | ||
7 | * it under the terms of the GNU General Public License as published by | ||
8 | * the Free Software Foundation; either version 2 of the License, or | ||
9 | * (at your option) any later version. | ||
10 | * | ||
11 | * This program is distributed in the hope that it will be useful, | ||
12 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
13 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
14 | * GNU General Public License for more details. | ||
15 | * | ||
16 | * You should have received a copy of the GNU General Public License | ||
17 | * along with this program; if not, write to the Free Software | ||
18 | * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA | ||
19 | */ | ||
20 | |||
21 | #ifndef _PMCRAID_H | ||
22 | #define _PMCRAID_H | ||
23 | |||
24 | #include <linux/version.h> | ||
25 | #include <linux/types.h> | ||
26 | #include <linux/completion.h> | ||
27 | #include <linux/list.h> | ||
28 | #include <scsi/scsi.h> | ||
29 | #include <linux/kref.h> | ||
30 | #include <scsi/scsi_cmnd.h> | ||
31 | #include <linux/cdev.h> | ||
32 | #include <net/netlink.h> | ||
33 | #include <net/genetlink.h> | ||
34 | #include <linux/connector.h> | ||
35 | /* | ||
36 | * Driver name : string representing the driver name | ||
37 | * Device file : /dev file to be used for management interfaces | ||
38 | * Driver version: version string in major_version.minor_version.patch format | ||
39 | * Driver date : date information in "Mon dd yyyy" format | ||
40 | */ | ||
41 | #define PMCRAID_DRIVER_NAME "PMC MaxRAID" | ||
42 | #define PMCRAID_DEVFILE "pmcsas" | ||
43 | #define PMCRAID_DRIVER_VERSION "1.0.2" | ||
44 | #define PMCRAID_DRIVER_DATE __DATE__ | ||
45 | |||
46 | /* Maximum number of adapters supported by current version of the driver */ | ||
47 | #define PMCRAID_MAX_ADAPTERS 1024 | ||
48 | |||
49 | /* Bit definitions as per firmware, bit position [0][1][2].....[31] */ | ||
50 | #define PMC_BIT8(n) (1 << (7-n)) | ||
51 | #define PMC_BIT16(n) (1 << (15-n)) | ||
52 | #define PMC_BIT32(n) (1 << (31-n)) | ||
53 | |||
54 | /* PMC PCI vendor ID and device ID values */ | ||
55 | #define PCI_VENDOR_ID_PMC 0x11F8 | ||
56 | #define PCI_DEVICE_ID_PMC_MAXRAID 0x5220 | ||
57 | |||
58 | /* | ||
59 | * MAX_CMD : maximum commands that can be outstanding with IOA | ||
60 | * MAX_IO_CMD : command blocks available for IO commands | ||
61 | * MAX_HCAM_CMD : command blocks avaibale for HCAMS | ||
62 | * MAX_INTERNAL_CMD : command blocks avaible for internal commands like reset | ||
63 | */ | ||
64 | #define PMCRAID_MAX_CMD 1024 | ||
65 | #define PMCRAID_MAX_IO_CMD 1020 | ||
66 | #define PMCRAID_MAX_HCAM_CMD 2 | ||
67 | #define PMCRAID_MAX_INTERNAL_CMD 2 | ||
68 | |||
69 | /* MAX_IOADLS : max number of scatter-gather lists supported by IOA | ||
70 | * IOADLS_INTERNAL : number of ioadls included as part of IOARCB. | ||
71 | * IOADLS_EXTERNAL : number of ioadls allocated external to IOARCB | ||
72 | */ | ||
73 | #define PMCRAID_IOADLS_INTERNAL 27 | ||
74 | #define PMCRAID_IOADLS_EXTERNAL 37 | ||
75 | #define PMCRAID_MAX_IOADLS PMCRAID_IOADLS_INTERNAL | ||
76 | |||
77 | /* HRRQ_ENTRY_SIZE : size of hrrq buffer | ||
78 | * IOARCB_ALIGNMENT : alignment required for IOARCB | ||
79 | * IOADL_ALIGNMENT : alignment requirement for IOADLs | ||
80 | * MSIX_VECTORS : number of MSIX vectors supported | ||
81 | */ | ||
82 | #define HRRQ_ENTRY_SIZE sizeof(__le32) | ||
83 | #define PMCRAID_IOARCB_ALIGNMENT 32 | ||
84 | #define PMCRAID_IOADL_ALIGNMENT 16 | ||
85 | #define PMCRAID_IOASA_ALIGNMENT 4 | ||
86 | #define PMCRAID_NUM_MSIX_VECTORS 1 | ||
87 | |||
88 | /* various other limits */ | ||
89 | #define PMCRAID_VENDOR_ID_LEN 8 | ||
90 | #define PMCRAID_PRODUCT_ID_LEN 16 | ||
91 | #define PMCRAID_SERIAL_NUM_LEN 8 | ||
92 | #define PMCRAID_LUN_LEN 8 | ||
93 | #define PMCRAID_MAX_CDB_LEN 16 | ||
94 | #define PMCRAID_DEVICE_ID_LEN 8 | ||
95 | #define PMCRAID_SENSE_DATA_LEN 256 | ||
96 | #define PMCRAID_ADD_CMD_PARAM_LEN 48 | ||
97 | |||
98 | #define PMCRAID_MAX_BUS_TO_SCAN 1 | ||
99 | #define PMCRAID_MAX_NUM_TARGETS_PER_BUS 256 | ||
100 | #define PMCRAID_MAX_NUM_LUNS_PER_TARGET 8 | ||
101 | |||
102 | /* IOA bus/target/lun number of IOA resources */ | ||
103 | #define PMCRAID_IOA_BUS_ID 0xfe | ||
104 | #define PMCRAID_IOA_TARGET_ID 0xff | ||
105 | #define PMCRAID_IOA_LUN_ID 0xff | ||
106 | #define PMCRAID_VSET_BUS_ID 0x1 | ||
107 | #define PMCRAID_VSET_LUN_ID 0x0 | ||
108 | #define PMCRAID_PHYS_BUS_ID 0x0 | ||
109 | #define PMCRAID_VIRTUAL_ENCL_BUS_ID 0x8 | ||
110 | #define PMCRAID_MAX_VSET_TARGETS 240 | ||
111 | #define PMCRAID_MAX_VSET_LUNS_PER_TARGET 8 | ||
112 | |||
113 | #define PMCRAID_IOA_MAX_SECTORS 32767 | ||
114 | #define PMCRAID_VSET_MAX_SECTORS 512 | ||
115 | #define PMCRAID_MAX_CMD_PER_LUN 254 | ||
116 | |||
117 | /* Number of configuration table entries (resources) */ | ||
118 | #define PMCRAID_MAX_NUM_OF_VSETS 240 | ||
119 | |||
120 | /* Todo : Check max limit for Phase 1 */ | ||
121 | #define PMCRAID_MAX_NUM_OF_PHY_DEVS 256 | ||
122 | |||
123 | /* MAX_NUM_OF_DEVS includes 1 FP, 1 Dummy Enclosure device */ | ||
124 | #define PMCRAID_MAX_NUM_OF_DEVS \ | ||
125 | (PMCRAID_MAX_NUM_OF_VSETS + PMCRAID_MAX_NUM_OF_PHY_DEVS + 2) | ||
126 | |||
127 | #define PMCRAID_MAX_RESOURCES PMCRAID_MAX_NUM_OF_DEVS | ||
128 | |||
129 | /* Adapter Commands used by driver */ | ||
130 | #define PMCRAID_QUERY_RESOURCE_STATE 0xC2 | ||
131 | #define PMCRAID_RESET_DEVICE 0xC3 | ||
132 | /* options to select reset target */ | ||
133 | #define ENABLE_RESET_MODIFIER 0x80 | ||
134 | #define RESET_DEVICE_LUN 0x40 | ||
135 | #define RESET_DEVICE_TARGET 0x20 | ||
136 | #define RESET_DEVICE_BUS 0x10 | ||
137 | |||
138 | #define PMCRAID_IDENTIFY_HRRQ 0xC4 | ||
139 | #define PMCRAID_QUERY_IOA_CONFIG 0xC5 | ||
140 | #define PMCRAID_QUERY_CMD_STATUS 0xCB | ||
141 | #define PMCRAID_ABORT_CMD 0xC7 | ||
142 | |||
143 | /* CANCEL ALL command, provides option for setting SYNC_COMPLETE | ||
144 | * on the target resources for which commands got cancelled | ||
145 | */ | ||
146 | #define PMCRAID_CANCEL_ALL_REQUESTS 0xCE | ||
147 | #define PMCRAID_SYNC_COMPLETE_AFTER_CANCEL PMC_BIT8(0) | ||
148 | |||
149 | /* HCAM command and types of HCAM supported by IOA */ | ||
150 | #define PMCRAID_HOST_CONTROLLED_ASYNC 0xCF | ||
151 | #define PMCRAID_HCAM_CODE_CONFIG_CHANGE 0x01 | ||
152 | #define PMCRAID_HCAM_CODE_LOG_DATA 0x02 | ||
153 | |||
154 | /* IOA shutdown command and various shutdown types */ | ||
155 | #define PMCRAID_IOA_SHUTDOWN 0xF7 | ||
156 | #define PMCRAID_SHUTDOWN_NORMAL 0x00 | ||
157 | #define PMCRAID_SHUTDOWN_PREPARE_FOR_NORMAL 0x40 | ||
158 | #define PMCRAID_SHUTDOWN_NONE 0x100 | ||
159 | #define PMCRAID_SHUTDOWN_ABBREV 0x80 | ||
160 | |||
161 | /* SET SUPPORTED DEVICES command and the option to select all the | ||
162 | * devices to be supported | ||
163 | */ | ||
164 | #define PMCRAID_SET_SUPPORTED_DEVICES 0xFB | ||
165 | #define ALL_DEVICES_SUPPORTED PMC_BIT8(0) | ||
166 | |||
167 | /* This option is used with SCSI WRITE_BUFFER command */ | ||
168 | #define PMCRAID_WR_BUF_DOWNLOAD_AND_SAVE 0x05 | ||
169 | |||
170 | /* IOASC Codes used by driver */ | ||
171 | #define PMCRAID_IOASC_SENSE_MASK 0xFFFFFF00 | ||
172 | #define PMCRAID_IOASC_SENSE_KEY(ioasc) ((ioasc) >> 24) | ||
173 | #define PMCRAID_IOASC_SENSE_CODE(ioasc) (((ioasc) & 0x00ff0000) >> 16) | ||
174 | #define PMCRAID_IOASC_SENSE_QUAL(ioasc) (((ioasc) & 0x0000ff00) >> 8) | ||
175 | #define PMCRAID_IOASC_SENSE_STATUS(ioasc) ((ioasc) & 0x000000ff) | ||
176 | |||
177 | #define PMCRAID_IOASC_GOOD_COMPLETION 0x00000000 | ||
178 | #define PMCRAID_IOASC_NR_INIT_CMD_REQUIRED 0x02040200 | ||
179 | #define PMCRAID_IOASC_NR_IOA_RESET_REQUIRED 0x02048000 | ||
180 | #define PMCRAID_IOASC_NR_SYNC_REQUIRED 0x023F0000 | ||
181 | #define PMCRAID_IOASC_ME_READ_ERROR_NO_REALLOC 0x03110C00 | ||
182 | #define PMCRAID_IOASC_HW_CANNOT_COMMUNICATE 0x04050000 | ||
183 | #define PMCRAID_IOASC_HW_DEVICE_TIMEOUT 0x04080100 | ||
184 | #define PMCRAID_IOASC_HW_DEVICE_BUS_STATUS_ERROR 0x04448500 | ||
185 | #define PMCRAID_IOASC_HW_IOA_RESET_REQUIRED 0x04448600 | ||
186 | #define PMCRAID_IOASC_IR_INVALID_RESOURCE_HANDLE 0x05250000 | ||
187 | #define PMCRAID_IOASC_AC_TERMINATED_BY_HOST 0x0B5A0000 | ||
188 | #define PMCRAID_IOASC_UA_BUS_WAS_RESET 0x06290000 | ||
189 | #define PMCRAID_IOASC_UA_BUS_WAS_RESET_BY_OTHER 0x06298000 | ||
190 | |||
191 | /* Driver defined IOASCs */ | ||
192 | #define PMCRAID_IOASC_IOA_WAS_RESET 0x10000001 | ||
193 | #define PMCRAID_IOASC_PCI_ACCESS_ERROR 0x10000002 | ||
194 | |||
195 | /* Various timeout values (in milliseconds) used. If any of these are chip | ||
196 | * specific, move them to pmcraid_chip_details structure. | ||
197 | */ | ||
198 | #define PMCRAID_PCI_DEASSERT_TIMEOUT 2000 | ||
199 | #define PMCRAID_BIST_TIMEOUT 2000 | ||
200 | #define PMCRAID_AENWAIT_TIMEOUT 5000 | ||
201 | #define PMCRAID_TRANSOP_TIMEOUT 60000 | ||
202 | |||
203 | #define PMCRAID_RESET_TIMEOUT (2 * HZ) | ||
204 | #define PMCRAID_CHECK_FOR_RESET_TIMEOUT ((HZ / 10)) | ||
205 | #define PMCRAID_VSET_IO_TIMEOUT (60 * HZ) | ||
206 | #define PMCRAID_INTERNAL_TIMEOUT (60 * HZ) | ||
207 | #define PMCRAID_SHUTDOWN_TIMEOUT (150 * HZ) | ||
208 | #define PMCRAID_RESET_BUS_TIMEOUT (60 * HZ) | ||
209 | #define PMCRAID_RESET_HOST_TIMEOUT (150 * HZ) | ||
210 | #define PMCRAID_REQUEST_SENSE_TIMEOUT (30 * HZ) | ||
211 | #define PMCRAID_SET_SUP_DEV_TIMEOUT (2 * 60 * HZ) | ||
212 | |||
213 | /* structure to represent a scatter-gather element (IOADL descriptor) */ | ||
214 | struct pmcraid_ioadl_desc { | ||
215 | __le64 address; | ||
216 | __le32 data_len; | ||
217 | __u8 reserved[3]; | ||
218 | __u8 flags; | ||
219 | } __attribute__((packed, aligned(PMCRAID_IOADL_ALIGNMENT))); | ||
220 | |||
221 | /* pmcraid_ioadl_desc.flags values */ | ||
222 | #define IOADL_FLAGS_CHAINED PMC_BIT8(0) | ||
223 | #define IOADL_FLAGS_LAST_DESC PMC_BIT8(1) | ||
224 | #define IOADL_FLAGS_READ_LAST PMC_BIT8(1) | ||
225 | #define IOADL_FLAGS_WRITE_LAST PMC_BIT8(1) | ||
226 | |||
227 | |||
228 | /* additional IOARCB data which can be CDB or additional request parameters | ||
229 | * or list of IOADLs. Firmware supports max of 512 bytes for IOARCB, hence then | ||
230 | * number of IOADLs are limted to 27. In case they are more than 27, they will | ||
231 | * be used in chained form | ||
232 | */ | ||
233 | struct pmcraid_ioarcb_add_data { | ||
234 | union { | ||
235 | struct pmcraid_ioadl_desc ioadl[PMCRAID_IOADLS_INTERNAL]; | ||
236 | __u8 add_cmd_params[PMCRAID_ADD_CMD_PARAM_LEN]; | ||
237 | } u; | ||
238 | }; | ||
239 | |||
240 | /* | ||
241 | * IOA Request Control Block | ||
242 | */ | ||
243 | struct pmcraid_ioarcb { | ||
244 | __le64 ioarcb_bus_addr; | ||
245 | __le32 resource_handle; | ||
246 | __le32 response_handle; | ||
247 | __le64 ioadl_bus_addr; | ||
248 | __le32 ioadl_length; | ||
249 | __le32 data_transfer_length; | ||
250 | __le64 ioasa_bus_addr; | ||
251 | __le16 ioasa_len; | ||
252 | __le16 cmd_timeout; | ||
253 | __le16 add_cmd_param_offset; | ||
254 | __le16 add_cmd_param_length; | ||
255 | __le32 reserved1[2]; | ||
256 | __le32 reserved2; | ||
257 | __u8 request_type; | ||
258 | __u8 request_flags0; | ||
259 | __u8 request_flags1; | ||
260 | __u8 hrrq_id; | ||
261 | __u8 cdb[PMCRAID_MAX_CDB_LEN]; | ||
262 | struct pmcraid_ioarcb_add_data add_data; | ||
263 | } __attribute__((packed, aligned(PMCRAID_IOARCB_ALIGNMENT))); | ||
264 | |||
265 | /* well known resource handle values */ | ||
266 | #define PMCRAID_IOA_RES_HANDLE 0xffffffff | ||
267 | #define PMCRAID_INVALID_RES_HANDLE 0 | ||
268 | |||
269 | /* pmcraid_ioarcb.request_type values */ | ||
270 | #define REQ_TYPE_SCSI 0x00 | ||
271 | #define REQ_TYPE_IOACMD 0x01 | ||
272 | #define REQ_TYPE_HCAM 0x02 | ||
273 | |||
274 | /* pmcraid_ioarcb.flags0 values */ | ||
275 | #define TRANSFER_DIR_WRITE PMC_BIT8(0) | ||
276 | #define INHIBIT_UL_CHECK PMC_BIT8(2) | ||
277 | #define SYNC_OVERRIDE PMC_BIT8(3) | ||
278 | #define SYNC_COMPLETE PMC_BIT8(4) | ||
279 | #define NO_LINK_DESCS PMC_BIT8(5) | ||
280 | |||
281 | /* pmcraid_ioarcb.flags1 values */ | ||
282 | #define DELAY_AFTER_RESET PMC_BIT8(0) | ||
283 | #define TASK_TAG_SIMPLE 0x10 | ||
284 | #define TASK_TAG_ORDERED 0x20 | ||
285 | #define TASK_TAG_QUEUE_HEAD 0x30 | ||
286 | |||
287 | /* toggle bit offset in response handle */ | ||
288 | #define HRRQ_TOGGLE_BIT 0x01 | ||
289 | #define HRRQ_RESPONSE_BIT 0x02 | ||
290 | |||
291 | /* IOA Status Area */ | ||
292 | struct pmcraid_ioasa_vset { | ||
293 | __le32 failing_lba_hi; | ||
294 | __le32 failing_lba_lo; | ||
295 | __le32 reserved; | ||
296 | } __attribute__((packed, aligned(4))); | ||
297 | |||
298 | struct pmcraid_ioasa { | ||
299 | __le32 ioasc; | ||
300 | __le16 returned_status_length; | ||
301 | __le16 available_status_length; | ||
302 | __le32 residual_data_length; | ||
303 | __le32 ilid; | ||
304 | __le32 fd_ioasc; | ||
305 | __le32 fd_res_address; | ||
306 | __le32 fd_res_handle; | ||
307 | __le32 reserved; | ||
308 | |||
309 | /* resource specific sense information */ | ||
310 | union { | ||
311 | struct pmcraid_ioasa_vset vset; | ||
312 | } u; | ||
313 | |||
314 | /* IOA autosense data */ | ||
315 | __le16 auto_sense_length; | ||
316 | __le16 error_data_length; | ||
317 | __u8 sense_data[PMCRAID_SENSE_DATA_LEN]; | ||
318 | } __attribute__((packed, aligned(4))); | ||
319 | |||
320 | #define PMCRAID_DRIVER_ILID 0xffffffff | ||
321 | |||
322 | /* Config Table Entry per Resource */ | ||
323 | struct pmcraid_config_table_entry { | ||
324 | __u8 resource_type; | ||
325 | __u8 bus_protocol; | ||
326 | __le16 array_id; | ||
327 | __u8 common_flags0; | ||
328 | __u8 common_flags1; | ||
329 | __u8 unique_flags0; | ||
330 | __u8 unique_flags1; /*also used as vset target_id */ | ||
331 | __le32 resource_handle; | ||
332 | __le32 resource_address; | ||
333 | __u8 device_id[PMCRAID_DEVICE_ID_LEN]; | ||
334 | __u8 lun[PMCRAID_LUN_LEN]; | ||
335 | } __attribute__((packed, aligned(4))); | ||
336 | |||
337 | /* resource types (config_table_entry.resource_type values) */ | ||
338 | #define RES_TYPE_AF_DASD 0x00 | ||
339 | #define RES_TYPE_GSCSI 0x01 | ||
340 | #define RES_TYPE_VSET 0x02 | ||
341 | #define RES_TYPE_IOA_FP 0xFF | ||
342 | |||
343 | #define RES_IS_IOA(res) ((res).resource_type == RES_TYPE_IOA_FP) | ||
344 | #define RES_IS_GSCSI(res) ((res).resource_type == RES_TYPE_GSCSI) | ||
345 | #define RES_IS_VSET(res) ((res).resource_type == RES_TYPE_VSET) | ||
346 | #define RES_IS_AFDASD(res) ((res).resource_type == RES_TYPE_AF_DASD) | ||
347 | |||
348 | /* bus_protocol values used by driver */ | ||
349 | #define RES_TYPE_VENCLOSURE 0x8 | ||
350 | |||
351 | /* config_table_entry.common_flags0 */ | ||
352 | #define MULTIPATH_RESOURCE PMC_BIT32(0) | ||
353 | |||
354 | /* unique_flags1 */ | ||
355 | #define IMPORT_MODE_MANUAL PMC_BIT8(0) | ||
356 | |||
357 | /* well known resource handle values */ | ||
358 | #define RES_HANDLE_IOA 0xFFFFFFFF | ||
359 | #define RES_HANDLE_NONE 0x00000000 | ||
360 | |||
361 | /* well known resource address values */ | ||
362 | #define RES_ADDRESS_IOAFP 0xFEFFFFFF | ||
363 | #define RES_ADDRESS_INVALID 0xFFFFFFFF | ||
364 | |||
365 | /* BUS/TARGET/LUN values from resource_addrr */ | ||
366 | #define RES_BUS(res_addr) (le32_to_cpu(res_addr) & 0xFF) | ||
367 | #define RES_TARGET(res_addr) ((le32_to_cpu(res_addr) >> 16) & 0xFF) | ||
368 | #define RES_LUN(res_addr) 0x0 | ||
369 | |||
370 | /* configuration table structure */ | ||
371 | struct pmcraid_config_table { | ||
372 | __le16 num_entries; | ||
373 | __u8 table_format; | ||
374 | __u8 reserved1; | ||
375 | __u8 flags; | ||
376 | __u8 reserved2[11]; | ||
377 | struct pmcraid_config_table_entry entries[PMCRAID_MAX_RESOURCES]; | ||
378 | } __attribute__((packed, aligned(4))); | ||
379 | |||
380 | /* config_table.flags value */ | ||
381 | #define MICROCODE_UPDATE_REQUIRED PMC_BIT32(0) | ||
382 | |||
383 | /* | ||
384 | * HCAM format | ||
385 | */ | ||
386 | #define PMCRAID_HOSTRCB_LDNSIZE 4056 | ||
387 | |||
388 | /* Error log notification format */ | ||
389 | struct pmcraid_hostrcb_error { | ||
390 | __le32 fd_ioasc; | ||
391 | __le32 fd_ra; | ||
392 | __le32 fd_rh; | ||
393 | __le32 prc; | ||
394 | union { | ||
395 | __u8 data[PMCRAID_HOSTRCB_LDNSIZE]; | ||
396 | } u; | ||
397 | } __attribute__ ((packed, aligned(4))); | ||
398 | |||
399 | struct pmcraid_hcam_hdr { | ||
400 | __u8 op_code; | ||
401 | __u8 notification_type; | ||
402 | __u8 notification_lost; | ||
403 | __u8 flags; | ||
404 | __u8 overlay_id; | ||
405 | __u8 reserved1[3]; | ||
406 | __le32 ilid; | ||
407 | __le32 timestamp1; | ||
408 | __le32 timestamp2; | ||
409 | __le32 data_len; | ||
410 | } __attribute__((packed, aligned(4))); | ||
411 | |||
412 | #define PMCRAID_AEN_GROUP 0x3 | ||
413 | |||
414 | struct pmcraid_hcam_ccn { | ||
415 | struct pmcraid_hcam_hdr header; | ||
416 | struct pmcraid_config_table_entry cfg_entry; | ||
417 | } __attribute__((packed, aligned(4))); | ||
418 | |||
419 | struct pmcraid_hcam_ldn { | ||
420 | struct pmcraid_hcam_hdr header; | ||
421 | struct pmcraid_hostrcb_error error_log; | ||
422 | } __attribute__((packed, aligned(4))); | ||
423 | |||
424 | /* pmcraid_hcam.op_code values */ | ||
425 | #define HOSTRCB_TYPE_CCN 0xE1 | ||
426 | #define HOSTRCB_TYPE_LDN 0xE2 | ||
427 | |||
428 | /* pmcraid_hcam.notification_type values */ | ||
429 | #define NOTIFICATION_TYPE_ENTRY_CHANGED 0x0 | ||
430 | #define NOTIFICATION_TYPE_ENTRY_NEW 0x1 | ||
431 | #define NOTIFICATION_TYPE_ENTRY_DELETED 0x2 | ||
432 | #define NOTIFICATION_TYPE_ERROR_LOG 0x10 | ||
433 | #define NOTIFICATION_TYPE_INFORMATION_LOG 0x11 | ||
434 | |||
435 | #define HOSTRCB_NOTIFICATIONS_LOST PMC_BIT8(0) | ||
436 | |||
437 | /* pmcraid_hcam.flags values */ | ||
438 | #define HOSTRCB_INTERNAL_OP_ERROR PMC_BIT8(0) | ||
439 | #define HOSTRCB_ERROR_RESPONSE_SENT PMC_BIT8(1) | ||
440 | |||
441 | /* pmcraid_hcam.overlay_id values */ | ||
442 | #define HOSTRCB_OVERLAY_ID_08 0x08 | ||
443 | #define HOSTRCB_OVERLAY_ID_09 0x09 | ||
444 | #define HOSTRCB_OVERLAY_ID_11 0x11 | ||
445 | #define HOSTRCB_OVERLAY_ID_12 0x12 | ||
446 | #define HOSTRCB_OVERLAY_ID_13 0x13 | ||
447 | #define HOSTRCB_OVERLAY_ID_14 0x14 | ||
448 | #define HOSTRCB_OVERLAY_ID_16 0x16 | ||
449 | #define HOSTRCB_OVERLAY_ID_17 0x17 | ||
450 | #define HOSTRCB_OVERLAY_ID_20 0x20 | ||
451 | #define HOSTRCB_OVERLAY_ID_FF 0xFF | ||
452 | |||
453 | /* Implementation specific card details */ | ||
454 | struct pmcraid_chip_details { | ||
455 | /* hardware register offsets */ | ||
456 | unsigned long ioastatus; | ||
457 | unsigned long ioarrin; | ||
458 | unsigned long mailbox; | ||
459 | unsigned long global_intr_mask; | ||
460 | unsigned long ioa_host_intr; | ||
461 | unsigned long ioa_host_intr_clr; | ||
462 | unsigned long ioa_host_mask; | ||
463 | unsigned long ioa_host_mask_clr; | ||
464 | unsigned long host_ioa_intr; | ||
465 | unsigned long host_ioa_intr_clr; | ||
466 | |||
467 | /* timeout used during transitional to operational state */ | ||
468 | unsigned long transop_timeout; | ||
469 | }; | ||
470 | |||
471 | /* IOA to HOST doorbells (interrupts) */ | ||
472 | #define INTRS_TRANSITION_TO_OPERATIONAL PMC_BIT32(0) | ||
473 | #define INTRS_IOARCB_TRANSFER_FAILED PMC_BIT32(3) | ||
474 | #define INTRS_IOA_UNIT_CHECK PMC_BIT32(4) | ||
475 | #define INTRS_NO_HRRQ_FOR_CMD_RESPONSE PMC_BIT32(5) | ||
476 | #define INTRS_CRITICAL_OP_IN_PROGRESS PMC_BIT32(6) | ||
477 | #define INTRS_IO_DEBUG_ACK PMC_BIT32(7) | ||
478 | #define INTRS_IOARRIN_LOST PMC_BIT32(27) | ||
479 | #define INTRS_SYSTEM_BUS_MMIO_ERROR PMC_BIT32(28) | ||
480 | #define INTRS_IOA_PROCESSOR_ERROR PMC_BIT32(29) | ||
481 | #define INTRS_HRRQ_VALID PMC_BIT32(30) | ||
482 | #define INTRS_OPERATIONAL_STATUS PMC_BIT32(0) | ||
483 | |||
484 | /* Host to IOA Doorbells */ | ||
485 | #define DOORBELL_RUNTIME_RESET PMC_BIT32(1) | ||
486 | #define DOORBELL_IOA_RESET_ALERT PMC_BIT32(7) | ||
487 | #define DOORBELL_IOA_DEBUG_ALERT PMC_BIT32(9) | ||
488 | #define DOORBELL_ENABLE_DESTRUCTIVE_DIAGS PMC_BIT32(8) | ||
489 | #define DOORBELL_IOA_START_BIST PMC_BIT32(23) | ||
490 | #define DOORBELL_RESET_IOA PMC_BIT32(31) | ||
491 | |||
492 | /* Global interrupt mask register value */ | ||
493 | #define GLOBAL_INTERRUPT_MASK 0x4ULL | ||
494 | |||
495 | #define PMCRAID_ERROR_INTERRUPTS (INTRS_IOARCB_TRANSFER_FAILED | \ | ||
496 | INTRS_IOA_UNIT_CHECK | \ | ||
497 | INTRS_NO_HRRQ_FOR_CMD_RESPONSE | \ | ||
498 | INTRS_IOARRIN_LOST | \ | ||
499 | INTRS_SYSTEM_BUS_MMIO_ERROR | \ | ||
500 | INTRS_IOA_PROCESSOR_ERROR) | ||
501 | |||
502 | #define PMCRAID_PCI_INTERRUPTS (PMCRAID_ERROR_INTERRUPTS | \ | ||
503 | INTRS_HRRQ_VALID | \ | ||
504 | INTRS_CRITICAL_OP_IN_PROGRESS |\ | ||
505 | INTRS_TRANSITION_TO_OPERATIONAL) | ||
506 | |||
507 | /* control_block, associated with each of the commands contains IOARCB, IOADLs | ||
508 | * memory for IOASA. Additional 3 * 16 bytes are allocated in order to support | ||
509 | * additional request parameters (of max size 48) any command. | ||
510 | */ | ||
511 | struct pmcraid_control_block { | ||
512 | struct pmcraid_ioarcb ioarcb; | ||
513 | struct pmcraid_ioadl_desc ioadl[PMCRAID_IOADLS_EXTERNAL + 3]; | ||
514 | struct pmcraid_ioasa ioasa; | ||
515 | } __attribute__ ((packed, aligned(PMCRAID_IOARCB_ALIGNMENT))); | ||
516 | |||
517 | /* pmcraid_sglist - Scatter-gather list allocated for passthrough ioctls | ||
518 | */ | ||
519 | struct pmcraid_sglist { | ||
520 | u32 order; | ||
521 | u32 num_sg; | ||
522 | u32 num_dma_sg; | ||
523 | u32 buffer_len; | ||
524 | struct scatterlist scatterlist[1]; | ||
525 | }; | ||
526 | |||
527 | /* pmcraid_cmd - LLD representation of SCSI command */ | ||
528 | struct pmcraid_cmd { | ||
529 | |||
530 | /* Ptr and bus address of DMA.able control block for this command */ | ||
531 | struct pmcraid_control_block *ioa_cb; | ||
532 | dma_addr_t ioa_cb_bus_addr; | ||
533 | |||
534 | /* sense buffer for REQUEST SENSE command if firmware is not sending | ||
535 | * auto sense data | ||
536 | */ | ||
537 | dma_addr_t sense_buffer_dma; | ||
538 | dma_addr_t dma_handle; | ||
539 | u8 *sense_buffer; | ||
540 | |||
541 | /* pointer to mid layer structure of SCSI commands */ | ||
542 | struct scsi_cmnd *scsi_cmd; | ||
543 | |||
544 | struct list_head free_list; | ||
545 | struct completion wait_for_completion; | ||
546 | struct timer_list timer; /* needed for internal commands */ | ||
547 | u32 timeout; /* current timeout value */ | ||
548 | u32 index; /* index into the command list */ | ||
549 | u8 completion_req; /* for handling internal commands */ | ||
550 | u8 release; /* for handling completions */ | ||
551 | |||
552 | void (*cmd_done) (struct pmcraid_cmd *); | ||
553 | struct pmcraid_instance *drv_inst; | ||
554 | |||
555 | struct pmcraid_sglist *sglist; /* used for passthrough IOCTLs */ | ||
556 | |||
557 | /* scratch used during reset sequence */ | ||
558 | union { | ||
559 | unsigned long time_left; | ||
560 | struct pmcraid_resource_entry *res; | ||
561 | } u; | ||
562 | }; | ||
563 | |||
564 | /* | ||
565 | * Interrupt registers of IOA | ||
566 | */ | ||
567 | struct pmcraid_interrupts { | ||
568 | void __iomem *ioa_host_interrupt_reg; | ||
569 | void __iomem *ioa_host_interrupt_clr_reg; | ||
570 | void __iomem *ioa_host_interrupt_mask_reg; | ||
571 | void __iomem *ioa_host_interrupt_mask_clr_reg; | ||
572 | void __iomem *global_interrupt_mask_reg; | ||
573 | void __iomem *host_ioa_interrupt_reg; | ||
574 | void __iomem *host_ioa_interrupt_clr_reg; | ||
575 | }; | ||
576 | |||
577 | /* ISR parameters LLD allocates (one for each MSI-X if enabled) vectors */ | ||
578 | struct pmcraid_isr_param { | ||
579 | u8 hrrq_id; /* hrrq entry index */ | ||
580 | u16 vector; /* allocated msi-x vector */ | ||
581 | struct pmcraid_instance *drv_inst; | ||
582 | }; | ||
583 | |||
584 | /* AEN message header sent as part of event data to applications */ | ||
585 | struct pmcraid_aen_msg { | ||
586 | u32 hostno; | ||
587 | u32 length; | ||
588 | u8 reserved[8]; | ||
589 | u8 data[0]; | ||
590 | }; | ||
591 | |||
592 | struct pmcraid_hostrcb { | ||
593 | struct pmcraid_instance *drv_inst; | ||
594 | struct pmcraid_aen_msg *msg; | ||
595 | struct pmcraid_hcam_hdr *hcam; /* pointer to hcam buffer */ | ||
596 | struct pmcraid_cmd *cmd; /* pointer to command block used */ | ||
597 | dma_addr_t baddr; /* system address of hcam buffer */ | ||
598 | atomic_t ignore; /* process HCAM response ? */ | ||
599 | }; | ||
600 | |||
601 | #define PMCRAID_AEN_HDR_SIZE sizeof(struct pmcraid_aen_msg) | ||
602 | |||
603 | |||
604 | |||
605 | /* | ||
606 | * Per adapter structure maintained by LLD | ||
607 | */ | ||
608 | struct pmcraid_instance { | ||
609 | /* Array of allowed-to-be-exposed resources, initialized from | ||
610 | * Configutation Table, later updated with CCNs | ||
611 | */ | ||
612 | struct pmcraid_resource_entry *res_entries; | ||
613 | |||
614 | struct list_head free_res_q; /* res_entries lists for easy lookup */ | ||
615 | struct list_head used_res_q; /* List of to be exposed resources */ | ||
616 | spinlock_t resource_lock; /* spinlock to protect resource list */ | ||
617 | |||
618 | void __iomem *mapped_dma_addr; | ||
619 | void __iomem *ioa_status; /* Iomapped IOA status register */ | ||
620 | void __iomem *mailbox; /* Iomapped mailbox register */ | ||
621 | void __iomem *ioarrin; /* IOmapped IOARR IN register */ | ||
622 | |||
623 | struct pmcraid_interrupts int_regs; | ||
624 | struct pmcraid_chip_details *chip_cfg; | ||
625 | |||
626 | /* HostRCBs needed for HCAM */ | ||
627 | struct pmcraid_hostrcb ldn; | ||
628 | struct pmcraid_hostrcb ccn; | ||
629 | |||
630 | |||
631 | /* Bus address of start of HRRQ */ | ||
632 | dma_addr_t hrrq_start_bus_addr[PMCRAID_NUM_MSIX_VECTORS]; | ||
633 | |||
634 | /* Pointer to 1st entry of HRRQ */ | ||
635 | __be32 *hrrq_start[PMCRAID_NUM_MSIX_VECTORS]; | ||
636 | |||
637 | /* Pointer to last entry of HRRQ */ | ||
638 | __be32 *hrrq_end[PMCRAID_NUM_MSIX_VECTORS]; | ||
639 | |||
640 | /* Pointer to current pointer of hrrq */ | ||
641 | __be32 *hrrq_curr[PMCRAID_NUM_MSIX_VECTORS]; | ||
642 | |||
643 | /* Lock for HRRQ access */ | ||
644 | spinlock_t hrrq_lock[PMCRAID_NUM_MSIX_VECTORS]; | ||
645 | |||
646 | /* Expected toggle bit at host */ | ||
647 | u8 host_toggle_bit[PMCRAID_NUM_MSIX_VECTORS]; | ||
648 | |||
649 | /* No of Reset IOA retries . IOA marked dead if threshold exceeds */ | ||
650 | u8 ioa_reset_attempts; | ||
651 | #define PMCRAID_RESET_ATTEMPTS 3 | ||
652 | |||
653 | /* Wait Q for threads to wait for Reset IOA completion */ | ||
654 | wait_queue_head_t reset_wait_q; | ||
655 | struct pmcraid_cmd *reset_cmd; | ||
656 | |||
657 | /* structures for supporting SIGIO based AEN. */ | ||
658 | struct fasync_struct *aen_queue; | ||
659 | struct mutex aen_queue_lock; /* lock for aen subscribers list */ | ||
660 | struct cdev cdev; | ||
661 | |||
662 | struct Scsi_Host *host; /* mid layer interface structure handle */ | ||
663 | struct pci_dev *pdev; /* PCI device structure handle */ | ||
664 | |||
665 | u8 current_log_level; /* default level for logging IOASC errors */ | ||
666 | |||
667 | u8 num_hrrq; /* Number of interrupt vectors allocated */ | ||
668 | dev_t dev; /* Major-Minor numbers for Char device */ | ||
669 | |||
670 | /* Used as ISR handler argument */ | ||
671 | struct pmcraid_isr_param hrrq_vector[PMCRAID_NUM_MSIX_VECTORS]; | ||
672 | |||
673 | /* configuration table */ | ||
674 | struct pmcraid_config_table *cfg_table; | ||
675 | dma_addr_t cfg_table_bus_addr; | ||
676 | |||
677 | /* structures related to command blocks */ | ||
678 | struct kmem_cache *cmd_cachep; /* cache for cmd blocks */ | ||
679 | struct pci_pool *control_pool; /* pool for control blocks */ | ||
680 | char cmd_pool_name[64]; /* name of cmd cache */ | ||
681 | char ctl_pool_name[64]; /* name of control cache */ | ||
682 | |||
683 | struct pmcraid_cmd *cmd_list[PMCRAID_MAX_CMD]; | ||
684 | |||
685 | struct list_head free_cmd_pool; | ||
686 | struct list_head pending_cmd_pool; | ||
687 | spinlock_t free_pool_lock; /* free pool lock */ | ||
688 | spinlock_t pending_pool_lock; /* pending pool lock */ | ||
689 | |||
690 | /* No of IO commands pending with FW */ | ||
691 | atomic_t outstanding_cmds; | ||
692 | |||
693 | /* should add/delete resources to mid-layer now ?*/ | ||
694 | atomic_t expose_resources; | ||
695 | |||
696 | /* Tasklet to handle deferred processing */ | ||
697 | struct tasklet_struct isr_tasklet[PMCRAID_NUM_MSIX_VECTORS]; | ||
698 | |||
699 | /* Work-queue (Shared) for deferred reset processing */ | ||
700 | struct work_struct worker_q; | ||
701 | |||
702 | |||
703 | u32 ioa_state:4; /* For IOA Reset sequence FSM */ | ||
704 | #define IOA_STATE_OPERATIONAL 0x0 | ||
705 | #define IOA_STATE_UNKNOWN 0x1 | ||
706 | #define IOA_STATE_DEAD 0x2 | ||
707 | #define IOA_STATE_IN_SOFT_RESET 0x3 | ||
708 | #define IOA_STATE_IN_HARD_RESET 0x4 | ||
709 | #define IOA_STATE_IN_RESET_ALERT 0x5 | ||
710 | #define IOA_STATE_IN_BRINGDOWN 0x6 | ||
711 | #define IOA_STATE_IN_BRINGUP 0x7 | ||
712 | |||
713 | u32 ioa_reset_in_progress:1; /* true if IOA reset is in progress */ | ||
714 | u32 ioa_hard_reset:1; /* TRUE if Hard Reset is needed */ | ||
715 | u32 ioa_unit_check:1; /* Indicates Unit Check condition */ | ||
716 | u32 ioa_bringdown:1; /* whether IOA needs to be brought down */ | ||
717 | u32 force_ioa_reset:1; /* force adapter reset ? */ | ||
718 | u32 reinit_cfg_table:1; /* reinit config table due to lost CCN */ | ||
719 | u32 ioa_shutdown_type:2;/* shutdown type used during reset */ | ||
720 | #define SHUTDOWN_NONE 0x0 | ||
721 | #define SHUTDOWN_NORMAL 0x1 | ||
722 | #define SHUTDOWN_ABBREV 0x2 | ||
723 | |||
724 | }; | ||
725 | |||
726 | /* LLD maintained resource entry structure */ | ||
727 | struct pmcraid_resource_entry { | ||
728 | struct list_head queue; /* link to "to be exposed" resources */ | ||
729 | struct pmcraid_config_table_entry cfg_entry; | ||
730 | struct scsi_device *scsi_dev; /* Link scsi_device structure */ | ||
731 | atomic_t read_failures; /* count of failed READ commands */ | ||
732 | atomic_t write_failures; /* count of failed WRITE commands */ | ||
733 | |||
734 | /* To indicate add/delete/modify during CCN */ | ||
735 | u8 change_detected; | ||
736 | #define RES_CHANGE_ADD 0x1 /* add this to mid-layer */ | ||
737 | #define RES_CHANGE_DEL 0x2 /* remove this from mid-layer */ | ||
738 | |||
739 | u8 reset_progress; /* Device is resetting */ | ||
740 | |||
741 | /* | ||
742 | * When IOA asks for sync (i.e. IOASC = Not Ready, Sync Required), this | ||
743 | * flag will be set, mid layer will be asked to retry. In the next | ||
744 | * attempt, this flag will be checked in queuecommand() to set | ||
745 | * SYNC_COMPLETE flag in IOARCB (flag_0). | ||
746 | */ | ||
747 | u8 sync_reqd; | ||
748 | |||
749 | /* target indicates the mapped target_id assigned to this resource if | ||
750 | * this is VSET resource. For non-VSET resources this will be un-used | ||
751 | * or zero | ||
752 | */ | ||
753 | u8 target; | ||
754 | }; | ||
755 | |||
756 | /* Data structures used in IOASC error code logging */ | ||
757 | struct pmcraid_ioasc_error { | ||
758 | u32 ioasc_code; /* IOASC code */ | ||
759 | u8 log_level; /* default log level assignment. */ | ||
760 | char *error_string; | ||
761 | }; | ||
762 | |||
763 | /* Initial log_level assignments for various IOASCs */ | ||
764 | #define IOASC_LOG_LEVEL_NONE 0x0 /* no logging */ | ||
765 | #define IOASC_LOG_LEVEL_MUST 0x1 /* must log: all high-severity errors */ | ||
766 | #define IOASC_LOG_LEVEL_HARD 0x2 /* optional – low severity errors */ | ||
767 | |||
768 | /* Error information maintained by LLD. LLD initializes the pmcraid_error_table | ||
769 | * statically. | ||
770 | */ | ||
771 | static struct pmcraid_ioasc_error pmcraid_ioasc_error_table[] = { | ||
772 | {0x01180600, IOASC_LOG_LEVEL_MUST, | ||
773 | "Recovered Error, soft media error, sector reassignment suggested"}, | ||
774 | {0x015D0000, IOASC_LOG_LEVEL_MUST, | ||
775 | "Recovered Error, failure prediction thresold exceeded"}, | ||
776 | {0x015D9200, IOASC_LOG_LEVEL_MUST, | ||
777 | "Recovered Error, soft Cache Card Battery error thresold"}, | ||
778 | {0x015D9200, IOASC_LOG_LEVEL_MUST, | ||
779 | "Recovered Error, soft Cache Card Battery error thresold"}, | ||
780 | {0x02048000, IOASC_LOG_LEVEL_MUST, | ||
781 | "Not Ready, IOA Reset Required"}, | ||
782 | {0x02408500, IOASC_LOG_LEVEL_MUST, | ||
783 | "Not Ready, IOA microcode download required"}, | ||
784 | {0x03110B00, IOASC_LOG_LEVEL_MUST, | ||
785 | "Medium Error, data unreadable, reassignment suggested"}, | ||
786 | {0x03110C00, IOASC_LOG_LEVEL_MUST, | ||
787 | "Medium Error, data unreadable do not reassign"}, | ||
788 | {0x03310000, IOASC_LOG_LEVEL_MUST, | ||
789 | "Medium Error, media corrupted"}, | ||
790 | {0x04050000, IOASC_LOG_LEVEL_MUST, | ||
791 | "Hardware Error, IOA can't communicate with device"}, | ||
792 | {0x04080000, IOASC_LOG_LEVEL_MUST, | ||
793 | "Hardware Error, device bus error"}, | ||
794 | {0x04080000, IOASC_LOG_LEVEL_MUST, | ||
795 | "Hardware Error, device bus is not functioning"}, | ||
796 | {0x04118000, IOASC_LOG_LEVEL_MUST, | ||
797 | "Hardware Error, IOA reserved area data check"}, | ||
798 | {0x04118100, IOASC_LOG_LEVEL_MUST, | ||
799 | "Hardware Error, IOA reserved area invalid data pattern"}, | ||
800 | {0x04118200, IOASC_LOG_LEVEL_MUST, | ||
801 | "Hardware Error, IOA reserved area LRC error"}, | ||
802 | {0x04320000, IOASC_LOG_LEVEL_MUST, | ||
803 | "Hardware Error, reassignment space exhausted"}, | ||
804 | {0x04330000, IOASC_LOG_LEVEL_MUST, | ||
805 | "Hardware Error, data transfer underlength error"}, | ||
806 | {0x04330000, IOASC_LOG_LEVEL_MUST, | ||
807 | "Hardware Error, data transfer overlength error"}, | ||
808 | {0x04418000, IOASC_LOG_LEVEL_MUST, | ||
809 | "Hardware Error, PCI bus error"}, | ||
810 | {0x04440000, IOASC_LOG_LEVEL_MUST, | ||
811 | "Hardware Error, device error"}, | ||
812 | {0x04448300, IOASC_LOG_LEVEL_MUST, | ||
813 | "Hardware Error, undefined device response"}, | ||
814 | {0x04448400, IOASC_LOG_LEVEL_MUST, | ||
815 | "Hardware Error, IOA microcode error"}, | ||
816 | {0x04448600, IOASC_LOG_LEVEL_MUST, | ||
817 | "Hardware Error, IOA reset required"}, | ||
818 | {0x04449200, IOASC_LOG_LEVEL_MUST, | ||
819 | "Hardware Error, hard Cache Fearuee Card Battery error"}, | ||
820 | {0x0444A000, IOASC_LOG_LEVEL_MUST, | ||
821 | "Hardware Error, failed device altered"}, | ||
822 | {0x0444A200, IOASC_LOG_LEVEL_MUST, | ||
823 | "Hardware Error, data check after reassignment"}, | ||
824 | {0x0444A300, IOASC_LOG_LEVEL_MUST, | ||
825 | "Hardware Error, LRC error after reassignment"}, | ||
826 | {0x044A0000, IOASC_LOG_LEVEL_MUST, | ||
827 | "Hardware Error, device bus error (msg/cmd phase)"}, | ||
828 | {0x04670400, IOASC_LOG_LEVEL_MUST, | ||
829 | "Hardware Error, new device can't be used"}, | ||
830 | {0x04678000, IOASC_LOG_LEVEL_MUST, | ||
831 | "Hardware Error, invalid multiadapter configuration"}, | ||
832 | {0x04678100, IOASC_LOG_LEVEL_MUST, | ||
833 | "Hardware Error, incorrect connection between enclosures"}, | ||
834 | {0x04678200, IOASC_LOG_LEVEL_MUST, | ||
835 | "Hardware Error, connections exceed IOA design limits"}, | ||
836 | {0x04678300, IOASC_LOG_LEVEL_MUST, | ||
837 | "Hardware Error, incorrect multipath connection"}, | ||
838 | {0x04679000, IOASC_LOG_LEVEL_MUST, | ||
839 | "Hardware Error, command to LUN failed"}, | ||
840 | {0x064C8000, IOASC_LOG_LEVEL_HARD, | ||
841 | "Unit Attention, cache exists for missing/failed device"}, | ||
842 | {0x06670100, IOASC_LOG_LEVEL_HARD, | ||
843 | "Unit Attention, incompatible exposed mode device"}, | ||
844 | {0x06670600, IOASC_LOG_LEVEL_HARD, | ||
845 | "Unit Attention, attachment of logical unit failed"}, | ||
846 | {0x06678000, IOASC_LOG_LEVEL_MUST, | ||
847 | "Unit Attention, cables exceed connective design limit"}, | ||
848 | {0x06678300, IOASC_LOG_LEVEL_MUST, | ||
849 | "Unit Attention, incomplete multipath connection between" \ | ||
850 | "IOA and enclosure"}, | ||
851 | {0x06678400, IOASC_LOG_LEVEL_MUST, | ||
852 | "Unit Attention, incomplete multipath connection between" \ | ||
853 | "device and enclosure"}, | ||
854 | {0x06678500, IOASC_LOG_LEVEL_MUST, | ||
855 | "Unit Attention, incomplete multipath connection between" \ | ||
856 | "IOA and remote IOA"}, | ||
857 | {0x06678600, IOASC_LOG_LEVEL_HARD, | ||
858 | "Unit Attention, missing remote IOA"}, | ||
859 | {0x06679100, IOASC_LOG_LEVEL_HARD, | ||
860 | "Unit Attention, enclosure doesn't support required multipath" \ | ||
861 | "function"}, | ||
862 | {0x06698200, IOASC_LOG_LEVEL_HARD, | ||
863 | "Unit Attention, corrupt array parity detected on device"}, | ||
864 | {0x066B0200, IOASC_LOG_LEVEL_MUST, | ||
865 | "Unit Attention, array exposed"}, | ||
866 | {0x066B8200, IOASC_LOG_LEVEL_HARD, | ||
867 | "Unit Attention, exposed array is still protected"}, | ||
868 | {0x066B9200, IOASC_LOG_LEVEL_MUST, | ||
869 | "Unit Attention, Multipath redundancy level got worse"}, | ||
870 | {0x07270000, IOASC_LOG_LEVEL_HARD, | ||
871 | "Data Protect, device is read/write protected by IOA"}, | ||
872 | {0x07278000, IOASC_LOG_LEVEL_HARD, | ||
873 | "Data Protect, IOA doesn't support device attribute"}, | ||
874 | {0x07278100, IOASC_LOG_LEVEL_HARD, | ||
875 | "Data Protect, NVRAM mirroring prohibited"}, | ||
876 | {0x07278400, IOASC_LOG_LEVEL_MUST, | ||
877 | "Data Protect, array is short 2 or more devices"}, | ||
878 | {0x07278600, IOASC_LOG_LEVEL_MUST, | ||
879 | "Data Protect, exposed array is short a required device"}, | ||
880 | {0x07278700, IOASC_LOG_LEVEL_MUST, | ||
881 | "Data Protect, array members not at required addresses"}, | ||
882 | {0x07278800, IOASC_LOG_LEVEL_MUST, | ||
883 | "Data Protect, exposed mode device resource address conflict"}, | ||
884 | {0x07278900, IOASC_LOG_LEVEL_MUST, | ||
885 | "Data Protect, incorrect resource address of exposed mode device"}, | ||
886 | {0x07278A00, IOASC_LOG_LEVEL_MUST, | ||
887 | "Data Protect, Array is missing a device and parity is out of sync"}, | ||
888 | {0x07278B00, IOASC_LOG_LEVEL_MUST, | ||
889 | "Data Protect, maximum number of arrays already exist"}, | ||
890 | {0x07278C00, IOASC_LOG_LEVEL_HARD, | ||
891 | "Data Protect, cannot locate cache data for device"}, | ||
892 | {0x07278D00, IOASC_LOG_LEVEL_HARD, | ||
893 | "Data Protect, cache data exits for a changed device"}, | ||
894 | {0x07279100, IOASC_LOG_LEVEL_MUST, | ||
895 | "Data Protect, detection of a device requiring format"}, | ||
896 | {0x07279200, IOASC_LOG_LEVEL_MUST, | ||
897 | "Data Protect, IOA exceeds maximum number of devices"}, | ||
898 | {0x07279600, IOASC_LOG_LEVEL_MUST, | ||
899 | "Data Protect, missing array, volume set is not functional"}, | ||
900 | {0x07279700, IOASC_LOG_LEVEL_MUST, | ||
901 | "Data Protect, single device for a volume set"}, | ||
902 | {0x07279800, IOASC_LOG_LEVEL_MUST, | ||
903 | "Data Protect, missing multiple devices for a volume set"}, | ||
904 | {0x07279900, IOASC_LOG_LEVEL_HARD, | ||
905 | "Data Protect, maximum number of volument sets already exists"}, | ||
906 | {0x07279A00, IOASC_LOG_LEVEL_MUST, | ||
907 | "Data Protect, other volume set problem"}, | ||
908 | }; | ||
909 | |||
910 | /* macros to help in debugging */ | ||
911 | #define pmcraid_err(...) \ | ||
912 | printk(KERN_ERR "MaxRAID: "__VA_ARGS__) | ||
913 | |||
914 | #define pmcraid_info(...) \ | ||
915 | if (pmcraid_debug_log) \ | ||
916 | printk(KERN_INFO "MaxRAID: "__VA_ARGS__) | ||
917 | |||
918 | /* check if given command is a SCSI READ or SCSI WRITE command */ | ||
919 | #define SCSI_READ_CMD 0x1 /* any of SCSI READ commands */ | ||
920 | #define SCSI_WRITE_CMD 0x2 /* any of SCSI WRITE commands */ | ||
921 | #define SCSI_CMD_TYPE(opcode) \ | ||
922 | ({ u8 op = opcode; u8 __type = 0;\ | ||
923 | if (op == READ_6 || op == READ_10 || op == READ_12 || op == READ_16)\ | ||
924 | __type = SCSI_READ_CMD;\ | ||
925 | else if (op == WRITE_6 || op == WRITE_10 || op == WRITE_12 || \ | ||
926 | op == WRITE_16)\ | ||
927 | __type = SCSI_WRITE_CMD;\ | ||
928 | __type;\ | ||
929 | }) | ||
930 | |||
931 | #define IS_SCSI_READ_WRITE(opcode) \ | ||
932 | ({ u8 __type = SCSI_CMD_TYPE(opcode); \ | ||
933 | (__type == SCSI_READ_CMD || __type == SCSI_WRITE_CMD) ? 1 : 0;\ | ||
934 | }) | ||
935 | |||
936 | |||
937 | /* | ||
938 | * pmcraid_ioctl_header - definition of header structure that preceeds all the | ||
939 | * buffers given as ioctl arguements. | ||
940 | * | ||
941 | * .signature : always ASCII string, "PMCRAID" | ||
942 | * .reserved : not used | ||
943 | * .buffer_length : length of the buffer following the header | ||
944 | */ | ||
945 | struct pmcraid_ioctl_header { | ||
946 | u8 signature[8]; | ||
947 | u32 reserved; | ||
948 | u32 buffer_length; | ||
949 | }; | ||
950 | |||
951 | #define PMCRAID_IOCTL_SIGNATURE "PMCRAID" | ||
952 | |||
953 | |||
954 | /* | ||
955 | * pmcraid_event_details - defines AEN details that apps can retrieve from LLD | ||
956 | * | ||
957 | * .rcb_ccn - complete RCB of CCN | ||
958 | * .rcb_ldn - complete RCB of CCN | ||
959 | */ | ||
960 | struct pmcraid_event_details { | ||
961 | struct pmcraid_hcam_ccn rcb_ccn; | ||
962 | struct pmcraid_hcam_ldn rcb_ldn; | ||
963 | }; | ||
964 | |||
965 | /* | ||
966 | * pmcraid_driver_ioctl_buffer - structure passed as argument to most of the | ||
967 | * PMC driver handled ioctls. | ||
968 | */ | ||
969 | struct pmcraid_driver_ioctl_buffer { | ||
970 | struct pmcraid_ioctl_header ioctl_header; | ||
971 | struct pmcraid_event_details event_details; | ||
972 | }; | ||
973 | |||
974 | /* | ||
975 | * pmcraid_passthrough_ioctl_buffer - structure given as argument to | ||
976 | * passthrough(or firmware handled) IOCTL commands. Note that ioarcb requires | ||
977 | * 32-byte alignment so, it is necessary to pack this structure to avoid any | ||
978 | * holes between ioctl_header and passthrough buffer | ||
979 | * | ||
980 | * .ioactl_header : ioctl header | ||
981 | * .ioarcb : filled-up ioarcb buffer, driver always reads this buffer | ||
982 | * .ioasa : buffer for ioasa, driver fills this with IOASA from firmware | ||
983 | * .request_buffer: The I/O buffer (flat), driver reads/writes to this based on | ||
984 | * the transfer directions passed in ioarcb.flags0. Contents | ||
985 | * of this buffer are valid only when ioarcb.data_transfer_len | ||
986 | * is not zero. | ||
987 | */ | ||
988 | struct pmcraid_passthrough_ioctl_buffer { | ||
989 | struct pmcraid_ioctl_header ioctl_header; | ||
990 | struct pmcraid_ioarcb ioarcb; | ||
991 | struct pmcraid_ioasa ioasa; | ||
992 | u8 request_buffer[1]; | ||
993 | } __attribute__ ((packed)); | ||
994 | |||
995 | /* | ||
996 | * keys to differentiate between driver handled IOCTLs and passthrough | ||
997 | * IOCTLs passed to IOA. driver determines the ioctl type using macro | ||
998 | * _IOC_TYPE | ||
999 | */ | ||
1000 | #define PMCRAID_DRIVER_IOCTL 'D' | ||
1001 | #define PMCRAID_PASSTHROUGH_IOCTL 'F' | ||
1002 | |||
1003 | #define DRV_IOCTL(n, size) \ | ||
1004 | _IOC(_IOC_READ|_IOC_WRITE, PMCRAID_DRIVER_IOCTL, (n), (size)) | ||
1005 | |||
1006 | #define FMW_IOCTL(n, size) \ | ||
1007 | _IOC(_IOC_READ|_IOC_WRITE, PMCRAID_PASSTHROUGH_IOCTL, (n), (size)) | ||
1008 | |||
1009 | /* | ||
1010 | * _ARGSIZE: macro that gives size of the argument type passed to an IOCTL cmd. | ||
1011 | * This is to facilitate applications avoiding un-necessary memory allocations. | ||
1012 | * For example, most of driver handled ioctls do not require ioarcb, ioasa. | ||
1013 | */ | ||
1014 | #define _ARGSIZE(arg) (sizeof(struct pmcraid_ioctl_header) + sizeof(arg)) | ||
1015 | |||
1016 | /* Driver handled IOCTL command definitions */ | ||
1017 | |||
1018 | #define PMCRAID_IOCTL_RESET_ADAPTER \ | ||
1019 | DRV_IOCTL(5, sizeof(struct pmcraid_ioctl_header)) | ||
1020 | |||
1021 | /* passthrough/firmware handled commands */ | ||
1022 | #define PMCRAID_IOCTL_PASSTHROUGH_COMMAND \ | ||
1023 | FMW_IOCTL(1, sizeof(struct pmcraid_passthrough_ioctl_buffer)) | ||
1024 | |||
1025 | #define PMCRAID_IOCTL_DOWNLOAD_MICROCODE \ | ||
1026 | FMW_IOCTL(2, sizeof(struct pmcraid_passthrough_ioctl_buffer)) | ||
1027 | |||
1028 | |||
1029 | #endif /* _PMCRAID_H */ | ||
diff --git a/drivers/scsi/qla2xxx/qla_attr.c b/drivers/scsi/qla2xxx/qla_attr.c index 0f8796201504..fbcb82a2f7f4 100644 --- a/drivers/scsi/qla2xxx/qla_attr.c +++ b/drivers/scsi/qla2xxx/qla_attr.c | |||
@@ -1670,7 +1670,7 @@ qla24xx_vport_create(struct fc_vport *fc_vport, bool disable) | |||
1670 | 1670 | ||
1671 | qla24xx_vport_disable(fc_vport, disable); | 1671 | qla24xx_vport_disable(fc_vport, disable); |
1672 | 1672 | ||
1673 | if (ql2xmultique_tag) { | 1673 | if (ha->flags.cpu_affinity_enabled) { |
1674 | req = ha->req_q_map[1]; | 1674 | req = ha->req_q_map[1]; |
1675 | goto vport_queue; | 1675 | goto vport_queue; |
1676 | } else if (ql2xmaxqueues == 1 || !ha->npiv_info) | 1676 | } else if (ql2xmaxqueues == 1 || !ha->npiv_info) |
@@ -1736,6 +1736,11 @@ qla24xx_vport_delete(struct fc_vport *fc_vport) | |||
1736 | 1736 | ||
1737 | qla24xx_deallocate_vp_id(vha); | 1737 | qla24xx_deallocate_vp_id(vha); |
1738 | 1738 | ||
1739 | mutex_lock(&ha->vport_lock); | ||
1740 | ha->cur_vport_count--; | ||
1741 | clear_bit(vha->vp_idx, ha->vp_idx_map); | ||
1742 | mutex_unlock(&ha->vport_lock); | ||
1743 | |||
1739 | if (vha->timer_active) { | 1744 | if (vha->timer_active) { |
1740 | qla2x00_vp_stop_timer(vha); | 1745 | qla2x00_vp_stop_timer(vha); |
1741 | DEBUG15(printk ("scsi(%ld): timer for the vport[%d] = %p " | 1746 | DEBUG15(printk ("scsi(%ld): timer for the vport[%d] = %p " |
@@ -1743,7 +1748,7 @@ qla24xx_vport_delete(struct fc_vport *fc_vport) | |||
1743 | vha->host_no, vha->vp_idx, vha)); | 1748 | vha->host_no, vha->vp_idx, vha)); |
1744 | } | 1749 | } |
1745 | 1750 | ||
1746 | if (vha->req->id && !ql2xmultique_tag) { | 1751 | if (vha->req->id && !ha->flags.cpu_affinity_enabled) { |
1747 | if (qla25xx_delete_req_que(vha, vha->req) != QLA_SUCCESS) | 1752 | if (qla25xx_delete_req_que(vha, vha->req) != QLA_SUCCESS) |
1748 | qla_printk(KERN_WARNING, ha, | 1753 | qla_printk(KERN_WARNING, ha, |
1749 | "Queue delete failed.\n"); | 1754 | "Queue delete failed.\n"); |
diff --git a/drivers/scsi/qla2xxx/qla_def.h b/drivers/scsi/qla2xxx/qla_def.h index 00aa48d975a6..215061861794 100644 --- a/drivers/scsi/qla2xxx/qla_def.h +++ b/drivers/scsi/qla2xxx/qla_def.h | |||
@@ -189,6 +189,7 @@ struct req_que; | |||
189 | */ | 189 | */ |
190 | typedef struct srb { | 190 | typedef struct srb { |
191 | struct fc_port *fcport; | 191 | struct fc_port *fcport; |
192 | uint32_t handle; | ||
192 | 193 | ||
193 | struct scsi_cmnd *cmd; /* Linux SCSI command pkt */ | 194 | struct scsi_cmnd *cmd; /* Linux SCSI command pkt */ |
194 | 195 | ||
@@ -196,6 +197,8 @@ typedef struct srb { | |||
196 | 197 | ||
197 | uint32_t request_sense_length; | 198 | uint32_t request_sense_length; |
198 | uint8_t *request_sense_ptr; | 199 | uint8_t *request_sense_ptr; |
200 | |||
201 | void *ctx; | ||
199 | } srb_t; | 202 | } srb_t; |
200 | 203 | ||
201 | /* | 204 | /* |
@@ -204,6 +207,28 @@ typedef struct srb { | |||
204 | #define SRB_DMA_VALID BIT_0 /* Command sent to ISP */ | 207 | #define SRB_DMA_VALID BIT_0 /* Command sent to ISP */ |
205 | 208 | ||
206 | /* | 209 | /* |
210 | * SRB extensions. | ||
211 | */ | ||
212 | struct srb_ctx { | ||
213 | #define SRB_LOGIN_CMD 1 | ||
214 | #define SRB_LOGOUT_CMD 2 | ||
215 | uint16_t type; | ||
216 | struct timer_list timer; | ||
217 | |||
218 | void (*free)(srb_t *sp); | ||
219 | void (*timeout)(srb_t *sp); | ||
220 | }; | ||
221 | |||
222 | struct srb_logio { | ||
223 | struct srb_ctx ctx; | ||
224 | |||
225 | #define SRB_LOGIN_RETRIED BIT_0 | ||
226 | #define SRB_LOGIN_COND_PLOGI BIT_1 | ||
227 | #define SRB_LOGIN_SKIP_PRLI BIT_2 | ||
228 | uint16_t flags; | ||
229 | }; | ||
230 | |||
231 | /* | ||
207 | * ISP I/O Register Set structure definitions. | 232 | * ISP I/O Register Set structure definitions. |
208 | */ | 233 | */ |
209 | struct device_reg_2xxx { | 234 | struct device_reg_2xxx { |
@@ -1482,7 +1507,7 @@ typedef union { | |||
1482 | uint8_t domain; | 1507 | uint8_t domain; |
1483 | uint8_t area; | 1508 | uint8_t area; |
1484 | uint8_t al_pa; | 1509 | uint8_t al_pa; |
1485 | #elif __LITTLE_ENDIAN | 1510 | #elif defined(__LITTLE_ENDIAN) |
1486 | uint8_t al_pa; | 1511 | uint8_t al_pa; |
1487 | uint8_t area; | 1512 | uint8_t area; |
1488 | uint8_t domain; | 1513 | uint8_t domain; |
@@ -1565,6 +1590,7 @@ typedef struct fc_port { | |||
1565 | #define FCF_FABRIC_DEVICE BIT_0 | 1590 | #define FCF_FABRIC_DEVICE BIT_0 |
1566 | #define FCF_LOGIN_NEEDED BIT_1 | 1591 | #define FCF_LOGIN_NEEDED BIT_1 |
1567 | #define FCF_TAPE_PRESENT BIT_2 | 1592 | #define FCF_TAPE_PRESENT BIT_2 |
1593 | #define FCF_FCP2_DEVICE BIT_3 | ||
1568 | 1594 | ||
1569 | /* No loop ID flag. */ | 1595 | /* No loop ID flag. */ |
1570 | #define FC_NO_LOOP_ID 0x1000 | 1596 | #define FC_NO_LOOP_ID 0x1000 |
@@ -2093,6 +2119,10 @@ struct qla_msix_entry { | |||
2093 | enum qla_work_type { | 2119 | enum qla_work_type { |
2094 | QLA_EVT_AEN, | 2120 | QLA_EVT_AEN, |
2095 | QLA_EVT_IDC_ACK, | 2121 | QLA_EVT_IDC_ACK, |
2122 | QLA_EVT_ASYNC_LOGIN, | ||
2123 | QLA_EVT_ASYNC_LOGIN_DONE, | ||
2124 | QLA_EVT_ASYNC_LOGOUT, | ||
2125 | QLA_EVT_ASYNC_LOGOUT_DONE, | ||
2096 | }; | 2126 | }; |
2097 | 2127 | ||
2098 | 2128 | ||
@@ -2111,6 +2141,11 @@ struct qla_work_evt { | |||
2111 | #define QLA_IDC_ACK_REGS 7 | 2141 | #define QLA_IDC_ACK_REGS 7 |
2112 | uint16_t mb[QLA_IDC_ACK_REGS]; | 2142 | uint16_t mb[QLA_IDC_ACK_REGS]; |
2113 | } idc_ack; | 2143 | } idc_ack; |
2144 | struct { | ||
2145 | struct fc_port *fcport; | ||
2146 | #define QLA_LOGIO_LOGIN_RETRIED BIT_0 | ||
2147 | u16 data[2]; | ||
2148 | } logio; | ||
2114 | } u; | 2149 | } u; |
2115 | }; | 2150 | }; |
2116 | 2151 | ||
@@ -2224,6 +2259,7 @@ struct qla_hw_data { | |||
2224 | uint32_t chip_reset_done :1; | 2259 | uint32_t chip_reset_done :1; |
2225 | uint32_t port0 :1; | 2260 | uint32_t port0 :1; |
2226 | uint32_t running_gold_fw :1; | 2261 | uint32_t running_gold_fw :1; |
2262 | uint32_t cpu_affinity_enabled :1; | ||
2227 | } flags; | 2263 | } flags; |
2228 | 2264 | ||
2229 | /* This spinlock is used to protect "io transactions", you must | 2265 | /* This spinlock is used to protect "io transactions", you must |
@@ -2350,6 +2386,7 @@ struct qla_hw_data { | |||
2350 | (ha)->flags.msix_enabled) | 2386 | (ha)->flags.msix_enabled) |
2351 | #define IS_FAC_REQUIRED(ha) (IS_QLA81XX(ha)) | 2387 | #define IS_FAC_REQUIRED(ha) (IS_QLA81XX(ha)) |
2352 | #define IS_NOCACHE_VPD_TYPE(ha) (IS_QLA81XX(ha)) | 2388 | #define IS_NOCACHE_VPD_TYPE(ha) (IS_QLA81XX(ha)) |
2389 | #define IS_ALOGIO_CAPABLE(ha) (IS_QLA23XX(ha) || IS_FWI2_CAPABLE(ha)) | ||
2353 | 2390 | ||
2354 | #define IS_IIDMA_CAPABLE(ha) ((ha)->device_type & DT_IIDMA) | 2391 | #define IS_IIDMA_CAPABLE(ha) ((ha)->device_type & DT_IIDMA) |
2355 | #define IS_FWI2_CAPABLE(ha) ((ha)->device_type & DT_FWI2) | 2392 | #define IS_FWI2_CAPABLE(ha) ((ha)->device_type & DT_FWI2) |
diff --git a/drivers/scsi/qla2xxx/qla_fw.h b/drivers/scsi/qla2xxx/qla_fw.h index dfde2dd865cb..66a8da5d7d08 100644 --- a/drivers/scsi/qla2xxx/qla_fw.h +++ b/drivers/scsi/qla2xxx/qla_fw.h | |||
@@ -1126,7 +1126,7 @@ struct vp_config_entry_24xx { | |||
1126 | uint16_t id; | 1126 | uint16_t id; |
1127 | uint16_t reserved_4; | 1127 | uint16_t reserved_4; |
1128 | uint16_t hopct; | 1128 | uint16_t hopct; |
1129 | uint8_t reserved_5; | 1129 | uint8_t reserved_5[2]; |
1130 | }; | 1130 | }; |
1131 | 1131 | ||
1132 | #define VP_RPT_ID_IOCB_TYPE 0x32 /* Report ID Acquisition entry. */ | 1132 | #define VP_RPT_ID_IOCB_TYPE 0x32 /* Report ID Acquisition entry. */ |
diff --git a/drivers/scsi/qla2xxx/qla_gbl.h b/drivers/scsi/qla2xxx/qla_gbl.h index 65b12d82867c..f3d1d1afa95b 100644 --- a/drivers/scsi/qla2xxx/qla_gbl.h +++ b/drivers/scsi/qla2xxx/qla_gbl.h | |||
@@ -52,6 +52,14 @@ extern void qla2x00_try_to_stop_firmware(scsi_qla_host_t *); | |||
52 | 52 | ||
53 | extern void qla84xx_put_chip(struct scsi_qla_host *); | 53 | extern void qla84xx_put_chip(struct scsi_qla_host *); |
54 | 54 | ||
55 | extern int qla2x00_async_login(struct scsi_qla_host *, fc_port_t *, | ||
56 | uint16_t *); | ||
57 | extern int qla2x00_async_logout(struct scsi_qla_host *, fc_port_t *); | ||
58 | extern int qla2x00_async_login_done(struct scsi_qla_host *, fc_port_t *, | ||
59 | uint16_t *); | ||
60 | extern int qla2x00_async_logout_done(struct scsi_qla_host *, fc_port_t *, | ||
61 | uint16_t *); | ||
62 | |||
55 | /* | 63 | /* |
56 | * Global Data in qla_os.c source file. | 64 | * Global Data in qla_os.c source file. |
57 | */ | 65 | */ |
@@ -76,6 +84,15 @@ extern void qla2x00_abort_all_cmds(scsi_qla_host_t *, int); | |||
76 | extern int qla2x00_post_aen_work(struct scsi_qla_host *, enum | 84 | extern int qla2x00_post_aen_work(struct scsi_qla_host *, enum |
77 | fc_host_event_code, u32); | 85 | fc_host_event_code, u32); |
78 | extern int qla2x00_post_idc_ack_work(struct scsi_qla_host *, uint16_t *); | 86 | extern int qla2x00_post_idc_ack_work(struct scsi_qla_host *, uint16_t *); |
87 | extern int qla2x00_post_async_login_work(struct scsi_qla_host *, fc_port_t *, | ||
88 | uint16_t *); | ||
89 | extern int qla2x00_post_async_login_done_work(struct scsi_qla_host *, | ||
90 | fc_port_t *, uint16_t *); | ||
91 | extern int qla2x00_post_async_logout_work(struct scsi_qla_host *, fc_port_t *, | ||
92 | uint16_t *); | ||
93 | extern int qla2x00_post_async_logout_done_work(struct scsi_qla_host *, | ||
94 | fc_port_t *, uint16_t *); | ||
95 | |||
79 | extern int qla81xx_restart_mpi_firmware(scsi_qla_host_t *); | 96 | extern int qla81xx_restart_mpi_firmware(scsi_qla_host_t *); |
80 | 97 | ||
81 | extern void qla2x00_abort_fcport_cmds(fc_port_t *); | 98 | extern void qla2x00_abort_fcport_cmds(fc_port_t *); |
@@ -83,6 +100,8 @@ extern struct scsi_qla_host *qla2x00_create_host(struct scsi_host_template *, | |||
83 | struct qla_hw_data *); | 100 | struct qla_hw_data *); |
84 | extern void qla2x00_free_host(struct scsi_qla_host *); | 101 | extern void qla2x00_free_host(struct scsi_qla_host *); |
85 | extern void qla2x00_relogin(struct scsi_qla_host *); | 102 | extern void qla2x00_relogin(struct scsi_qla_host *); |
103 | extern void qla2x00_do_work(struct scsi_qla_host *); | ||
104 | |||
86 | /* | 105 | /* |
87 | * Global Functions in qla_mid.c source file. | 106 | * Global Functions in qla_mid.c source file. |
88 | */ | 107 | */ |
@@ -135,6 +154,7 @@ int qla2x00_marker(struct scsi_qla_host *, struct req_que *, struct rsp_que *, | |||
135 | uint16_t, uint16_t, uint8_t); | 154 | uint16_t, uint16_t, uint8_t); |
136 | int __qla2x00_marker(struct scsi_qla_host *, struct req_que *, struct rsp_que *, | 155 | int __qla2x00_marker(struct scsi_qla_host *, struct req_que *, struct rsp_que *, |
137 | uint16_t, uint16_t, uint8_t); | 156 | uint16_t, uint16_t, uint8_t); |
157 | extern int qla2x00_start_sp(srb_t *); | ||
138 | 158 | ||
139 | /* | 159 | /* |
140 | * Global Function Prototypes in qla_mbx.c source file. | 160 | * Global Function Prototypes in qla_mbx.c source file. |
diff --git a/drivers/scsi/qla2xxx/qla_gs.c b/drivers/scsi/qla2xxx/qla_gs.c index 917534b9f221..4647015eba63 100644 --- a/drivers/scsi/qla2xxx/qla_gs.c +++ b/drivers/scsi/qla2xxx/qla_gs.c | |||
@@ -1674,6 +1674,10 @@ int | |||
1674 | qla2x00_fdmi_register(scsi_qla_host_t *vha) | 1674 | qla2x00_fdmi_register(scsi_qla_host_t *vha) |
1675 | { | 1675 | { |
1676 | int rval; | 1676 | int rval; |
1677 | struct qla_hw_data *ha = vha->hw; | ||
1678 | |||
1679 | if (IS_QLA2100(ha) || IS_QLA2200(ha)) | ||
1680 | return QLA_FUNCTION_FAILED; | ||
1677 | 1681 | ||
1678 | rval = qla2x00_mgmt_svr_login(vha); | 1682 | rval = qla2x00_mgmt_svr_login(vha); |
1679 | if (rval) | 1683 | if (rval) |
diff --git a/drivers/scsi/qla2xxx/qla_init.c b/drivers/scsi/qla2xxx/qla_init.c index f2ce8e3cc91b..9e3eaac25596 100644 --- a/drivers/scsi/qla2xxx/qla_init.c +++ b/drivers/scsi/qla2xxx/qla_init.c | |||
@@ -40,6 +40,210 @@ static struct qla_chip_state_84xx *qla84xx_get_chip(struct scsi_qla_host *); | |||
40 | static int qla84xx_init_chip(scsi_qla_host_t *); | 40 | static int qla84xx_init_chip(scsi_qla_host_t *); |
41 | static int qla25xx_init_queues(struct qla_hw_data *); | 41 | static int qla25xx_init_queues(struct qla_hw_data *); |
42 | 42 | ||
43 | /* SRB Extensions ---------------------------------------------------------- */ | ||
44 | |||
45 | static void | ||
46 | qla2x00_ctx_sp_timeout(unsigned long __data) | ||
47 | { | ||
48 | srb_t *sp = (srb_t *)__data; | ||
49 | struct srb_ctx *ctx; | ||
50 | fc_port_t *fcport = sp->fcport; | ||
51 | struct qla_hw_data *ha = fcport->vha->hw; | ||
52 | struct req_que *req; | ||
53 | unsigned long flags; | ||
54 | |||
55 | spin_lock_irqsave(&ha->hardware_lock, flags); | ||
56 | req = ha->req_q_map[0]; | ||
57 | req->outstanding_cmds[sp->handle] = NULL; | ||
58 | ctx = sp->ctx; | ||
59 | ctx->timeout(sp); | ||
60 | spin_unlock_irqrestore(&ha->hardware_lock, flags); | ||
61 | |||
62 | ctx->free(sp); | ||
63 | } | ||
64 | |||
65 | static void | ||
66 | qla2x00_ctx_sp_free(srb_t *sp) | ||
67 | { | ||
68 | struct srb_ctx *ctx = sp->ctx; | ||
69 | |||
70 | kfree(ctx); | ||
71 | mempool_free(sp, sp->fcport->vha->hw->srb_mempool); | ||
72 | } | ||
73 | |||
74 | inline srb_t * | ||
75 | qla2x00_get_ctx_sp(scsi_qla_host_t *vha, fc_port_t *fcport, size_t size, | ||
76 | unsigned long tmo) | ||
77 | { | ||
78 | srb_t *sp; | ||
79 | struct qla_hw_data *ha = vha->hw; | ||
80 | struct srb_ctx *ctx; | ||
81 | |||
82 | sp = mempool_alloc(ha->srb_mempool, GFP_KERNEL); | ||
83 | if (!sp) | ||
84 | goto done; | ||
85 | ctx = kzalloc(size, GFP_KERNEL); | ||
86 | if (!ctx) { | ||
87 | mempool_free(sp, ha->srb_mempool); | ||
88 | goto done; | ||
89 | } | ||
90 | |||
91 | memset(sp, 0, sizeof(*sp)); | ||
92 | sp->fcport = fcport; | ||
93 | sp->ctx = ctx; | ||
94 | ctx->free = qla2x00_ctx_sp_free; | ||
95 | |||
96 | init_timer(&ctx->timer); | ||
97 | if (!tmo) | ||
98 | goto done; | ||
99 | ctx->timer.expires = jiffies + tmo * HZ; | ||
100 | ctx->timer.data = (unsigned long)sp; | ||
101 | ctx->timer.function = qla2x00_ctx_sp_timeout; | ||
102 | add_timer(&ctx->timer); | ||
103 | done: | ||
104 | return sp; | ||
105 | } | ||
106 | |||
107 | /* Asynchronous Login/Logout Routines -------------------------------------- */ | ||
108 | |||
109 | #define ELS_TMO_2_RATOV(ha) ((ha)->r_a_tov / 10 * 2) | ||
110 | |||
111 | static void | ||
112 | qla2x00_async_logio_timeout(srb_t *sp) | ||
113 | { | ||
114 | fc_port_t *fcport = sp->fcport; | ||
115 | struct srb_logio *lio = sp->ctx; | ||
116 | |||
117 | DEBUG2(printk(KERN_WARNING | ||
118 | "scsi(%ld:%x): Async-%s timeout.\n", | ||
119 | fcport->vha->host_no, sp->handle, | ||
120 | lio->ctx.type == SRB_LOGIN_CMD ? "login": "logout")); | ||
121 | |||
122 | if (lio->ctx.type == SRB_LOGIN_CMD) | ||
123 | qla2x00_post_async_logout_work(fcport->vha, fcport, NULL); | ||
124 | } | ||
125 | |||
126 | int | ||
127 | qla2x00_async_login(struct scsi_qla_host *vha, fc_port_t *fcport, | ||
128 | uint16_t *data) | ||
129 | { | ||
130 | struct qla_hw_data *ha = vha->hw; | ||
131 | srb_t *sp; | ||
132 | struct srb_logio *lio; | ||
133 | int rval; | ||
134 | |||
135 | rval = QLA_FUNCTION_FAILED; | ||
136 | sp = qla2x00_get_ctx_sp(vha, fcport, sizeof(struct srb_logio), | ||
137 | ELS_TMO_2_RATOV(ha) + 2); | ||
138 | if (!sp) | ||
139 | goto done; | ||
140 | |||
141 | lio = sp->ctx; | ||
142 | lio->ctx.type = SRB_LOGIN_CMD; | ||
143 | lio->ctx.timeout = qla2x00_async_logio_timeout; | ||
144 | lio->flags |= SRB_LOGIN_COND_PLOGI; | ||
145 | if (data[1] & QLA_LOGIO_LOGIN_RETRIED) | ||
146 | lio->flags |= SRB_LOGIN_RETRIED; | ||
147 | rval = qla2x00_start_sp(sp); | ||
148 | if (rval != QLA_SUCCESS) | ||
149 | goto done_free_sp; | ||
150 | |||
151 | DEBUG2(printk(KERN_DEBUG | ||
152 | "scsi(%ld:%x): Async-login - loop-id=%x portid=%02x%02x%02x " | ||
153 | "retries=%d.\n", fcport->vha->host_no, sp->handle, fcport->loop_id, | ||
154 | fcport->d_id.b.domain, fcport->d_id.b.area, fcport->d_id.b.al_pa, | ||
155 | fcport->login_retry)); | ||
156 | return rval; | ||
157 | |||
158 | done_free_sp: | ||
159 | del_timer_sync(&lio->ctx.timer); | ||
160 | lio->ctx.free(sp); | ||
161 | done: | ||
162 | return rval; | ||
163 | } | ||
164 | |||
165 | int | ||
166 | qla2x00_async_logout(struct scsi_qla_host *vha, fc_port_t *fcport) | ||
167 | { | ||
168 | struct qla_hw_data *ha = vha->hw; | ||
169 | srb_t *sp; | ||
170 | struct srb_logio *lio; | ||
171 | int rval; | ||
172 | |||
173 | rval = QLA_FUNCTION_FAILED; | ||
174 | sp = qla2x00_get_ctx_sp(vha, fcport, sizeof(struct srb_logio), | ||
175 | ELS_TMO_2_RATOV(ha) + 2); | ||
176 | if (!sp) | ||
177 | goto done; | ||
178 | |||
179 | lio = sp->ctx; | ||
180 | lio->ctx.type = SRB_LOGOUT_CMD; | ||
181 | lio->ctx.timeout = qla2x00_async_logio_timeout; | ||
182 | rval = qla2x00_start_sp(sp); | ||
183 | if (rval != QLA_SUCCESS) | ||
184 | goto done_free_sp; | ||
185 | |||
186 | DEBUG2(printk(KERN_DEBUG | ||
187 | "scsi(%ld:%x): Async-logout - loop-id=%x portid=%02x%02x%02x.\n", | ||
188 | fcport->vha->host_no, sp->handle, fcport->loop_id, | ||
189 | fcport->d_id.b.domain, fcport->d_id.b.area, fcport->d_id.b.al_pa)); | ||
190 | return rval; | ||
191 | |||
192 | done_free_sp: | ||
193 | del_timer_sync(&lio->ctx.timer); | ||
194 | lio->ctx.free(sp); | ||
195 | done: | ||
196 | return rval; | ||
197 | } | ||
198 | |||
199 | int | ||
200 | qla2x00_async_login_done(struct scsi_qla_host *vha, fc_port_t *fcport, | ||
201 | uint16_t *data) | ||
202 | { | ||
203 | int rval; | ||
204 | uint8_t opts = 0; | ||
205 | |||
206 | switch (data[0]) { | ||
207 | case MBS_COMMAND_COMPLETE: | ||
208 | if (fcport->flags & FCF_TAPE_PRESENT) | ||
209 | opts |= BIT_1; | ||
210 | rval = qla2x00_get_port_database(vha, fcport, opts); | ||
211 | if (rval != QLA_SUCCESS) | ||
212 | qla2x00_mark_device_lost(vha, fcport, 1, 0); | ||
213 | else | ||
214 | qla2x00_update_fcport(vha, fcport); | ||
215 | break; | ||
216 | case MBS_COMMAND_ERROR: | ||
217 | if (data[1] & QLA_LOGIO_LOGIN_RETRIED) | ||
218 | set_bit(RELOGIN_NEEDED, &vha->dpc_flags); | ||
219 | else | ||
220 | qla2x00_mark_device_lost(vha, fcport, 1, 0); | ||
221 | break; | ||
222 | case MBS_PORT_ID_USED: | ||
223 | fcport->loop_id = data[1]; | ||
224 | qla2x00_post_async_login_work(vha, fcport, NULL); | ||
225 | break; | ||
226 | case MBS_LOOP_ID_USED: | ||
227 | fcport->loop_id++; | ||
228 | rval = qla2x00_find_new_loop_id(vha, fcport); | ||
229 | if (rval != QLA_SUCCESS) { | ||
230 | qla2x00_mark_device_lost(vha, fcport, 1, 0); | ||
231 | break; | ||
232 | } | ||
233 | qla2x00_post_async_login_work(vha, fcport, NULL); | ||
234 | break; | ||
235 | } | ||
236 | return QLA_SUCCESS; | ||
237 | } | ||
238 | |||
239 | int | ||
240 | qla2x00_async_logout_done(struct scsi_qla_host *vha, fc_port_t *fcport, | ||
241 | uint16_t *data) | ||
242 | { | ||
243 | qla2x00_mark_device_lost(vha, fcport, 1, 0); | ||
244 | return QLA_SUCCESS; | ||
245 | } | ||
246 | |||
43 | /****************************************************************************/ | 247 | /****************************************************************************/ |
44 | /* QLogic ISP2x00 Hardware Support Functions. */ | 248 | /* QLogic ISP2x00 Hardware Support Functions. */ |
45 | /****************************************************************************/ | 249 | /****************************************************************************/ |
@@ -987,7 +1191,6 @@ qla2x00_setup_chip(scsi_qla_host_t *vha) | |||
987 | ha->phy_version); | 1191 | ha->phy_version); |
988 | if (rval != QLA_SUCCESS) | 1192 | if (rval != QLA_SUCCESS) |
989 | goto failed; | 1193 | goto failed; |
990 | |||
991 | ha->flags.npiv_supported = 0; | 1194 | ha->flags.npiv_supported = 0; |
992 | if (IS_QLA2XXX_MIDTYPE(ha) && | 1195 | if (IS_QLA2XXX_MIDTYPE(ha) && |
993 | (ha->fw_attributes & BIT_2)) { | 1196 | (ha->fw_attributes & BIT_2)) { |
@@ -1591,7 +1794,8 @@ qla2x00_set_model_info(scsi_qla_host_t *vha, uint8_t *model, size_t len, | |||
1591 | char *st, *en; | 1794 | char *st, *en; |
1592 | uint16_t index; | 1795 | uint16_t index; |
1593 | struct qla_hw_data *ha = vha->hw; | 1796 | struct qla_hw_data *ha = vha->hw; |
1594 | int use_tbl = !IS_QLA25XX(ha) && !IS_QLA81XX(ha); | 1797 | int use_tbl = !IS_QLA24XX_TYPE(ha) && !IS_QLA25XX(ha) && |
1798 | !IS_QLA81XX(ha); | ||
1595 | 1799 | ||
1596 | if (memcmp(model, BINZERO, len) != 0) { | 1800 | if (memcmp(model, BINZERO, len) != 0) { |
1597 | strncpy(ha->model_number, model, len); | 1801 | strncpy(ha->model_number, model, len); |
@@ -1978,7 +2182,7 @@ qla2x00_rport_del(void *data) | |||
1978 | struct fc_rport *rport; | 2182 | struct fc_rport *rport; |
1979 | 2183 | ||
1980 | spin_lock_irq(fcport->vha->host->host_lock); | 2184 | spin_lock_irq(fcport->vha->host->host_lock); |
1981 | rport = fcport->drport; | 2185 | rport = fcport->drport ? fcport->drport: fcport->rport; |
1982 | fcport->drport = NULL; | 2186 | fcport->drport = NULL; |
1983 | spin_unlock_irq(fcport->vha->host->host_lock); | 2187 | spin_unlock_irq(fcport->vha->host->host_lock); |
1984 | if (rport) | 2188 | if (rport) |
@@ -2345,8 +2549,7 @@ qla2x00_reg_remote_port(scsi_qla_host_t *vha, fc_port_t *fcport) | |||
2345 | struct fc_rport *rport; | 2549 | struct fc_rport *rport; |
2346 | struct qla_hw_data *ha = vha->hw; | 2550 | struct qla_hw_data *ha = vha->hw; |
2347 | 2551 | ||
2348 | if (fcport->drport) | 2552 | qla2x00_rport_del(fcport); |
2349 | qla2x00_rport_del(fcport); | ||
2350 | 2553 | ||
2351 | rport_ids.node_name = wwn_to_u64(fcport->node_name); | 2554 | rport_ids.node_name = wwn_to_u64(fcport->node_name); |
2352 | rport_ids.port_name = wwn_to_u64(fcport->port_name); | 2555 | rport_ids.port_name = wwn_to_u64(fcport->port_name); |
@@ -3039,6 +3242,12 @@ qla2x00_fabric_dev_login(scsi_qla_host_t *vha, fc_port_t *fcport, | |||
3039 | rval = QLA_SUCCESS; | 3242 | rval = QLA_SUCCESS; |
3040 | retry = 0; | 3243 | retry = 0; |
3041 | 3244 | ||
3245 | if (IS_ALOGIO_CAPABLE(ha)) { | ||
3246 | rval = qla2x00_post_async_login_work(vha, fcport, NULL); | ||
3247 | if (!rval) | ||
3248 | return rval; | ||
3249 | } | ||
3250 | |||
3042 | rval = qla2x00_fabric_login(vha, fcport, next_loopid); | 3251 | rval = qla2x00_fabric_login(vha, fcport, next_loopid); |
3043 | if (rval == QLA_SUCCESS) { | 3252 | if (rval == QLA_SUCCESS) { |
3044 | /* Send an ADISC to tape devices.*/ | 3253 | /* Send an ADISC to tape devices.*/ |
@@ -3133,7 +3342,7 @@ qla2x00_fabric_login(scsi_qla_host_t *vha, fc_port_t *fcport, | |||
3133 | } else { | 3342 | } else { |
3134 | fcport->port_type = FCT_TARGET; | 3343 | fcport->port_type = FCT_TARGET; |
3135 | if (mb[1] & BIT_1) { | 3344 | if (mb[1] & BIT_1) { |
3136 | fcport->flags |= FCF_TAPE_PRESENT; | 3345 | fcport->flags |= FCF_FCP2_DEVICE; |
3137 | } | 3346 | } |
3138 | } | 3347 | } |
3139 | 3348 | ||
@@ -3244,7 +3453,7 @@ qla2x00_loop_resync(scsi_qla_host_t *vha) | |||
3244 | struct req_que *req; | 3453 | struct req_que *req; |
3245 | struct rsp_que *rsp; | 3454 | struct rsp_que *rsp; |
3246 | 3455 | ||
3247 | if (ql2xmultique_tag) | 3456 | if (vha->hw->flags.cpu_affinity_enabled) |
3248 | req = vha->hw->req_q_map[0]; | 3457 | req = vha->hw->req_q_map[0]; |
3249 | else | 3458 | else |
3250 | req = vha->req; | 3459 | req = vha->req; |
@@ -3286,15 +3495,17 @@ qla2x00_loop_resync(scsi_qla_host_t *vha) | |||
3286 | } | 3495 | } |
3287 | 3496 | ||
3288 | void | 3497 | void |
3289 | qla2x00_update_fcports(scsi_qla_host_t *vha) | 3498 | qla2x00_update_fcports(scsi_qla_host_t *base_vha) |
3290 | { | 3499 | { |
3291 | fc_port_t *fcport; | 3500 | fc_port_t *fcport; |
3501 | struct scsi_qla_host *tvp, *vha; | ||
3292 | 3502 | ||
3293 | /* Go with deferred removal of rport references. */ | 3503 | /* Go with deferred removal of rport references. */ |
3294 | list_for_each_entry(fcport, &vha->vp_fcports, list) | 3504 | list_for_each_entry_safe(vha, tvp, &base_vha->hw->vp_list, list) |
3295 | if (fcport && fcport->drport && | 3505 | list_for_each_entry(fcport, &vha->vp_fcports, list) |
3296 | atomic_read(&fcport->state) != FCS_UNCONFIGURED) | 3506 | if (fcport && fcport->drport && |
3297 | qla2x00_rport_del(fcport); | 3507 | atomic_read(&fcport->state) != FCS_UNCONFIGURED) |
3508 | qla2x00_rport_del(fcport); | ||
3298 | } | 3509 | } |
3299 | 3510 | ||
3300 | /* | 3511 | /* |
@@ -3331,8 +3542,6 @@ qla2x00_abort_isp(scsi_qla_host_t *vha) | |||
3331 | if (atomic_read(&vha->loop_state) != LOOP_DOWN) { | 3542 | if (atomic_read(&vha->loop_state) != LOOP_DOWN) { |
3332 | atomic_set(&vha->loop_state, LOOP_DOWN); | 3543 | atomic_set(&vha->loop_state, LOOP_DOWN); |
3333 | qla2x00_mark_all_devices_lost(vha, 0); | 3544 | qla2x00_mark_all_devices_lost(vha, 0); |
3334 | list_for_each_entry_safe(vp, tvp, &ha->vp_list, list) | ||
3335 | qla2x00_mark_all_devices_lost(vp, 0); | ||
3336 | } else { | 3545 | } else { |
3337 | if (!atomic_read(&vha->loop_down_timer)) | 3546 | if (!atomic_read(&vha->loop_down_timer)) |
3338 | atomic_set(&vha->loop_down_timer, | 3547 | atomic_set(&vha->loop_down_timer, |
@@ -4264,7 +4473,7 @@ qla24xx_configure_vhba(scsi_qla_host_t *vha) | |||
4264 | return -EINVAL; | 4473 | return -EINVAL; |
4265 | 4474 | ||
4266 | rval = qla2x00_fw_ready(base_vha); | 4475 | rval = qla2x00_fw_ready(base_vha); |
4267 | if (ql2xmultique_tag) | 4476 | if (ha->flags.cpu_affinity_enabled) |
4268 | req = ha->req_q_map[0]; | 4477 | req = ha->req_q_map[0]; |
4269 | else | 4478 | else |
4270 | req = vha->req; | 4479 | req = vha->req; |
diff --git a/drivers/scsi/qla2xxx/qla_iocb.c b/drivers/scsi/qla2xxx/qla_iocb.c index 13396beae2ce..c5ccac0bef76 100644 --- a/drivers/scsi/qla2xxx/qla_iocb.c +++ b/drivers/scsi/qla2xxx/qla_iocb.c | |||
@@ -350,6 +350,7 @@ qla2x00_start_scsi(srb_t *sp) | |||
350 | /* Build command packet */ | 350 | /* Build command packet */ |
351 | req->current_outstanding_cmd = handle; | 351 | req->current_outstanding_cmd = handle; |
352 | req->outstanding_cmds[handle] = sp; | 352 | req->outstanding_cmds[handle] = sp; |
353 | sp->handle = handle; | ||
353 | sp->cmd->host_scribble = (unsigned char *)(unsigned long)handle; | 354 | sp->cmd->host_scribble = (unsigned char *)(unsigned long)handle; |
354 | req->cnt -= req_cnt; | 355 | req->cnt -= req_cnt; |
355 | 356 | ||
@@ -778,6 +779,7 @@ qla24xx_start_scsi(srb_t *sp) | |||
778 | /* Build command packet. */ | 779 | /* Build command packet. */ |
779 | req->current_outstanding_cmd = handle; | 780 | req->current_outstanding_cmd = handle; |
780 | req->outstanding_cmds[handle] = sp; | 781 | req->outstanding_cmds[handle] = sp; |
782 | sp->handle = handle; | ||
781 | sp->cmd->host_scribble = (unsigned char *)(unsigned long)handle; | 783 | sp->cmd->host_scribble = (unsigned char *)(unsigned long)handle; |
782 | req->cnt -= req_cnt; | 784 | req->cnt -= req_cnt; |
783 | 785 | ||
@@ -852,9 +854,211 @@ static void qla25xx_set_que(srb_t *sp, struct rsp_que **rsp) | |||
852 | struct qla_hw_data *ha = sp->fcport->vha->hw; | 854 | struct qla_hw_data *ha = sp->fcport->vha->hw; |
853 | int affinity = cmd->request->cpu; | 855 | int affinity = cmd->request->cpu; |
854 | 856 | ||
855 | if (ql2xmultique_tag && affinity >= 0 && | 857 | if (ha->flags.cpu_affinity_enabled && affinity >= 0 && |
856 | affinity < ha->max_rsp_queues - 1) | 858 | affinity < ha->max_rsp_queues - 1) |
857 | *rsp = ha->rsp_q_map[affinity + 1]; | 859 | *rsp = ha->rsp_q_map[affinity + 1]; |
858 | else | 860 | else |
859 | *rsp = ha->rsp_q_map[0]; | 861 | *rsp = ha->rsp_q_map[0]; |
860 | } | 862 | } |
863 | |||
864 | /* Generic Control-SRB manipulation functions. */ | ||
865 | |||
866 | static void * | ||
867 | qla2x00_alloc_iocbs(srb_t *sp) | ||
868 | { | ||
869 | scsi_qla_host_t *vha = sp->fcport->vha; | ||
870 | struct qla_hw_data *ha = vha->hw; | ||
871 | struct req_que *req = ha->req_q_map[0]; | ||
872 | device_reg_t __iomem *reg = ISP_QUE_REG(ha, req->id); | ||
873 | uint32_t index, handle; | ||
874 | request_t *pkt; | ||
875 | uint16_t cnt, req_cnt; | ||
876 | |||
877 | pkt = NULL; | ||
878 | req_cnt = 1; | ||
879 | |||
880 | /* Check for room in outstanding command list. */ | ||
881 | handle = req->current_outstanding_cmd; | ||
882 | for (index = 1; index < MAX_OUTSTANDING_COMMANDS; index++) { | ||
883 | handle++; | ||
884 | if (handle == MAX_OUTSTANDING_COMMANDS) | ||
885 | handle = 1; | ||
886 | if (!req->outstanding_cmds[handle]) | ||
887 | break; | ||
888 | } | ||
889 | if (index == MAX_OUTSTANDING_COMMANDS) | ||
890 | goto queuing_error; | ||
891 | |||
892 | /* Check for room on request queue. */ | ||
893 | if (req->cnt < req_cnt) { | ||
894 | if (ha->mqenable) | ||
895 | cnt = RD_REG_DWORD(®->isp25mq.req_q_out); | ||
896 | else if (IS_FWI2_CAPABLE(ha)) | ||
897 | cnt = RD_REG_DWORD(®->isp24.req_q_out); | ||
898 | else | ||
899 | cnt = qla2x00_debounce_register( | ||
900 | ISP_REQ_Q_OUT(ha, ®->isp)); | ||
901 | |||
902 | if (req->ring_index < cnt) | ||
903 | req->cnt = cnt - req->ring_index; | ||
904 | else | ||
905 | req->cnt = req->length - | ||
906 | (req->ring_index - cnt); | ||
907 | } | ||
908 | if (req->cnt < req_cnt) | ||
909 | goto queuing_error; | ||
910 | |||
911 | /* Prep packet */ | ||
912 | req->current_outstanding_cmd = handle; | ||
913 | req->outstanding_cmds[handle] = sp; | ||
914 | req->cnt -= req_cnt; | ||
915 | |||
916 | pkt = req->ring_ptr; | ||
917 | memset(pkt, 0, REQUEST_ENTRY_SIZE); | ||
918 | pkt->entry_count = req_cnt; | ||
919 | pkt->handle = handle; | ||
920 | sp->handle = handle; | ||
921 | |||
922 | queuing_error: | ||
923 | return pkt; | ||
924 | } | ||
925 | |||
926 | static void | ||
927 | qla2x00_start_iocbs(srb_t *sp) | ||
928 | { | ||
929 | struct qla_hw_data *ha = sp->fcport->vha->hw; | ||
930 | struct req_que *req = ha->req_q_map[0]; | ||
931 | device_reg_t __iomem *reg = ISP_QUE_REG(ha, req->id); | ||
932 | struct device_reg_2xxx __iomem *ioreg = &ha->iobase->isp; | ||
933 | |||
934 | /* Adjust ring index. */ | ||
935 | req->ring_index++; | ||
936 | if (req->ring_index == req->length) { | ||
937 | req->ring_index = 0; | ||
938 | req->ring_ptr = req->ring; | ||
939 | } else | ||
940 | req->ring_ptr++; | ||
941 | |||
942 | /* Set chip new ring index. */ | ||
943 | if (ha->mqenable) { | ||
944 | WRT_REG_DWORD(®->isp25mq.req_q_in, req->ring_index); | ||
945 | RD_REG_DWORD(&ioreg->hccr); | ||
946 | } else if (IS_FWI2_CAPABLE(ha)) { | ||
947 | WRT_REG_DWORD(®->isp24.req_q_in, req->ring_index); | ||
948 | RD_REG_DWORD_RELAXED(®->isp24.req_q_in); | ||
949 | } else { | ||
950 | WRT_REG_WORD(ISP_REQ_Q_IN(ha, ®->isp), req->ring_index); | ||
951 | RD_REG_WORD_RELAXED(ISP_REQ_Q_IN(ha, ®->isp)); | ||
952 | } | ||
953 | } | ||
954 | |||
955 | static void | ||
956 | qla24xx_login_iocb(srb_t *sp, struct logio_entry_24xx *logio) | ||
957 | { | ||
958 | struct srb_logio *lio = sp->ctx; | ||
959 | |||
960 | logio->entry_type = LOGINOUT_PORT_IOCB_TYPE; | ||
961 | logio->control_flags = cpu_to_le16(LCF_COMMAND_PLOGI); | ||
962 | if (lio->flags & SRB_LOGIN_COND_PLOGI) | ||
963 | logio->control_flags |= cpu_to_le16(LCF_COND_PLOGI); | ||
964 | if (lio->flags & SRB_LOGIN_SKIP_PRLI) | ||
965 | logio->control_flags |= cpu_to_le16(LCF_SKIP_PRLI); | ||
966 | logio->nport_handle = cpu_to_le16(sp->fcport->loop_id); | ||
967 | logio->port_id[0] = sp->fcport->d_id.b.al_pa; | ||
968 | logio->port_id[1] = sp->fcport->d_id.b.area; | ||
969 | logio->port_id[2] = sp->fcport->d_id.b.domain; | ||
970 | logio->vp_index = sp->fcport->vp_idx; | ||
971 | } | ||
972 | |||
973 | static void | ||
974 | qla2x00_login_iocb(srb_t *sp, struct mbx_entry *mbx) | ||
975 | { | ||
976 | struct qla_hw_data *ha = sp->fcport->vha->hw; | ||
977 | struct srb_logio *lio = sp->ctx; | ||
978 | uint16_t opts; | ||
979 | |||
980 | mbx->entry_type = MBX_IOCB_TYPE;; | ||
981 | SET_TARGET_ID(ha, mbx->loop_id, sp->fcport->loop_id); | ||
982 | mbx->mb0 = cpu_to_le16(MBC_LOGIN_FABRIC_PORT); | ||
983 | opts = lio->flags & SRB_LOGIN_COND_PLOGI ? BIT_0: 0; | ||
984 | opts |= lio->flags & SRB_LOGIN_SKIP_PRLI ? BIT_1: 0; | ||
985 | if (HAS_EXTENDED_IDS(ha)) { | ||
986 | mbx->mb1 = cpu_to_le16(sp->fcport->loop_id); | ||
987 | mbx->mb10 = cpu_to_le16(opts); | ||
988 | } else { | ||
989 | mbx->mb1 = cpu_to_le16((sp->fcport->loop_id << 8) | opts); | ||
990 | } | ||
991 | mbx->mb2 = cpu_to_le16(sp->fcport->d_id.b.domain); | ||
992 | mbx->mb3 = cpu_to_le16(sp->fcport->d_id.b.area << 8 | | ||
993 | sp->fcport->d_id.b.al_pa); | ||
994 | mbx->mb9 = cpu_to_le16(sp->fcport->vp_idx); | ||
995 | } | ||
996 | |||
997 | static void | ||
998 | qla24xx_logout_iocb(srb_t *sp, struct logio_entry_24xx *logio) | ||
999 | { | ||
1000 | logio->entry_type = LOGINOUT_PORT_IOCB_TYPE; | ||
1001 | logio->control_flags = | ||
1002 | cpu_to_le16(LCF_COMMAND_LOGO|LCF_IMPL_LOGO); | ||
1003 | logio->nport_handle = cpu_to_le16(sp->fcport->loop_id); | ||
1004 | logio->port_id[0] = sp->fcport->d_id.b.al_pa; | ||
1005 | logio->port_id[1] = sp->fcport->d_id.b.area; | ||
1006 | logio->port_id[2] = sp->fcport->d_id.b.domain; | ||
1007 | logio->vp_index = sp->fcport->vp_idx; | ||
1008 | } | ||
1009 | |||
1010 | static void | ||
1011 | qla2x00_logout_iocb(srb_t *sp, struct mbx_entry *mbx) | ||
1012 | { | ||
1013 | struct qla_hw_data *ha = sp->fcport->vha->hw; | ||
1014 | |||
1015 | mbx->entry_type = MBX_IOCB_TYPE;; | ||
1016 | SET_TARGET_ID(ha, mbx->loop_id, sp->fcport->loop_id); | ||
1017 | mbx->mb0 = cpu_to_le16(MBC_LOGOUT_FABRIC_PORT); | ||
1018 | mbx->mb1 = HAS_EXTENDED_IDS(ha) ? | ||
1019 | cpu_to_le16(sp->fcport->loop_id): | ||
1020 | cpu_to_le16(sp->fcport->loop_id << 8); | ||
1021 | mbx->mb2 = cpu_to_le16(sp->fcport->d_id.b.domain); | ||
1022 | mbx->mb3 = cpu_to_le16(sp->fcport->d_id.b.area << 8 | | ||
1023 | sp->fcport->d_id.b.al_pa); | ||
1024 | mbx->mb9 = cpu_to_le16(sp->fcport->vp_idx); | ||
1025 | /* Implicit: mbx->mbx10 = 0. */ | ||
1026 | } | ||
1027 | |||
1028 | int | ||
1029 | qla2x00_start_sp(srb_t *sp) | ||
1030 | { | ||
1031 | int rval; | ||
1032 | struct qla_hw_data *ha = sp->fcport->vha->hw; | ||
1033 | void *pkt; | ||
1034 | struct srb_ctx *ctx = sp->ctx; | ||
1035 | unsigned long flags; | ||
1036 | |||
1037 | rval = QLA_FUNCTION_FAILED; | ||
1038 | spin_lock_irqsave(&ha->hardware_lock, flags); | ||
1039 | pkt = qla2x00_alloc_iocbs(sp); | ||
1040 | if (!pkt) | ||
1041 | goto done; | ||
1042 | |||
1043 | rval = QLA_SUCCESS; | ||
1044 | switch (ctx->type) { | ||
1045 | case SRB_LOGIN_CMD: | ||
1046 | IS_FWI2_CAPABLE(ha) ? | ||
1047 | qla24xx_login_iocb(sp, pkt): | ||
1048 | qla2x00_login_iocb(sp, pkt); | ||
1049 | break; | ||
1050 | case SRB_LOGOUT_CMD: | ||
1051 | IS_FWI2_CAPABLE(ha) ? | ||
1052 | qla24xx_logout_iocb(sp, pkt): | ||
1053 | qla2x00_logout_iocb(sp, pkt); | ||
1054 | break; | ||
1055 | default: | ||
1056 | break; | ||
1057 | } | ||
1058 | |||
1059 | wmb(); | ||
1060 | qla2x00_start_iocbs(sp); | ||
1061 | done: | ||
1062 | spin_unlock_irqrestore(&ha->hardware_lock, flags); | ||
1063 | return rval; | ||
1064 | } | ||
diff --git a/drivers/scsi/qla2xxx/qla_isr.c b/drivers/scsi/qla2xxx/qla_isr.c index 245e7afb4c4d..b20a7169aac2 100644 --- a/drivers/scsi/qla2xxx/qla_isr.c +++ b/drivers/scsi/qla2xxx/qla_isr.c | |||
@@ -598,9 +598,54 @@ skip_rio: | |||
598 | break; | 598 | break; |
599 | 599 | ||
600 | case MBA_PORT_UPDATE: /* Port database update */ | 600 | case MBA_PORT_UPDATE: /* Port database update */ |
601 | /* Only handle SCNs for our Vport index. */ | 601 | /* |
602 | if (vha->vp_idx && vha->vp_idx != (mb[3] & 0xff)) | 602 | * Handle only global and vn-port update events |
603 | * | ||
604 | * Relevant inputs: | ||
605 | * mb[1] = N_Port handle of changed port | ||
606 | * OR 0xffff for global event | ||
607 | * mb[2] = New login state | ||
608 | * 7 = Port logged out | ||
609 | * mb[3] = LSB is vp_idx, 0xff = all vps | ||
610 | * | ||
611 | * Skip processing if: | ||
612 | * Event is global, vp_idx is NOT all vps, | ||
613 | * vp_idx does not match | ||
614 | * Event is not global, vp_idx does not match | ||
615 | */ | ||
616 | if ((mb[1] == 0xffff && (mb[3] & 0xff) != 0xff) | ||
617 | || (mb[1] != 0xffff)) { | ||
618 | if (vha->vp_idx != (mb[3] & 0xff)) | ||
619 | break; | ||
620 | } | ||
621 | |||
622 | /* Global event -- port logout or port unavailable. */ | ||
623 | if (mb[1] == 0xffff && mb[2] == 0x7) { | ||
624 | DEBUG2(printk("scsi(%ld): Asynchronous PORT UPDATE.\n", | ||
625 | vha->host_no)); | ||
626 | DEBUG(printk(KERN_INFO | ||
627 | "scsi(%ld): Port unavailable %04x %04x %04x.\n", | ||
628 | vha->host_no, mb[1], mb[2], mb[3])); | ||
629 | |||
630 | if (atomic_read(&vha->loop_state) != LOOP_DOWN) { | ||
631 | atomic_set(&vha->loop_state, LOOP_DOWN); | ||
632 | atomic_set(&vha->loop_down_timer, | ||
633 | LOOP_DOWN_TIME); | ||
634 | vha->device_flags |= DFLG_NO_CABLE; | ||
635 | qla2x00_mark_all_devices_lost(vha, 1); | ||
636 | } | ||
637 | |||
638 | if (vha->vp_idx) { | ||
639 | atomic_set(&vha->vp_state, VP_FAILED); | ||
640 | fc_vport_set_state(vha->fc_vport, | ||
641 | FC_VPORT_FAILED); | ||
642 | qla2x00_mark_all_devices_lost(vha, 1); | ||
643 | } | ||
644 | |||
645 | vha->flags.management_server_logged_in = 0; | ||
646 | ha->link_data_rate = PORT_SPEED_UNKNOWN; | ||
603 | break; | 647 | break; |
648 | } | ||
604 | 649 | ||
605 | /* | 650 | /* |
606 | * If PORT UPDATE is global (received LIP_OCCURRED/LIP_RESET | 651 | * If PORT UPDATE is global (received LIP_OCCURRED/LIP_RESET |
@@ -640,8 +685,9 @@ skip_rio: | |||
640 | if (vha->vp_idx && test_bit(VP_SCR_NEEDED, &vha->vp_flags)) | 685 | if (vha->vp_idx && test_bit(VP_SCR_NEEDED, &vha->vp_flags)) |
641 | break; | 686 | break; |
642 | /* Only handle SCNs for our Vport index. */ | 687 | /* Only handle SCNs for our Vport index. */ |
643 | if (vha->vp_idx && vha->vp_idx != (mb[3] & 0xff)) | 688 | if (ha->flags.npiv_supported && vha->vp_idx != (mb[3] & 0xff)) |
644 | break; | 689 | break; |
690 | |||
645 | DEBUG2(printk("scsi(%ld): Asynchronous RSCR UPDATE.\n", | 691 | DEBUG2(printk("scsi(%ld): Asynchronous RSCR UPDATE.\n", |
646 | vha->host_no)); | 692 | vha->host_no)); |
647 | DEBUG(printk(KERN_INFO | 693 | DEBUG(printk(KERN_INFO |
@@ -874,6 +920,249 @@ qla2x00_process_completed_request(struct scsi_qla_host *vha, | |||
874 | } | 920 | } |
875 | } | 921 | } |
876 | 922 | ||
923 | static srb_t * | ||
924 | qla2x00_get_sp_from_handle(scsi_qla_host_t *vha, const char *func, | ||
925 | struct req_que *req, void *iocb) | ||
926 | { | ||
927 | struct qla_hw_data *ha = vha->hw; | ||
928 | sts_entry_t *pkt = iocb; | ||
929 | srb_t *sp = NULL; | ||
930 | uint16_t index; | ||
931 | |||
932 | index = LSW(pkt->handle); | ||
933 | if (index >= MAX_OUTSTANDING_COMMANDS) { | ||
934 | qla_printk(KERN_WARNING, ha, | ||
935 | "%s: Invalid completion handle (%x).\n", func, index); | ||
936 | set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); | ||
937 | goto done; | ||
938 | } | ||
939 | sp = req->outstanding_cmds[index]; | ||
940 | if (!sp) { | ||
941 | qla_printk(KERN_WARNING, ha, | ||
942 | "%s: Invalid completion handle (%x) -- timed-out.\n", func, | ||
943 | index); | ||
944 | return sp; | ||
945 | } | ||
946 | if (sp->handle != index) { | ||
947 | qla_printk(KERN_WARNING, ha, | ||
948 | "%s: SRB handle (%x) mismatch %x.\n", func, sp->handle, | ||
949 | index); | ||
950 | return NULL; | ||
951 | } | ||
952 | req->outstanding_cmds[index] = NULL; | ||
953 | done: | ||
954 | return sp; | ||
955 | } | ||
956 | |||
957 | static void | ||
958 | qla2x00_mbx_iocb_entry(scsi_qla_host_t *vha, struct req_que *req, | ||
959 | struct mbx_entry *mbx) | ||
960 | { | ||
961 | const char func[] = "MBX-IOCB"; | ||
962 | const char *type; | ||
963 | struct qla_hw_data *ha = vha->hw; | ||
964 | fc_port_t *fcport; | ||
965 | srb_t *sp; | ||
966 | struct srb_logio *lio; | ||
967 | uint16_t data[2]; | ||
968 | |||
969 | sp = qla2x00_get_sp_from_handle(vha, func, req, mbx); | ||
970 | if (!sp) | ||
971 | return; | ||
972 | |||
973 | type = NULL; | ||
974 | lio = sp->ctx; | ||
975 | switch (lio->ctx.type) { | ||
976 | case SRB_LOGIN_CMD: | ||
977 | type = "login"; | ||
978 | break; | ||
979 | case SRB_LOGOUT_CMD: | ||
980 | type = "logout"; | ||
981 | break; | ||
982 | default: | ||
983 | qla_printk(KERN_WARNING, ha, | ||
984 | "%s: Unrecognized SRB: (%p) type=%d.\n", func, sp, | ||
985 | lio->ctx.type); | ||
986 | return; | ||
987 | } | ||
988 | |||
989 | del_timer(&lio->ctx.timer); | ||
990 | fcport = sp->fcport; | ||
991 | |||
992 | data[0] = data[1] = 0; | ||
993 | if (mbx->entry_status) { | ||
994 | DEBUG2(printk(KERN_WARNING | ||
995 | "scsi(%ld:%x): Async-%s error entry - entry-status=%x " | ||
996 | "status=%x state-flag=%x status-flags=%x.\n", | ||
997 | fcport->vha->host_no, sp->handle, type, | ||
998 | mbx->entry_status, le16_to_cpu(mbx->status), | ||
999 | le16_to_cpu(mbx->state_flags), | ||
1000 | le16_to_cpu(mbx->status_flags))); | ||
1001 | DEBUG2(qla2x00_dump_buffer((uint8_t *)mbx, sizeof(*mbx))); | ||
1002 | |||
1003 | data[0] = MBS_COMMAND_ERROR; | ||
1004 | data[1] = lio->flags & SRB_LOGIN_RETRIED ? | ||
1005 | QLA_LOGIO_LOGIN_RETRIED: 0; | ||
1006 | goto done_post_logio_done_work; | ||
1007 | } | ||
1008 | |||
1009 | if (!mbx->status && le16_to_cpu(mbx->mb0) == MBS_COMMAND_COMPLETE) { | ||
1010 | DEBUG2(printk(KERN_DEBUG | ||
1011 | "scsi(%ld:%x): Async-%s complete - mbx1=%x.\n", | ||
1012 | fcport->vha->host_no, sp->handle, type, | ||
1013 | le16_to_cpu(mbx->mb1))); | ||
1014 | |||
1015 | data[0] = MBS_COMMAND_COMPLETE; | ||
1016 | if (lio->ctx.type == SRB_LOGIN_CMD && le16_to_cpu(mbx->mb1) & BIT_1) | ||
1017 | fcport->flags |= FCF_FCP2_DEVICE; | ||
1018 | |||
1019 | goto done_post_logio_done_work; | ||
1020 | } | ||
1021 | |||
1022 | data[0] = le16_to_cpu(mbx->mb0); | ||
1023 | switch (data[0]) { | ||
1024 | case MBS_PORT_ID_USED: | ||
1025 | data[1] = le16_to_cpu(mbx->mb1); | ||
1026 | break; | ||
1027 | case MBS_LOOP_ID_USED: | ||
1028 | break; | ||
1029 | default: | ||
1030 | data[0] = MBS_COMMAND_ERROR; | ||
1031 | data[1] = lio->flags & SRB_LOGIN_RETRIED ? | ||
1032 | QLA_LOGIO_LOGIN_RETRIED: 0; | ||
1033 | break; | ||
1034 | } | ||
1035 | |||
1036 | DEBUG2(printk(KERN_WARNING | ||
1037 | "scsi(%ld:%x): Async-%s failed - status=%x mb0=%x mb1=%x mb2=%x " | ||
1038 | "mb6=%x mb7=%x.\n", | ||
1039 | fcport->vha->host_no, sp->handle, type, le16_to_cpu(mbx->status), | ||
1040 | le16_to_cpu(mbx->mb0), le16_to_cpu(mbx->mb1), | ||
1041 | le16_to_cpu(mbx->mb2), le16_to_cpu(mbx->mb6), | ||
1042 | le16_to_cpu(mbx->mb7))); | ||
1043 | |||
1044 | done_post_logio_done_work: | ||
1045 | lio->ctx.type == SRB_LOGIN_CMD ? | ||
1046 | qla2x00_post_async_login_done_work(fcport->vha, fcport, data): | ||
1047 | qla2x00_post_async_logout_done_work(fcport->vha, fcport, data); | ||
1048 | |||
1049 | lio->ctx.free(sp); | ||
1050 | } | ||
1051 | |||
1052 | static void | ||
1053 | qla24xx_logio_entry(scsi_qla_host_t *vha, struct req_que *req, | ||
1054 | struct logio_entry_24xx *logio) | ||
1055 | { | ||
1056 | const char func[] = "LOGIO-IOCB"; | ||
1057 | const char *type; | ||
1058 | struct qla_hw_data *ha = vha->hw; | ||
1059 | fc_port_t *fcport; | ||
1060 | srb_t *sp; | ||
1061 | struct srb_logio *lio; | ||
1062 | uint16_t data[2]; | ||
1063 | uint32_t iop[2]; | ||
1064 | |||
1065 | sp = qla2x00_get_sp_from_handle(vha, func, req, logio); | ||
1066 | if (!sp) | ||
1067 | return; | ||
1068 | |||
1069 | type = NULL; | ||
1070 | lio = sp->ctx; | ||
1071 | switch (lio->ctx.type) { | ||
1072 | case SRB_LOGIN_CMD: | ||
1073 | type = "login"; | ||
1074 | break; | ||
1075 | case SRB_LOGOUT_CMD: | ||
1076 | type = "logout"; | ||
1077 | break; | ||
1078 | default: | ||
1079 | qla_printk(KERN_WARNING, ha, | ||
1080 | "%s: Unrecognized SRB: (%p) type=%d.\n", func, sp, | ||
1081 | lio->ctx.type); | ||
1082 | return; | ||
1083 | } | ||
1084 | |||
1085 | del_timer(&lio->ctx.timer); | ||
1086 | fcport = sp->fcport; | ||
1087 | |||
1088 | data[0] = data[1] = 0; | ||
1089 | if (logio->entry_status) { | ||
1090 | DEBUG2(printk(KERN_WARNING | ||
1091 | "scsi(%ld:%x): Async-%s error entry - entry-status=%x.\n", | ||
1092 | fcport->vha->host_no, sp->handle, type, | ||
1093 | logio->entry_status)); | ||
1094 | DEBUG2(qla2x00_dump_buffer((uint8_t *)logio, sizeof(*logio))); | ||
1095 | |||
1096 | data[0] = MBS_COMMAND_ERROR; | ||
1097 | data[1] = lio->flags & SRB_LOGIN_RETRIED ? | ||
1098 | QLA_LOGIO_LOGIN_RETRIED: 0; | ||
1099 | goto done_post_logio_done_work; | ||
1100 | } | ||
1101 | |||
1102 | if (le16_to_cpu(logio->comp_status) == CS_COMPLETE) { | ||
1103 | DEBUG2(printk(KERN_DEBUG | ||
1104 | "scsi(%ld:%x): Async-%s complete - iop0=%x.\n", | ||
1105 | fcport->vha->host_no, sp->handle, type, | ||
1106 | le32_to_cpu(logio->io_parameter[0]))); | ||
1107 | |||
1108 | data[0] = MBS_COMMAND_COMPLETE; | ||
1109 | if (lio->ctx.type == SRB_LOGOUT_CMD) | ||
1110 | goto done_post_logio_done_work; | ||
1111 | |||
1112 | iop[0] = le32_to_cpu(logio->io_parameter[0]); | ||
1113 | if (iop[0] & BIT_4) { | ||
1114 | fcport->port_type = FCT_TARGET; | ||
1115 | if (iop[0] & BIT_8) | ||
1116 | fcport->flags |= FCF_FCP2_DEVICE; | ||
1117 | } | ||
1118 | if (iop[0] & BIT_5) | ||
1119 | fcport->port_type = FCT_INITIATOR; | ||
1120 | if (logio->io_parameter[7] || logio->io_parameter[8]) | ||
1121 | fcport->supported_classes |= FC_COS_CLASS2; | ||
1122 | if (logio->io_parameter[9] || logio->io_parameter[10]) | ||
1123 | fcport->supported_classes |= FC_COS_CLASS3; | ||
1124 | |||
1125 | goto done_post_logio_done_work; | ||
1126 | } | ||
1127 | |||
1128 | iop[0] = le32_to_cpu(logio->io_parameter[0]); | ||
1129 | iop[1] = le32_to_cpu(logio->io_parameter[1]); | ||
1130 | switch (iop[0]) { | ||
1131 | case LSC_SCODE_PORTID_USED: | ||
1132 | data[0] = MBS_PORT_ID_USED; | ||
1133 | data[1] = LSW(iop[1]); | ||
1134 | break; | ||
1135 | case LSC_SCODE_NPORT_USED: | ||
1136 | data[0] = MBS_LOOP_ID_USED; | ||
1137 | break; | ||
1138 | case LSC_SCODE_CMD_FAILED: | ||
1139 | if ((iop[1] & 0xff) == 0x05) { | ||
1140 | data[0] = MBS_NOT_LOGGED_IN; | ||
1141 | break; | ||
1142 | } | ||
1143 | /* Fall through. */ | ||
1144 | default: | ||
1145 | data[0] = MBS_COMMAND_ERROR; | ||
1146 | data[1] = lio->flags & SRB_LOGIN_RETRIED ? | ||
1147 | QLA_LOGIO_LOGIN_RETRIED: 0; | ||
1148 | break; | ||
1149 | } | ||
1150 | |||
1151 | DEBUG2(printk(KERN_WARNING | ||
1152 | "scsi(%ld:%x): Async-%s failed - comp=%x iop0=%x iop1=%x.\n", | ||
1153 | fcport->vha->host_no, sp->handle, type, | ||
1154 | le16_to_cpu(logio->comp_status), | ||
1155 | le32_to_cpu(logio->io_parameter[0]), | ||
1156 | le32_to_cpu(logio->io_parameter[1]))); | ||
1157 | |||
1158 | done_post_logio_done_work: | ||
1159 | lio->ctx.type == SRB_LOGIN_CMD ? | ||
1160 | qla2x00_post_async_login_done_work(fcport->vha, fcport, data): | ||
1161 | qla2x00_post_async_logout_done_work(fcport->vha, fcport, data); | ||
1162 | |||
1163 | lio->ctx.free(sp); | ||
1164 | } | ||
1165 | |||
877 | /** | 1166 | /** |
878 | * qla2x00_process_response_queue() - Process response queue entries. | 1167 | * qla2x00_process_response_queue() - Process response queue entries. |
879 | * @ha: SCSI driver HA context | 1168 | * @ha: SCSI driver HA context |
@@ -935,6 +1224,9 @@ qla2x00_process_response_queue(struct rsp_que *rsp) | |||
935 | case STATUS_CONT_TYPE: | 1224 | case STATUS_CONT_TYPE: |
936 | qla2x00_status_cont_entry(rsp, (sts_cont_entry_t *)pkt); | 1225 | qla2x00_status_cont_entry(rsp, (sts_cont_entry_t *)pkt); |
937 | break; | 1226 | break; |
1227 | case MBX_IOCB_TYPE: | ||
1228 | qla2x00_mbx_iocb_entry(vha, rsp->req, | ||
1229 | (struct mbx_entry *)pkt); | ||
938 | default: | 1230 | default: |
939 | /* Type Not Supported. */ | 1231 | /* Type Not Supported. */ |
940 | DEBUG4(printk(KERN_WARNING | 1232 | DEBUG4(printk(KERN_WARNING |
@@ -1223,6 +1515,7 @@ qla2x00_status_entry(scsi_qla_host_t *vha, struct rsp_que *rsp, void *pkt) | |||
1223 | cp->device->id, cp->device->lun, resid, | 1515 | cp->device->id, cp->device->lun, resid, |
1224 | scsi_bufflen(cp))); | 1516 | scsi_bufflen(cp))); |
1225 | 1517 | ||
1518 | scsi_set_resid(cp, resid); | ||
1226 | cp->result = DID_ERROR << 16; | 1519 | cp->result = DID_ERROR << 16; |
1227 | break; | 1520 | break; |
1228 | } | 1521 | } |
@@ -1544,6 +1837,10 @@ void qla24xx_process_response_queue(struct scsi_qla_host *vha, | |||
1544 | qla24xx_report_id_acquisition(vha, | 1837 | qla24xx_report_id_acquisition(vha, |
1545 | (struct vp_rpt_id_entry_24xx *)pkt); | 1838 | (struct vp_rpt_id_entry_24xx *)pkt); |
1546 | break; | 1839 | break; |
1840 | case LOGINOUT_PORT_IOCB_TYPE: | ||
1841 | qla24xx_logio_entry(vha, rsp->req, | ||
1842 | (struct logio_entry_24xx *)pkt); | ||
1843 | break; | ||
1547 | default: | 1844 | default: |
1548 | /* Type Not Supported. */ | 1845 | /* Type Not Supported. */ |
1549 | DEBUG4(printk(KERN_WARNING | 1846 | DEBUG4(printk(KERN_WARNING |
@@ -1723,8 +2020,10 @@ qla24xx_msix_rsp_q(int irq, void *dev_id) | |||
1723 | 2020 | ||
1724 | vha = qla25xx_get_host(rsp); | 2021 | vha = qla25xx_get_host(rsp); |
1725 | qla24xx_process_response_queue(vha, rsp); | 2022 | qla24xx_process_response_queue(vha, rsp); |
1726 | WRT_REG_DWORD(®->hccr, HCCRX_CLR_RISC_INT); | 2023 | if (!ha->mqenable) { |
1727 | 2024 | WRT_REG_DWORD(®->hccr, HCCRX_CLR_RISC_INT); | |
2025 | RD_REG_DWORD_RELAXED(®->hccr); | ||
2026 | } | ||
1728 | spin_unlock_irq(&ha->hardware_lock); | 2027 | spin_unlock_irq(&ha->hardware_lock); |
1729 | 2028 | ||
1730 | return IRQ_HANDLED; | 2029 | return IRQ_HANDLED; |
diff --git a/drivers/scsi/qla2xxx/qla_mbx.c b/drivers/scsi/qla2xxx/qla_mbx.c index fe69f3057671..b6202fe118ac 100644 --- a/drivers/scsi/qla2xxx/qla_mbx.c +++ b/drivers/scsi/qla2xxx/qla_mbx.c | |||
@@ -1507,7 +1507,7 @@ qla24xx_login_fabric(scsi_qla_host_t *vha, uint16_t loop_id, uint8_t domain, | |||
1507 | 1507 | ||
1508 | DEBUG11(printk("%s(%ld): entered.\n", __func__, vha->host_no)); | 1508 | DEBUG11(printk("%s(%ld): entered.\n", __func__, vha->host_no)); |
1509 | 1509 | ||
1510 | if (ql2xmultique_tag) | 1510 | if (ha->flags.cpu_affinity_enabled) |
1511 | req = ha->req_q_map[0]; | 1511 | req = ha->req_q_map[0]; |
1512 | else | 1512 | else |
1513 | req = vha->req; | 1513 | req = vha->req; |
@@ -2324,7 +2324,7 @@ __qla24xx_issue_tmf(char *name, uint32_t type, struct fc_port *fcport, | |||
2324 | vha = fcport->vha; | 2324 | vha = fcport->vha; |
2325 | ha = vha->hw; | 2325 | ha = vha->hw; |
2326 | req = vha->req; | 2326 | req = vha->req; |
2327 | if (ql2xmultique_tag) | 2327 | if (ha->flags.cpu_affinity_enabled) |
2328 | rsp = ha->rsp_q_map[tag + 1]; | 2328 | rsp = ha->rsp_q_map[tag + 1]; |
2329 | else | 2329 | else |
2330 | rsp = req->rsp; | 2330 | rsp = req->rsp; |
@@ -2746,7 +2746,8 @@ qla24xx_report_id_acquisition(scsi_qla_host_t *vha, | |||
2746 | if (rptid_entry->format == 0) { | 2746 | if (rptid_entry->format == 0) { |
2747 | DEBUG15(printk("%s:format 0 : scsi(%ld) number of VPs setup %d," | 2747 | DEBUG15(printk("%s:format 0 : scsi(%ld) number of VPs setup %d," |
2748 | " number of VPs acquired %d\n", __func__, vha->host_no, | 2748 | " number of VPs acquired %d\n", __func__, vha->host_no, |
2749 | MSB(rptid_entry->vp_count), LSB(rptid_entry->vp_count))); | 2749 | MSB(le16_to_cpu(rptid_entry->vp_count)), |
2750 | LSB(le16_to_cpu(rptid_entry->vp_count)))); | ||
2750 | DEBUG15(printk("%s primary port id %02x%02x%02x\n", __func__, | 2751 | DEBUG15(printk("%s primary port id %02x%02x%02x\n", __func__, |
2751 | rptid_entry->port_id[2], rptid_entry->port_id[1], | 2752 | rptid_entry->port_id[2], rptid_entry->port_id[1], |
2752 | rptid_entry->port_id[0])); | 2753 | rptid_entry->port_id[0])); |
diff --git a/drivers/scsi/qla2xxx/qla_mid.c b/drivers/scsi/qla2xxx/qla_mid.c index cd78c501803a..42b799abba57 100644 --- a/drivers/scsi/qla2xxx/qla_mid.c +++ b/drivers/scsi/qla2xxx/qla_mid.c | |||
@@ -42,7 +42,6 @@ qla24xx_allocate_vp_id(scsi_qla_host_t *vha) | |||
42 | 42 | ||
43 | set_bit(vp_id, ha->vp_idx_map); | 43 | set_bit(vp_id, ha->vp_idx_map); |
44 | ha->num_vhosts++; | 44 | ha->num_vhosts++; |
45 | ha->cur_vport_count++; | ||
46 | vha->vp_idx = vp_id; | 45 | vha->vp_idx = vp_id; |
47 | list_add_tail(&vha->list, &ha->vp_list); | 46 | list_add_tail(&vha->list, &ha->vp_list); |
48 | mutex_unlock(&ha->vport_lock); | 47 | mutex_unlock(&ha->vport_lock); |
@@ -58,7 +57,6 @@ qla24xx_deallocate_vp_id(scsi_qla_host_t *vha) | |||
58 | mutex_lock(&ha->vport_lock); | 57 | mutex_lock(&ha->vport_lock); |
59 | vp_id = vha->vp_idx; | 58 | vp_id = vha->vp_idx; |
60 | ha->num_vhosts--; | 59 | ha->num_vhosts--; |
61 | ha->cur_vport_count--; | ||
62 | clear_bit(vp_id, ha->vp_idx_map); | 60 | clear_bit(vp_id, ha->vp_idx_map); |
63 | list_del(&vha->list); | 61 | list_del(&vha->list); |
64 | mutex_unlock(&ha->vport_lock); | 62 | mutex_unlock(&ha->vport_lock); |
@@ -235,7 +233,11 @@ qla2x00_vp_abort_isp(scsi_qla_host_t *vha) | |||
235 | atomic_set(&vha->loop_down_timer, LOOP_DOWN_TIME); | 233 | atomic_set(&vha->loop_down_timer, LOOP_DOWN_TIME); |
236 | } | 234 | } |
237 | 235 | ||
238 | /* To exclusively reset vport, we need to log it out first.*/ | 236 | /* |
237 | * To exclusively reset vport, we need to log it out first. Note: this | ||
238 | * control_vp can fail if ISP reset is already issued, this is | ||
239 | * expected, as the vp would be already logged out due to ISP reset. | ||
240 | */ | ||
239 | if (!test_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags)) | 241 | if (!test_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags)) |
240 | qla24xx_control_vp(vha, VCE_COMMAND_DISABLE_VPS_LOGO_ALL); | 242 | qla24xx_control_vp(vha, VCE_COMMAND_DISABLE_VPS_LOGO_ALL); |
241 | 243 | ||
@@ -247,18 +249,11 @@ qla2x00_vp_abort_isp(scsi_qla_host_t *vha) | |||
247 | static int | 249 | static int |
248 | qla2x00_do_dpc_vp(scsi_qla_host_t *vha) | 250 | qla2x00_do_dpc_vp(scsi_qla_host_t *vha) |
249 | { | 251 | { |
250 | struct qla_hw_data *ha = vha->hw; | 252 | qla2x00_do_work(vha); |
251 | scsi_qla_host_t *base_vha = pci_get_drvdata(ha->pdev); | ||
252 | 253 | ||
253 | if (test_and_clear_bit(VP_IDX_ACQUIRED, &vha->vp_flags)) { | 254 | if (test_and_clear_bit(VP_IDX_ACQUIRED, &vha->vp_flags)) { |
254 | /* VP acquired. complete port configuration */ | 255 | /* VP acquired. complete port configuration */ |
255 | if (atomic_read(&base_vha->loop_state) == LOOP_READY) { | 256 | qla24xx_configure_vp(vha); |
256 | qla24xx_configure_vp(vha); | ||
257 | } else { | ||
258 | set_bit(VP_IDX_ACQUIRED, &vha->vp_flags); | ||
259 | set_bit(VP_DPC_NEEDED, &base_vha->dpc_flags); | ||
260 | } | ||
261 | |||
262 | return 0; | 257 | return 0; |
263 | } | 258 | } |
264 | 259 | ||
@@ -309,6 +304,9 @@ qla2x00_do_dpc_all_vps(scsi_qla_host_t *vha) | |||
309 | 304 | ||
310 | clear_bit(VP_DPC_NEEDED, &vha->dpc_flags); | 305 | clear_bit(VP_DPC_NEEDED, &vha->dpc_flags); |
311 | 306 | ||
307 | if (!(ha->current_topology & ISP_CFG_F)) | ||
308 | return; | ||
309 | |||
312 | list_for_each_entry_safe(vp, tvp, &ha->vp_list, list) { | 310 | list_for_each_entry_safe(vp, tvp, &ha->vp_list, list) { |
313 | if (vp->vp_idx) | 311 | if (vp->vp_idx) |
314 | ret = qla2x00_do_dpc_vp(vp); | 312 | ret = qla2x00_do_dpc_vp(vp); |
@@ -413,6 +411,11 @@ qla24xx_create_vhost(struct fc_vport *fc_vport) | |||
413 | 411 | ||
414 | vha->flags.init_done = 1; | 412 | vha->flags.init_done = 1; |
415 | 413 | ||
414 | mutex_lock(&ha->vport_lock); | ||
415 | set_bit(vha->vp_idx, ha->vp_idx_map); | ||
416 | ha->cur_vport_count++; | ||
417 | mutex_unlock(&ha->vport_lock); | ||
418 | |||
416 | return vha; | 419 | return vha; |
417 | 420 | ||
418 | create_vhost_failed: | 421 | create_vhost_failed: |
diff --git a/drivers/scsi/qla2xxx/qla_os.c b/drivers/scsi/qla2xxx/qla_os.c index f0396e79b6fa..b79fca7d461b 100644 --- a/drivers/scsi/qla2xxx/qla_os.c +++ b/drivers/scsi/qla2xxx/qla_os.c | |||
@@ -287,9 +287,12 @@ static int qla25xx_setup_mode(struct scsi_qla_host *vha) | |||
287 | int ques, req, ret; | 287 | int ques, req, ret; |
288 | struct qla_hw_data *ha = vha->hw; | 288 | struct qla_hw_data *ha = vha->hw; |
289 | 289 | ||
290 | if (!(ha->fw_attributes & BIT_6)) { | ||
291 | qla_printk(KERN_INFO, ha, | ||
292 | "Firmware is not multi-queue capable\n"); | ||
293 | goto fail; | ||
294 | } | ||
290 | if (ql2xmultique_tag) { | 295 | if (ql2xmultique_tag) { |
291 | /* CPU affinity mode */ | ||
292 | ha->wq = create_workqueue("qla2xxx_wq"); | ||
293 | /* create a request queue for IO */ | 296 | /* create a request queue for IO */ |
294 | options |= BIT_7; | 297 | options |= BIT_7; |
295 | req = qla25xx_create_req_que(ha, options, 0, 0, -1, | 298 | req = qla25xx_create_req_que(ha, options, 0, 0, -1, |
@@ -299,6 +302,7 @@ static int qla25xx_setup_mode(struct scsi_qla_host *vha) | |||
299 | "Can't create request queue\n"); | 302 | "Can't create request queue\n"); |
300 | goto fail; | 303 | goto fail; |
301 | } | 304 | } |
305 | ha->wq = create_workqueue("qla2xxx_wq"); | ||
302 | vha->req = ha->req_q_map[req]; | 306 | vha->req = ha->req_q_map[req]; |
303 | options |= BIT_1; | 307 | options |= BIT_1; |
304 | for (ques = 1; ques < ha->max_rsp_queues; ques++) { | 308 | for (ques = 1; ques < ha->max_rsp_queues; ques++) { |
@@ -309,6 +313,8 @@ static int qla25xx_setup_mode(struct scsi_qla_host *vha) | |||
309 | goto fail2; | 313 | goto fail2; |
310 | } | 314 | } |
311 | } | 315 | } |
316 | ha->flags.cpu_affinity_enabled = 1; | ||
317 | |||
312 | DEBUG2(qla_printk(KERN_INFO, ha, | 318 | DEBUG2(qla_printk(KERN_INFO, ha, |
313 | "CPU affinity mode enabled, no. of response" | 319 | "CPU affinity mode enabled, no. of response" |
314 | " queues:%d, no. of request queues:%d\n", | 320 | " queues:%d, no. of request queues:%d\n", |
@@ -317,8 +323,13 @@ static int qla25xx_setup_mode(struct scsi_qla_host *vha) | |||
317 | return 0; | 323 | return 0; |
318 | fail2: | 324 | fail2: |
319 | qla25xx_delete_queues(vha); | 325 | qla25xx_delete_queues(vha); |
326 | destroy_workqueue(ha->wq); | ||
327 | ha->wq = NULL; | ||
320 | fail: | 328 | fail: |
321 | ha->mqenable = 0; | 329 | ha->mqenable = 0; |
330 | kfree(ha->req_q_map); | ||
331 | kfree(ha->rsp_q_map); | ||
332 | ha->max_req_queues = ha->max_rsp_queues = 1; | ||
322 | return 1; | 333 | return 1; |
323 | } | 334 | } |
324 | 335 | ||
@@ -462,6 +473,7 @@ qla2x00_get_new_sp(scsi_qla_host_t *vha, fc_port_t *fcport, | |||
462 | sp->flags = 0; | 473 | sp->flags = 0; |
463 | CMD_SP(cmd) = (void *)sp; | 474 | CMD_SP(cmd) = (void *)sp; |
464 | cmd->scsi_done = done; | 475 | cmd->scsi_done = done; |
476 | sp->ctx = NULL; | ||
465 | 477 | ||
466 | return sp; | 478 | return sp; |
467 | } | 479 | } |
@@ -556,11 +568,8 @@ qla2x00_eh_wait_on_command(struct scsi_cmnd *cmd) | |||
556 | unsigned long wait_iter = ABORT_WAIT_ITER; | 568 | unsigned long wait_iter = ABORT_WAIT_ITER; |
557 | int ret = QLA_SUCCESS; | 569 | int ret = QLA_SUCCESS; |
558 | 570 | ||
559 | while (CMD_SP(cmd)) { | 571 | while (CMD_SP(cmd) && wait_iter--) { |
560 | msleep(ABORT_POLLING_PERIOD); | 572 | msleep(ABORT_POLLING_PERIOD); |
561 | |||
562 | if (--wait_iter) | ||
563 | break; | ||
564 | } | 573 | } |
565 | if (CMD_SP(cmd)) | 574 | if (CMD_SP(cmd)) |
566 | ret = QLA_FUNCTION_FAILED; | 575 | ret = QLA_FUNCTION_FAILED; |
@@ -698,6 +707,8 @@ qla2x00_abort_fcport_cmds(fc_port_t *fcport) | |||
698 | continue; | 707 | continue; |
699 | if (sp->fcport != fcport) | 708 | if (sp->fcport != fcport) |
700 | continue; | 709 | continue; |
710 | if (sp->ctx) | ||
711 | continue; | ||
701 | 712 | ||
702 | spin_unlock_irqrestore(&ha->hardware_lock, flags); | 713 | spin_unlock_irqrestore(&ha->hardware_lock, flags); |
703 | if (ha->isp_ops->abort_command(sp)) { | 714 | if (ha->isp_ops->abort_command(sp)) { |
@@ -783,7 +794,8 @@ qla2xxx_eh_abort(struct scsi_cmnd *cmd) | |||
783 | 794 | ||
784 | if (sp == NULL) | 795 | if (sp == NULL) |
785 | continue; | 796 | continue; |
786 | 797 | if (sp->ctx) | |
798 | continue; | ||
787 | if (sp->cmd != cmd) | 799 | if (sp->cmd != cmd) |
788 | continue; | 800 | continue; |
789 | 801 | ||
@@ -848,7 +860,8 @@ qla2x00_eh_wait_for_pending_commands(scsi_qla_host_t *vha, unsigned int t, | |||
848 | sp = req->outstanding_cmds[cnt]; | 860 | sp = req->outstanding_cmds[cnt]; |
849 | if (!sp) | 861 | if (!sp) |
850 | continue; | 862 | continue; |
851 | 863 | if (sp->ctx) | |
864 | continue; | ||
852 | if (vha->vp_idx != sp->fcport->vha->vp_idx) | 865 | if (vha->vp_idx != sp->fcport->vha->vp_idx) |
853 | continue; | 866 | continue; |
854 | match = 0; | 867 | match = 0; |
@@ -1106,8 +1119,7 @@ qla2x00_loop_reset(scsi_qla_host_t *vha) | |||
1106 | struct fc_port *fcport; | 1119 | struct fc_port *fcport; |
1107 | struct qla_hw_data *ha = vha->hw; | 1120 | struct qla_hw_data *ha = vha->hw; |
1108 | 1121 | ||
1109 | if (ha->flags.enable_lip_full_login && !vha->vp_idx && | 1122 | if (ha->flags.enable_lip_full_login && !IS_QLA81XX(ha)) { |
1110 | !IS_QLA81XX(ha)) { | ||
1111 | ret = qla2x00_full_login_lip(vha); | 1123 | ret = qla2x00_full_login_lip(vha); |
1112 | if (ret != QLA_SUCCESS) { | 1124 | if (ret != QLA_SUCCESS) { |
1113 | DEBUG2_3(printk("%s(%ld): failed: " | 1125 | DEBUG2_3(printk("%s(%ld): failed: " |
@@ -1120,7 +1132,7 @@ qla2x00_loop_reset(scsi_qla_host_t *vha) | |||
1120 | qla2x00_wait_for_loop_ready(vha); | 1132 | qla2x00_wait_for_loop_ready(vha); |
1121 | } | 1133 | } |
1122 | 1134 | ||
1123 | if (ha->flags.enable_lip_reset && !vha->vp_idx) { | 1135 | if (ha->flags.enable_lip_reset) { |
1124 | ret = qla2x00_lip_reset(vha); | 1136 | ret = qla2x00_lip_reset(vha); |
1125 | if (ret != QLA_SUCCESS) { | 1137 | if (ret != QLA_SUCCESS) { |
1126 | DEBUG2_3(printk("%s(%ld): failed: " | 1138 | DEBUG2_3(printk("%s(%ld): failed: " |
@@ -1154,6 +1166,7 @@ qla2x00_abort_all_cmds(scsi_qla_host_t *vha, int res) | |||
1154 | int que, cnt; | 1166 | int que, cnt; |
1155 | unsigned long flags; | 1167 | unsigned long flags; |
1156 | srb_t *sp; | 1168 | srb_t *sp; |
1169 | struct srb_ctx *ctx; | ||
1157 | struct qla_hw_data *ha = vha->hw; | 1170 | struct qla_hw_data *ha = vha->hw; |
1158 | struct req_que *req; | 1171 | struct req_que *req; |
1159 | 1172 | ||
@@ -1166,8 +1179,14 @@ qla2x00_abort_all_cmds(scsi_qla_host_t *vha, int res) | |||
1166 | sp = req->outstanding_cmds[cnt]; | 1179 | sp = req->outstanding_cmds[cnt]; |
1167 | if (sp) { | 1180 | if (sp) { |
1168 | req->outstanding_cmds[cnt] = NULL; | 1181 | req->outstanding_cmds[cnt] = NULL; |
1169 | sp->cmd->result = res; | 1182 | if (!sp->ctx) { |
1170 | qla2x00_sp_compl(ha, sp); | 1183 | sp->cmd->result = res; |
1184 | qla2x00_sp_compl(ha, sp); | ||
1185 | } else { | ||
1186 | ctx = sp->ctx; | ||
1187 | del_timer_sync(&ctx->timer); | ||
1188 | ctx->free(sp); | ||
1189 | } | ||
1171 | } | 1190 | } |
1172 | } | 1191 | } |
1173 | } | 1192 | } |
@@ -1193,6 +1212,7 @@ qla2xxx_slave_configure(struct scsi_device *sdev) | |||
1193 | scsi_qla_host_t *vha = shost_priv(sdev->host); | 1212 | scsi_qla_host_t *vha = shost_priv(sdev->host); |
1194 | struct qla_hw_data *ha = vha->hw; | 1213 | struct qla_hw_data *ha = vha->hw; |
1195 | struct fc_rport *rport = starget_to_rport(sdev->sdev_target); | 1214 | struct fc_rport *rport = starget_to_rport(sdev->sdev_target); |
1215 | fc_port_t *fcport = *(fc_port_t **)rport->dd_data; | ||
1196 | struct req_que *req = vha->req; | 1216 | struct req_que *req = vha->req; |
1197 | 1217 | ||
1198 | if (sdev->tagged_supported) | 1218 | if (sdev->tagged_supported) |
@@ -1201,6 +1221,8 @@ qla2xxx_slave_configure(struct scsi_device *sdev) | |||
1201 | scsi_deactivate_tcq(sdev, req->max_q_depth); | 1221 | scsi_deactivate_tcq(sdev, req->max_q_depth); |
1202 | 1222 | ||
1203 | rport->dev_loss_tmo = ha->port_down_retry_count; | 1223 | rport->dev_loss_tmo = ha->port_down_retry_count; |
1224 | if (sdev->type == TYPE_TAPE) | ||
1225 | fcport->flags |= FCF_TAPE_PRESENT; | ||
1204 | 1226 | ||
1205 | return 0; | 1227 | return 0; |
1206 | } | 1228 | } |
@@ -1923,6 +1945,7 @@ qla2x00_probe_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
1923 | if (ret) | 1945 | if (ret) |
1924 | goto probe_init_failed; | 1946 | goto probe_init_failed; |
1925 | /* Alloc arrays of request and response ring ptrs */ | 1947 | /* Alloc arrays of request and response ring ptrs */ |
1948 | que_init: | ||
1926 | if (!qla2x00_alloc_queues(ha)) { | 1949 | if (!qla2x00_alloc_queues(ha)) { |
1927 | qla_printk(KERN_WARNING, ha, | 1950 | qla_printk(KERN_WARNING, ha, |
1928 | "[ERROR] Failed to allocate memory for queue" | 1951 | "[ERROR] Failed to allocate memory for queue" |
@@ -1959,11 +1982,14 @@ qla2x00_probe_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
1959 | goto probe_failed; | 1982 | goto probe_failed; |
1960 | } | 1983 | } |
1961 | 1984 | ||
1962 | if (ha->mqenable) | 1985 | if (ha->mqenable) { |
1963 | if (qla25xx_setup_mode(base_vha)) | 1986 | if (qla25xx_setup_mode(base_vha)) { |
1964 | qla_printk(KERN_WARNING, ha, | 1987 | qla_printk(KERN_WARNING, ha, |
1965 | "Can't create queues, falling back to single" | 1988 | "Can't create queues, falling back to single" |
1966 | " queue mode\n"); | 1989 | " queue mode\n"); |
1990 | goto que_init; | ||
1991 | } | ||
1992 | } | ||
1967 | 1993 | ||
1968 | if (ha->flags.running_gold_fw) | 1994 | if (ha->flags.running_gold_fw) |
1969 | goto skip_dpc; | 1995 | goto skip_dpc; |
@@ -2155,17 +2181,19 @@ qla2x00_schedule_rport_del(struct scsi_qla_host *vha, fc_port_t *fcport, | |||
2155 | int defer) | 2181 | int defer) |
2156 | { | 2182 | { |
2157 | struct fc_rport *rport; | 2183 | struct fc_rport *rport; |
2184 | scsi_qla_host_t *base_vha; | ||
2158 | 2185 | ||
2159 | if (!fcport->rport) | 2186 | if (!fcport->rport) |
2160 | return; | 2187 | return; |
2161 | 2188 | ||
2162 | rport = fcport->rport; | 2189 | rport = fcport->rport; |
2163 | if (defer) { | 2190 | if (defer) { |
2191 | base_vha = pci_get_drvdata(vha->hw->pdev); | ||
2164 | spin_lock_irq(vha->host->host_lock); | 2192 | spin_lock_irq(vha->host->host_lock); |
2165 | fcport->drport = rport; | 2193 | fcport->drport = rport; |
2166 | spin_unlock_irq(vha->host->host_lock); | 2194 | spin_unlock_irq(vha->host->host_lock); |
2167 | set_bit(FCPORT_UPDATE_NEEDED, &vha->dpc_flags); | 2195 | set_bit(FCPORT_UPDATE_NEEDED, &base_vha->dpc_flags); |
2168 | qla2xxx_wake_dpc(vha); | 2196 | qla2xxx_wake_dpc(base_vha); |
2169 | } else | 2197 | } else |
2170 | fc_remote_port_delete(rport); | 2198 | fc_remote_port_delete(rport); |
2171 | } | 2199 | } |
@@ -2237,8 +2265,9 @@ qla2x00_mark_all_devices_lost(scsi_qla_host_t *vha, int defer) | |||
2237 | fc_port_t *fcport; | 2265 | fc_port_t *fcport; |
2238 | 2266 | ||
2239 | list_for_each_entry(fcport, &vha->vp_fcports, list) { | 2267 | list_for_each_entry(fcport, &vha->vp_fcports, list) { |
2240 | if (vha->vp_idx != fcport->vp_idx) | 2268 | if (vha->vp_idx != 0 && vha->vp_idx != fcport->vp_idx) |
2241 | continue; | 2269 | continue; |
2270 | |||
2242 | /* | 2271 | /* |
2243 | * No point in marking the device as lost, if the device is | 2272 | * No point in marking the device as lost, if the device is |
2244 | * already DEAD. | 2273 | * already DEAD. |
@@ -2246,10 +2275,12 @@ qla2x00_mark_all_devices_lost(scsi_qla_host_t *vha, int defer) | |||
2246 | if (atomic_read(&fcport->state) == FCS_DEVICE_DEAD) | 2275 | if (atomic_read(&fcport->state) == FCS_DEVICE_DEAD) |
2247 | continue; | 2276 | continue; |
2248 | if (atomic_read(&fcport->state) == FCS_ONLINE) { | 2277 | if (atomic_read(&fcport->state) == FCS_ONLINE) { |
2249 | atomic_set(&fcport->state, FCS_DEVICE_LOST); | 2278 | if (defer) |
2250 | qla2x00_schedule_rport_del(vha, fcport, defer); | 2279 | qla2x00_schedule_rport_del(vha, fcport, defer); |
2251 | } else | 2280 | else if (vha->vp_idx == fcport->vp_idx) |
2252 | atomic_set(&fcport->state, FCS_DEVICE_LOST); | 2281 | qla2x00_schedule_rport_del(vha, fcport, defer); |
2282 | } | ||
2283 | atomic_set(&fcport->state, FCS_DEVICE_LOST); | ||
2253 | } | 2284 | } |
2254 | } | 2285 | } |
2255 | 2286 | ||
@@ -2598,7 +2629,31 @@ qla2x00_post_idc_ack_work(struct scsi_qla_host *vha, uint16_t *mb) | |||
2598 | return qla2x00_post_work(vha, e); | 2629 | return qla2x00_post_work(vha, e); |
2599 | } | 2630 | } |
2600 | 2631 | ||
2601 | static void | 2632 | #define qla2x00_post_async_work(name, type) \ |
2633 | int qla2x00_post_async_##name##_work( \ | ||
2634 | struct scsi_qla_host *vha, \ | ||
2635 | fc_port_t *fcport, uint16_t *data) \ | ||
2636 | { \ | ||
2637 | struct qla_work_evt *e; \ | ||
2638 | \ | ||
2639 | e = qla2x00_alloc_work(vha, type); \ | ||
2640 | if (!e) \ | ||
2641 | return QLA_FUNCTION_FAILED; \ | ||
2642 | \ | ||
2643 | e->u.logio.fcport = fcport; \ | ||
2644 | if (data) { \ | ||
2645 | e->u.logio.data[0] = data[0]; \ | ||
2646 | e->u.logio.data[1] = data[1]; \ | ||
2647 | } \ | ||
2648 | return qla2x00_post_work(vha, e); \ | ||
2649 | } | ||
2650 | |||
2651 | qla2x00_post_async_work(login, QLA_EVT_ASYNC_LOGIN); | ||
2652 | qla2x00_post_async_work(login_done, QLA_EVT_ASYNC_LOGIN_DONE); | ||
2653 | qla2x00_post_async_work(logout, QLA_EVT_ASYNC_LOGOUT); | ||
2654 | qla2x00_post_async_work(logout_done, QLA_EVT_ASYNC_LOGOUT_DONE); | ||
2655 | |||
2656 | void | ||
2602 | qla2x00_do_work(struct scsi_qla_host *vha) | 2657 | qla2x00_do_work(struct scsi_qla_host *vha) |
2603 | { | 2658 | { |
2604 | struct qla_work_evt *e, *tmp; | 2659 | struct qla_work_evt *e, *tmp; |
@@ -2620,6 +2675,21 @@ qla2x00_do_work(struct scsi_qla_host *vha) | |||
2620 | case QLA_EVT_IDC_ACK: | 2675 | case QLA_EVT_IDC_ACK: |
2621 | qla81xx_idc_ack(vha, e->u.idc_ack.mb); | 2676 | qla81xx_idc_ack(vha, e->u.idc_ack.mb); |
2622 | break; | 2677 | break; |
2678 | case QLA_EVT_ASYNC_LOGIN: | ||
2679 | qla2x00_async_login(vha, e->u.logio.fcport, | ||
2680 | e->u.logio.data); | ||
2681 | break; | ||
2682 | case QLA_EVT_ASYNC_LOGIN_DONE: | ||
2683 | qla2x00_async_login_done(vha, e->u.logio.fcport, | ||
2684 | e->u.logio.data); | ||
2685 | break; | ||
2686 | case QLA_EVT_ASYNC_LOGOUT: | ||
2687 | qla2x00_async_logout(vha, e->u.logio.fcport); | ||
2688 | break; | ||
2689 | case QLA_EVT_ASYNC_LOGOUT_DONE: | ||
2690 | qla2x00_async_logout_done(vha, e->u.logio.fcport, | ||
2691 | e->u.logio.data); | ||
2692 | break; | ||
2623 | } | 2693 | } |
2624 | if (e->flags & QLA_EVT_FLAG_FREE) | 2694 | if (e->flags & QLA_EVT_FLAG_FREE) |
2625 | kfree(e); | 2695 | kfree(e); |
@@ -2635,6 +2705,7 @@ void qla2x00_relogin(struct scsi_qla_host *vha) | |||
2635 | int status; | 2705 | int status; |
2636 | uint16_t next_loopid = 0; | 2706 | uint16_t next_loopid = 0; |
2637 | struct qla_hw_data *ha = vha->hw; | 2707 | struct qla_hw_data *ha = vha->hw; |
2708 | uint16_t data[2]; | ||
2638 | 2709 | ||
2639 | list_for_each_entry(fcport, &vha->vp_fcports, list) { | 2710 | list_for_each_entry(fcport, &vha->vp_fcports, list) { |
2640 | /* | 2711 | /* |
@@ -2644,6 +2715,7 @@ void qla2x00_relogin(struct scsi_qla_host *vha) | |||
2644 | if (atomic_read(&fcport->state) != | 2715 | if (atomic_read(&fcport->state) != |
2645 | FCS_ONLINE && fcport->login_retry) { | 2716 | FCS_ONLINE && fcport->login_retry) { |
2646 | 2717 | ||
2718 | fcport->login_retry--; | ||
2647 | if (fcport->flags & FCF_FABRIC_DEVICE) { | 2719 | if (fcport->flags & FCF_FABRIC_DEVICE) { |
2648 | if (fcport->flags & FCF_TAPE_PRESENT) | 2720 | if (fcport->flags & FCF_TAPE_PRESENT) |
2649 | ha->isp_ops->fabric_logout(vha, | 2721 | ha->isp_ops->fabric_logout(vha, |
@@ -2652,13 +2724,22 @@ void qla2x00_relogin(struct scsi_qla_host *vha) | |||
2652 | fcport->d_id.b.area, | 2724 | fcport->d_id.b.area, |
2653 | fcport->d_id.b.al_pa); | 2725 | fcport->d_id.b.al_pa); |
2654 | 2726 | ||
2655 | status = qla2x00_fabric_login(vha, fcport, | 2727 | if (IS_ALOGIO_CAPABLE(ha)) { |
2656 | &next_loopid); | 2728 | data[0] = 0; |
2729 | data[1] = QLA_LOGIO_LOGIN_RETRIED; | ||
2730 | status = qla2x00_post_async_login_work( | ||
2731 | vha, fcport, data); | ||
2732 | if (status == QLA_SUCCESS) | ||
2733 | continue; | ||
2734 | /* Attempt a retry. */ | ||
2735 | status = 1; | ||
2736 | } else | ||
2737 | status = qla2x00_fabric_login(vha, | ||
2738 | fcport, &next_loopid); | ||
2657 | } else | 2739 | } else |
2658 | status = qla2x00_local_device_login(vha, | 2740 | status = qla2x00_local_device_login(vha, |
2659 | fcport); | 2741 | fcport); |
2660 | 2742 | ||
2661 | fcport->login_retry--; | ||
2662 | if (status == QLA_SUCCESS) { | 2743 | if (status == QLA_SUCCESS) { |
2663 | fcport->old_loop_id = fcport->loop_id; | 2744 | fcport->old_loop_id = fcport->loop_id; |
2664 | 2745 | ||
@@ -2831,6 +2912,9 @@ qla2x00_do_dpc(void *data) | |||
2831 | */ | 2912 | */ |
2832 | ha->dpc_active = 0; | 2913 | ha->dpc_active = 0; |
2833 | 2914 | ||
2915 | /* Cleanup any residual CTX SRBs. */ | ||
2916 | qla2x00_abort_all_cmds(base_vha, DID_NO_CONNECT << 16); | ||
2917 | |||
2834 | return 0; | 2918 | return 0; |
2835 | } | 2919 | } |
2836 | 2920 | ||
@@ -2971,6 +3055,8 @@ qla2x00_timer(scsi_qla_host_t *vha) | |||
2971 | sp = req->outstanding_cmds[index]; | 3055 | sp = req->outstanding_cmds[index]; |
2972 | if (!sp) | 3056 | if (!sp) |
2973 | continue; | 3057 | continue; |
3058 | if (sp->ctx) | ||
3059 | continue; | ||
2974 | sfcp = sp->fcport; | 3060 | sfcp = sp->fcport; |
2975 | if (!(sfcp->flags & FCF_TAPE_PRESENT)) | 3061 | if (!(sfcp->flags & FCF_TAPE_PRESENT)) |
2976 | continue; | 3062 | continue; |
@@ -2987,8 +3073,7 @@ qla2x00_timer(scsi_qla_host_t *vha) | |||
2987 | 3073 | ||
2988 | /* if the loop has been down for 4 minutes, reinit adapter */ | 3074 | /* if the loop has been down for 4 minutes, reinit adapter */ |
2989 | if (atomic_dec_and_test(&vha->loop_down_timer) != 0) { | 3075 | if (atomic_dec_and_test(&vha->loop_down_timer) != 0) { |
2990 | if (!(vha->device_flags & DFLG_NO_CABLE) && | 3076 | if (!(vha->device_flags & DFLG_NO_CABLE)) { |
2991 | !vha->vp_idx) { | ||
2992 | DEBUG(printk("scsi(%ld): Loop down - " | 3077 | DEBUG(printk("scsi(%ld): Loop down - " |
2993 | "aborting ISP.\n", | 3078 | "aborting ISP.\n", |
2994 | vha->host_no)); | 3079 | vha->host_no)); |
diff --git a/drivers/scsi/qla2xxx/qla_version.h b/drivers/scsi/qla2xxx/qla_version.h index 84369705a9ad..ac107a2c34a4 100644 --- a/drivers/scsi/qla2xxx/qla_version.h +++ b/drivers/scsi/qla2xxx/qla_version.h | |||
@@ -7,7 +7,7 @@ | |||
7 | /* | 7 | /* |
8 | * Driver version | 8 | * Driver version |
9 | */ | 9 | */ |
10 | #define QLA2XXX_VERSION "8.03.01-k4" | 10 | #define QLA2XXX_VERSION "8.03.01-k6" |
11 | 11 | ||
12 | #define QLA_DRIVER_MAJOR_VER 8 | 12 | #define QLA_DRIVER_MAJOR_VER 8 |
13 | #define QLA_DRIVER_MINOR_VER 3 | 13 | #define QLA_DRIVER_MINOR_VER 3 |
diff --git a/drivers/scsi/qla4xxx/ql4_isr.c b/drivers/scsi/qla4xxx/ql4_isr.c index 8025ee16588e..c196d55eae39 100644 --- a/drivers/scsi/qla4xxx/ql4_isr.c +++ b/drivers/scsi/qla4xxx/ql4_isr.c | |||
@@ -227,11 +227,11 @@ static void qla4xxx_status_entry(struct scsi_qla_host *ha, | |||
227 | case SCS_DATA_UNDERRUN: | 227 | case SCS_DATA_UNDERRUN: |
228 | case SCS_DATA_OVERRUN: | 228 | case SCS_DATA_OVERRUN: |
229 | if ((sts_entry->iscsiFlags & ISCSI_FLAG_RESIDUAL_OVER) || | 229 | if ((sts_entry->iscsiFlags & ISCSI_FLAG_RESIDUAL_OVER) || |
230 | (sts_entry->completionStatus == SCS_DATA_OVERRUN)) { | 230 | (sts_entry->completionStatus == SCS_DATA_OVERRUN)) { |
231 | DEBUG2(printk("scsi%ld:%d:%d:%d: %s: " "Data overrun, " | 231 | DEBUG2(printk("scsi%ld:%d:%d:%d: %s: " "Data overrun\n", |
232 | "residual = 0x%x\n", ha->host_no, | 232 | ha->host_no, |
233 | cmd->device->channel, cmd->device->id, | 233 | cmd->device->channel, cmd->device->id, |
234 | cmd->device->lun, __func__, residual)); | 234 | cmd->device->lun, __func__)); |
235 | 235 | ||
236 | cmd->result = DID_ERROR << 16; | 236 | cmd->result = DID_ERROR << 16; |
237 | break; | 237 | break; |
diff --git a/drivers/scsi/scsi.c b/drivers/scsi/scsi.c index 2de5f3ad640b..b6e03074cb8f 100644 --- a/drivers/scsi/scsi.c +++ b/drivers/scsi/scsi.c | |||
@@ -994,7 +994,7 @@ static int scsi_vpd_inquiry(struct scsi_device *sdev, unsigned char *buffer, | |||
994 | * all the existing users tried this hard. | 994 | * all the existing users tried this hard. |
995 | */ | 995 | */ |
996 | result = scsi_execute_req(sdev, cmd, DMA_FROM_DEVICE, buffer, | 996 | result = scsi_execute_req(sdev, cmd, DMA_FROM_DEVICE, buffer, |
997 | len + 4, NULL, 30 * HZ, 3, NULL); | 997 | len, NULL, 30 * HZ, 3, NULL); |
998 | if (result) | 998 | if (result) |
999 | return result; | 999 | return result; |
1000 | 1000 | ||
@@ -1021,13 +1021,14 @@ unsigned char *scsi_get_vpd_page(struct scsi_device *sdev, u8 page) | |||
1021 | { | 1021 | { |
1022 | int i, result; | 1022 | int i, result; |
1023 | unsigned int len; | 1023 | unsigned int len; |
1024 | unsigned char *buf = kmalloc(259, GFP_KERNEL); | 1024 | const unsigned int init_vpd_len = 255; |
1025 | unsigned char *buf = kmalloc(init_vpd_len, GFP_KERNEL); | ||
1025 | 1026 | ||
1026 | if (!buf) | 1027 | if (!buf) |
1027 | return NULL; | 1028 | return NULL; |
1028 | 1029 | ||
1029 | /* Ask for all the pages supported by this device */ | 1030 | /* Ask for all the pages supported by this device */ |
1030 | result = scsi_vpd_inquiry(sdev, buf, 0, 255); | 1031 | result = scsi_vpd_inquiry(sdev, buf, 0, init_vpd_len); |
1031 | if (result) | 1032 | if (result) |
1032 | goto fail; | 1033 | goto fail; |
1033 | 1034 | ||
@@ -1050,12 +1051,12 @@ unsigned char *scsi_get_vpd_page(struct scsi_device *sdev, u8 page) | |||
1050 | * Some pages are longer than 255 bytes. The actual length of | 1051 | * Some pages are longer than 255 bytes. The actual length of |
1051 | * the page is returned in the header. | 1052 | * the page is returned in the header. |
1052 | */ | 1053 | */ |
1053 | len = (buf[2] << 8) | buf[3]; | 1054 | len = ((buf[2] << 8) | buf[3]) + 4; |
1054 | if (len <= 255) | 1055 | if (len <= init_vpd_len) |
1055 | return buf; | 1056 | return buf; |
1056 | 1057 | ||
1057 | kfree(buf); | 1058 | kfree(buf); |
1058 | buf = kmalloc(len + 4, GFP_KERNEL); | 1059 | buf = kmalloc(len, GFP_KERNEL); |
1059 | result = scsi_vpd_inquiry(sdev, buf, page, len); | 1060 | result = scsi_vpd_inquiry(sdev, buf, page, len); |
1060 | if (result) | 1061 | if (result) |
1061 | goto fail; | 1062 | goto fail; |
diff --git a/drivers/scsi/scsi_error.c b/drivers/scsi/scsi_error.c index a1689353d7fd..877204daf549 100644 --- a/drivers/scsi/scsi_error.c +++ b/drivers/scsi/scsi_error.c | |||
@@ -382,9 +382,13 @@ static int scsi_eh_completed_normally(struct scsi_cmnd *scmd) | |||
382 | * who knows? FIXME(eric) | 382 | * who knows? FIXME(eric) |
383 | */ | 383 | */ |
384 | return SUCCESS; | 384 | return SUCCESS; |
385 | case RESERVATION_CONFLICT: | ||
386 | /* | ||
387 | * let issuer deal with this, it could be just fine | ||
388 | */ | ||
389 | return SUCCESS; | ||
385 | case BUSY: | 390 | case BUSY: |
386 | case QUEUE_FULL: | 391 | case QUEUE_FULL: |
387 | case RESERVATION_CONFLICT: | ||
388 | default: | 392 | default: |
389 | return FAILED; | 393 | return FAILED; |
390 | } | 394 | } |
diff --git a/drivers/scsi/scsi_lib.c b/drivers/scsi/scsi_lib.c index f3c40898fc7d..662024d86949 100644 --- a/drivers/scsi/scsi_lib.c +++ b/drivers/scsi/scsi_lib.c | |||
@@ -896,6 +896,7 @@ void scsi_io_completion(struct scsi_cmnd *cmd, unsigned int good_bytes) | |||
896 | scsi_print_result(cmd); | 896 | scsi_print_result(cmd); |
897 | if (driver_byte(result) & DRIVER_SENSE) | 897 | if (driver_byte(result) & DRIVER_SENSE) |
898 | scsi_print_sense("", cmd); | 898 | scsi_print_sense("", cmd); |
899 | scsi_print_command(cmd); | ||
899 | } | 900 | } |
900 | blk_end_request_all(req, -EIO); | 901 | blk_end_request_all(req, -EIO); |
901 | scsi_next_command(cmd); | 902 | scsi_next_command(cmd); |
diff --git a/drivers/scsi/scsi_transport_fc.c b/drivers/scsi/scsi_transport_fc.c index 292c02f810d0..b98885de6876 100644 --- a/drivers/scsi/scsi_transport_fc.c +++ b/drivers/scsi/scsi_transport_fc.c | |||
@@ -291,7 +291,7 @@ static void fc_scsi_scan_rport(struct work_struct *work); | |||
291 | #define FC_STARGET_NUM_ATTRS 3 | 291 | #define FC_STARGET_NUM_ATTRS 3 |
292 | #define FC_RPORT_NUM_ATTRS 10 | 292 | #define FC_RPORT_NUM_ATTRS 10 |
293 | #define FC_VPORT_NUM_ATTRS 9 | 293 | #define FC_VPORT_NUM_ATTRS 9 |
294 | #define FC_HOST_NUM_ATTRS 21 | 294 | #define FC_HOST_NUM_ATTRS 22 |
295 | 295 | ||
296 | struct fc_internal { | 296 | struct fc_internal { |
297 | struct scsi_transport_template t; | 297 | struct scsi_transport_template t; |
@@ -3432,7 +3432,7 @@ fc_bsg_jobdone(struct fc_bsg_job *job) | |||
3432 | 3432 | ||
3433 | /** | 3433 | /** |
3434 | * fc_bsg_softirq_done - softirq done routine for destroying the bsg requests | 3434 | * fc_bsg_softirq_done - softirq done routine for destroying the bsg requests |
3435 | * @req: BSG request that holds the job to be destroyed | 3435 | * @rq: BSG request that holds the job to be destroyed |
3436 | */ | 3436 | */ |
3437 | static void fc_bsg_softirq_done(struct request *rq) | 3437 | static void fc_bsg_softirq_done(struct request *rq) |
3438 | { | 3438 | { |
diff --git a/drivers/scsi/scsi_transport_iscsi.c b/drivers/scsi/scsi_transport_iscsi.c index b47240ca4b19..ad897df36615 100644 --- a/drivers/scsi/scsi_transport_iscsi.c +++ b/drivers/scsi/scsi_transport_iscsi.c | |||
@@ -36,6 +36,38 @@ | |||
36 | 36 | ||
37 | #define ISCSI_TRANSPORT_VERSION "2.0-870" | 37 | #define ISCSI_TRANSPORT_VERSION "2.0-870" |
38 | 38 | ||
39 | static int dbg_session; | ||
40 | module_param_named(debug_session, dbg_session, int, | ||
41 | S_IRUGO | S_IWUSR); | ||
42 | MODULE_PARM_DESC(debug_session, | ||
43 | "Turn on debugging for sessions in scsi_transport_iscsi " | ||
44 | "module. Set to 1 to turn on, and zero to turn off. Default " | ||
45 | "is off."); | ||
46 | |||
47 | static int dbg_conn; | ||
48 | module_param_named(debug_conn, dbg_conn, int, | ||
49 | S_IRUGO | S_IWUSR); | ||
50 | MODULE_PARM_DESC(debug_conn, | ||
51 | "Turn on debugging for connections in scsi_transport_iscsi " | ||
52 | "module. Set to 1 to turn on, and zero to turn off. Default " | ||
53 | "is off."); | ||
54 | |||
55 | #define ISCSI_DBG_TRANS_SESSION(_session, dbg_fmt, arg...) \ | ||
56 | do { \ | ||
57 | if (dbg_session) \ | ||
58 | iscsi_cls_session_printk(KERN_INFO, _session, \ | ||
59 | "%s: " dbg_fmt, \ | ||
60 | __func__, ##arg); \ | ||
61 | } while (0); | ||
62 | |||
63 | #define ISCSI_DBG_TRANS_CONN(_conn, dbg_fmt, arg...) \ | ||
64 | do { \ | ||
65 | if (dbg_conn) \ | ||
66 | iscsi_cls_conn_printk(KERN_INFO, _conn, \ | ||
67 | "%s: " dbg_fmt, \ | ||
68 | __func__, ##arg); \ | ||
69 | } while (0); | ||
70 | |||
39 | struct iscsi_internal { | 71 | struct iscsi_internal { |
40 | struct scsi_transport_template t; | 72 | struct scsi_transport_template t; |
41 | struct iscsi_transport *iscsi_transport; | 73 | struct iscsi_transport *iscsi_transport; |
@@ -377,6 +409,7 @@ static void iscsi_session_release(struct device *dev) | |||
377 | 409 | ||
378 | shost = iscsi_session_to_shost(session); | 410 | shost = iscsi_session_to_shost(session); |
379 | scsi_host_put(shost); | 411 | scsi_host_put(shost); |
412 | ISCSI_DBG_TRANS_SESSION(session, "Completing session release\n"); | ||
380 | kfree(session); | 413 | kfree(session); |
381 | } | 414 | } |
382 | 415 | ||
@@ -441,6 +474,9 @@ static int iscsi_user_scan_session(struct device *dev, void *data) | |||
441 | return 0; | 474 | return 0; |
442 | 475 | ||
443 | session = iscsi_dev_to_session(dev); | 476 | session = iscsi_dev_to_session(dev); |
477 | |||
478 | ISCSI_DBG_TRANS_SESSION(session, "Scanning session\n"); | ||
479 | |||
444 | shost = iscsi_session_to_shost(session); | 480 | shost = iscsi_session_to_shost(session); |
445 | ihost = shost->shost_data; | 481 | ihost = shost->shost_data; |
446 | 482 | ||
@@ -448,8 +484,7 @@ static int iscsi_user_scan_session(struct device *dev, void *data) | |||
448 | spin_lock_irqsave(&session->lock, flags); | 484 | spin_lock_irqsave(&session->lock, flags); |
449 | if (session->state != ISCSI_SESSION_LOGGED_IN) { | 485 | if (session->state != ISCSI_SESSION_LOGGED_IN) { |
450 | spin_unlock_irqrestore(&session->lock, flags); | 486 | spin_unlock_irqrestore(&session->lock, flags); |
451 | mutex_unlock(&ihost->mutex); | 487 | goto user_scan_exit; |
452 | return 0; | ||
453 | } | 488 | } |
454 | id = session->target_id; | 489 | id = session->target_id; |
455 | spin_unlock_irqrestore(&session->lock, flags); | 490 | spin_unlock_irqrestore(&session->lock, flags); |
@@ -462,7 +497,10 @@ static int iscsi_user_scan_session(struct device *dev, void *data) | |||
462 | scsi_scan_target(&session->dev, 0, id, | 497 | scsi_scan_target(&session->dev, 0, id, |
463 | scan_data->lun, 1); | 498 | scan_data->lun, 1); |
464 | } | 499 | } |
500 | |||
501 | user_scan_exit: | ||
465 | mutex_unlock(&ihost->mutex); | 502 | mutex_unlock(&ihost->mutex); |
503 | ISCSI_DBG_TRANS_SESSION(session, "Completed session scan\n"); | ||
466 | return 0; | 504 | return 0; |
467 | } | 505 | } |
468 | 506 | ||
@@ -522,7 +560,9 @@ static void session_recovery_timedout(struct work_struct *work) | |||
522 | if (session->transport->session_recovery_timedout) | 560 | if (session->transport->session_recovery_timedout) |
523 | session->transport->session_recovery_timedout(session); | 561 | session->transport->session_recovery_timedout(session); |
524 | 562 | ||
563 | ISCSI_DBG_TRANS_SESSION(session, "Unblocking SCSI target\n"); | ||
525 | scsi_target_unblock(&session->dev); | 564 | scsi_target_unblock(&session->dev); |
565 | ISCSI_DBG_TRANS_SESSION(session, "Completed unblocking SCSI target\n"); | ||
526 | } | 566 | } |
527 | 567 | ||
528 | static void __iscsi_unblock_session(struct work_struct *work) | 568 | static void __iscsi_unblock_session(struct work_struct *work) |
@@ -534,6 +574,7 @@ static void __iscsi_unblock_session(struct work_struct *work) | |||
534 | struct iscsi_cls_host *ihost = shost->shost_data; | 574 | struct iscsi_cls_host *ihost = shost->shost_data; |
535 | unsigned long flags; | 575 | unsigned long flags; |
536 | 576 | ||
577 | ISCSI_DBG_TRANS_SESSION(session, "Unblocking session\n"); | ||
537 | /* | 578 | /* |
538 | * The recovery and unblock work get run from the same workqueue, | 579 | * The recovery and unblock work get run from the same workqueue, |
539 | * so try to cancel it if it was going to run after this unblock. | 580 | * so try to cancel it if it was going to run after this unblock. |
@@ -553,6 +594,7 @@ static void __iscsi_unblock_session(struct work_struct *work) | |||
553 | if (scsi_queue_work(shost, &session->scan_work)) | 594 | if (scsi_queue_work(shost, &session->scan_work)) |
554 | atomic_inc(&ihost->nr_scans); | 595 | atomic_inc(&ihost->nr_scans); |
555 | } | 596 | } |
597 | ISCSI_DBG_TRANS_SESSION(session, "Completed unblocking session\n"); | ||
556 | } | 598 | } |
557 | 599 | ||
558 | /** | 600 | /** |
@@ -579,10 +621,12 @@ static void __iscsi_block_session(struct work_struct *work) | |||
579 | block_work); | 621 | block_work); |
580 | unsigned long flags; | 622 | unsigned long flags; |
581 | 623 | ||
624 | ISCSI_DBG_TRANS_SESSION(session, "Blocking session\n"); | ||
582 | spin_lock_irqsave(&session->lock, flags); | 625 | spin_lock_irqsave(&session->lock, flags); |
583 | session->state = ISCSI_SESSION_FAILED; | 626 | session->state = ISCSI_SESSION_FAILED; |
584 | spin_unlock_irqrestore(&session->lock, flags); | 627 | spin_unlock_irqrestore(&session->lock, flags); |
585 | scsi_target_block(&session->dev); | 628 | scsi_target_block(&session->dev); |
629 | ISCSI_DBG_TRANS_SESSION(session, "Completed SCSI target blocking\n"); | ||
586 | queue_delayed_work(iscsi_eh_timer_workq, &session->recovery_work, | 630 | queue_delayed_work(iscsi_eh_timer_workq, &session->recovery_work, |
587 | session->recovery_tmo * HZ); | 631 | session->recovery_tmo * HZ); |
588 | } | 632 | } |
@@ -602,6 +646,8 @@ static void __iscsi_unbind_session(struct work_struct *work) | |||
602 | struct iscsi_cls_host *ihost = shost->shost_data; | 646 | struct iscsi_cls_host *ihost = shost->shost_data; |
603 | unsigned long flags; | 647 | unsigned long flags; |
604 | 648 | ||
649 | ISCSI_DBG_TRANS_SESSION(session, "Unbinding session\n"); | ||
650 | |||
605 | /* Prevent new scans and make sure scanning is not in progress */ | 651 | /* Prevent new scans and make sure scanning is not in progress */ |
606 | mutex_lock(&ihost->mutex); | 652 | mutex_lock(&ihost->mutex); |
607 | spin_lock_irqsave(&session->lock, flags); | 653 | spin_lock_irqsave(&session->lock, flags); |
@@ -616,6 +662,7 @@ static void __iscsi_unbind_session(struct work_struct *work) | |||
616 | 662 | ||
617 | scsi_remove_target(&session->dev); | 663 | scsi_remove_target(&session->dev); |
618 | iscsi_session_event(session, ISCSI_KEVENT_UNBIND_SESSION); | 664 | iscsi_session_event(session, ISCSI_KEVENT_UNBIND_SESSION); |
665 | ISCSI_DBG_TRANS_SESSION(session, "Completed target removal\n"); | ||
619 | } | 666 | } |
620 | 667 | ||
621 | struct iscsi_cls_session * | 668 | struct iscsi_cls_session * |
@@ -647,6 +694,8 @@ iscsi_alloc_session(struct Scsi_Host *shost, struct iscsi_transport *transport, | |||
647 | device_initialize(&session->dev); | 694 | device_initialize(&session->dev); |
648 | if (dd_size) | 695 | if (dd_size) |
649 | session->dd_data = &session[1]; | 696 | session->dd_data = &session[1]; |
697 | |||
698 | ISCSI_DBG_TRANS_SESSION(session, "Completed session allocation\n"); | ||
650 | return session; | 699 | return session; |
651 | } | 700 | } |
652 | EXPORT_SYMBOL_GPL(iscsi_alloc_session); | 701 | EXPORT_SYMBOL_GPL(iscsi_alloc_session); |
@@ -712,6 +761,7 @@ int iscsi_add_session(struct iscsi_cls_session *session, unsigned int target_id) | |||
712 | spin_unlock_irqrestore(&sesslock, flags); | 761 | spin_unlock_irqrestore(&sesslock, flags); |
713 | 762 | ||
714 | iscsi_session_event(session, ISCSI_KEVENT_CREATE_SESSION); | 763 | iscsi_session_event(session, ISCSI_KEVENT_CREATE_SESSION); |
764 | ISCSI_DBG_TRANS_SESSION(session, "Completed session adding\n"); | ||
715 | return 0; | 765 | return 0; |
716 | 766 | ||
717 | release_host: | 767 | release_host: |
@@ -752,6 +802,7 @@ static void iscsi_conn_release(struct device *dev) | |||
752 | struct iscsi_cls_conn *conn = iscsi_dev_to_conn(dev); | 802 | struct iscsi_cls_conn *conn = iscsi_dev_to_conn(dev); |
753 | struct device *parent = conn->dev.parent; | 803 | struct device *parent = conn->dev.parent; |
754 | 804 | ||
805 | ISCSI_DBG_TRANS_CONN(conn, "Releasing conn\n"); | ||
755 | kfree(conn); | 806 | kfree(conn); |
756 | put_device(parent); | 807 | put_device(parent); |
757 | } | 808 | } |
@@ -774,6 +825,8 @@ void iscsi_remove_session(struct iscsi_cls_session *session) | |||
774 | unsigned long flags; | 825 | unsigned long flags; |
775 | int err; | 826 | int err; |
776 | 827 | ||
828 | ISCSI_DBG_TRANS_SESSION(session, "Removing session\n"); | ||
829 | |||
777 | spin_lock_irqsave(&sesslock, flags); | 830 | spin_lock_irqsave(&sesslock, flags); |
778 | list_del(&session->sess_list); | 831 | list_del(&session->sess_list); |
779 | spin_unlock_irqrestore(&sesslock, flags); | 832 | spin_unlock_irqrestore(&sesslock, flags); |
@@ -807,12 +860,15 @@ void iscsi_remove_session(struct iscsi_cls_session *session) | |||
807 | "for session. Error %d.\n", err); | 860 | "for session. Error %d.\n", err); |
808 | 861 | ||
809 | transport_unregister_device(&session->dev); | 862 | transport_unregister_device(&session->dev); |
863 | |||
864 | ISCSI_DBG_TRANS_SESSION(session, "Completing session removal\n"); | ||
810 | device_del(&session->dev); | 865 | device_del(&session->dev); |
811 | } | 866 | } |
812 | EXPORT_SYMBOL_GPL(iscsi_remove_session); | 867 | EXPORT_SYMBOL_GPL(iscsi_remove_session); |
813 | 868 | ||
814 | void iscsi_free_session(struct iscsi_cls_session *session) | 869 | void iscsi_free_session(struct iscsi_cls_session *session) |
815 | { | 870 | { |
871 | ISCSI_DBG_TRANS_SESSION(session, "Freeing session\n"); | ||
816 | iscsi_session_event(session, ISCSI_KEVENT_DESTROY_SESSION); | 872 | iscsi_session_event(session, ISCSI_KEVENT_DESTROY_SESSION); |
817 | put_device(&session->dev); | 873 | put_device(&session->dev); |
818 | } | 874 | } |
@@ -828,6 +884,7 @@ EXPORT_SYMBOL_GPL(iscsi_free_session); | |||
828 | int iscsi_destroy_session(struct iscsi_cls_session *session) | 884 | int iscsi_destroy_session(struct iscsi_cls_session *session) |
829 | { | 885 | { |
830 | iscsi_remove_session(session); | 886 | iscsi_remove_session(session); |
887 | ISCSI_DBG_TRANS_SESSION(session, "Completing session destruction\n"); | ||
831 | iscsi_free_session(session); | 888 | iscsi_free_session(session); |
832 | return 0; | 889 | return 0; |
833 | } | 890 | } |
@@ -885,6 +942,8 @@ iscsi_create_conn(struct iscsi_cls_session *session, int dd_size, uint32_t cid) | |||
885 | list_add(&conn->conn_list, &connlist); | 942 | list_add(&conn->conn_list, &connlist); |
886 | conn->active = 1; | 943 | conn->active = 1; |
887 | spin_unlock_irqrestore(&connlock, flags); | 944 | spin_unlock_irqrestore(&connlock, flags); |
945 | |||
946 | ISCSI_DBG_TRANS_CONN(conn, "Completed conn creation\n"); | ||
888 | return conn; | 947 | return conn; |
889 | 948 | ||
890 | release_parent_ref: | 949 | release_parent_ref: |
@@ -912,6 +971,7 @@ int iscsi_destroy_conn(struct iscsi_cls_conn *conn) | |||
912 | spin_unlock_irqrestore(&connlock, flags); | 971 | spin_unlock_irqrestore(&connlock, flags); |
913 | 972 | ||
914 | transport_unregister_device(&conn->dev); | 973 | transport_unregister_device(&conn->dev); |
974 | ISCSI_DBG_TRANS_CONN(conn, "Completing conn destruction\n"); | ||
915 | device_unregister(&conn->dev); | 975 | device_unregister(&conn->dev); |
916 | return 0; | 976 | return 0; |
917 | } | 977 | } |
@@ -1200,6 +1260,9 @@ int iscsi_session_event(struct iscsi_cls_session *session, | |||
1200 | "Cannot notify userspace of session " | 1260 | "Cannot notify userspace of session " |
1201 | "event %u. Check iscsi daemon\n", | 1261 | "event %u. Check iscsi daemon\n", |
1202 | event); | 1262 | event); |
1263 | |||
1264 | ISCSI_DBG_TRANS_SESSION(session, "Completed handling event %d rc %d\n", | ||
1265 | event, rc); | ||
1203 | return rc; | 1266 | return rc; |
1204 | } | 1267 | } |
1205 | EXPORT_SYMBOL_GPL(iscsi_session_event); | 1268 | EXPORT_SYMBOL_GPL(iscsi_session_event); |
@@ -1221,6 +1284,8 @@ iscsi_if_create_session(struct iscsi_internal *priv, struct iscsi_endpoint *ep, | |||
1221 | shost = iscsi_session_to_shost(session); | 1284 | shost = iscsi_session_to_shost(session); |
1222 | ev->r.c_session_ret.host_no = shost->host_no; | 1285 | ev->r.c_session_ret.host_no = shost->host_no; |
1223 | ev->r.c_session_ret.sid = session->sid; | 1286 | ev->r.c_session_ret.sid = session->sid; |
1287 | ISCSI_DBG_TRANS_SESSION(session, | ||
1288 | "Completed creating transport session\n"); | ||
1224 | return 0; | 1289 | return 0; |
1225 | } | 1290 | } |
1226 | 1291 | ||
@@ -1246,6 +1311,8 @@ iscsi_if_create_conn(struct iscsi_transport *transport, struct iscsi_uevent *ev) | |||
1246 | 1311 | ||
1247 | ev->r.c_conn_ret.sid = session->sid; | 1312 | ev->r.c_conn_ret.sid = session->sid; |
1248 | ev->r.c_conn_ret.cid = conn->cid; | 1313 | ev->r.c_conn_ret.cid = conn->cid; |
1314 | |||
1315 | ISCSI_DBG_TRANS_CONN(conn, "Completed creating transport conn\n"); | ||
1249 | return 0; | 1316 | return 0; |
1250 | } | 1317 | } |
1251 | 1318 | ||
@@ -1258,8 +1325,10 @@ iscsi_if_destroy_conn(struct iscsi_transport *transport, struct iscsi_uevent *ev | |||
1258 | if (!conn) | 1325 | if (!conn) |
1259 | return -EINVAL; | 1326 | return -EINVAL; |
1260 | 1327 | ||
1328 | ISCSI_DBG_TRANS_CONN(conn, "Destroying transport conn\n"); | ||
1261 | if (transport->destroy_conn) | 1329 | if (transport->destroy_conn) |
1262 | transport->destroy_conn(conn); | 1330 | transport->destroy_conn(conn); |
1331 | |||
1263 | return 0; | 1332 | return 0; |
1264 | } | 1333 | } |
1265 | 1334 | ||
diff --git a/drivers/scsi/scsi_transport_sas.c b/drivers/scsi/scsi_transport_sas.c index 0895d3c71b03..fd47cb1bee1b 100644 --- a/drivers/scsi/scsi_transport_sas.c +++ b/drivers/scsi/scsi_transport_sas.c | |||
@@ -1692,10 +1692,6 @@ sas_attach_transport(struct sas_function_template *ft) | |||
1692 | i->f = ft; | 1692 | i->f = ft; |
1693 | 1693 | ||
1694 | count = 0; | 1694 | count = 0; |
1695 | SETUP_PORT_ATTRIBUTE(num_phys); | ||
1696 | i->host_attrs[count] = NULL; | ||
1697 | |||
1698 | count = 0; | ||
1699 | SETUP_PHY_ATTRIBUTE(initiator_port_protocols); | 1695 | SETUP_PHY_ATTRIBUTE(initiator_port_protocols); |
1700 | SETUP_PHY_ATTRIBUTE(target_port_protocols); | 1696 | SETUP_PHY_ATTRIBUTE(target_port_protocols); |
1701 | SETUP_PHY_ATTRIBUTE(device_type); | 1697 | SETUP_PHY_ATTRIBUTE(device_type); |
diff --git a/drivers/scsi/sd.c b/drivers/scsi/sd.c index b7b9fec67a98..a89c421dab51 100644 --- a/drivers/scsi/sd.c +++ b/drivers/scsi/sd.c | |||
@@ -2021,6 +2021,7 @@ static void sd_probe_async(void *data, async_cookie_t cookie) | |||
2021 | 2021 | ||
2022 | sd_printk(KERN_NOTICE, sdkp, "Attached SCSI %sdisk\n", | 2022 | sd_printk(KERN_NOTICE, sdkp, "Attached SCSI %sdisk\n", |
2023 | sdp->removable ? "removable " : ""); | 2023 | sdp->removable ? "removable " : ""); |
2024 | put_device(&sdkp->dev); | ||
2024 | } | 2025 | } |
2025 | 2026 | ||
2026 | /** | 2027 | /** |
@@ -2106,6 +2107,7 @@ static int sd_probe(struct device *dev) | |||
2106 | 2107 | ||
2107 | get_device(&sdp->sdev_gendev); | 2108 | get_device(&sdp->sdev_gendev); |
2108 | 2109 | ||
2110 | get_device(&sdkp->dev); /* prevent release before async_schedule */ | ||
2109 | async_schedule(sd_probe_async, sdkp); | 2111 | async_schedule(sd_probe_async, sdkp); |
2110 | 2112 | ||
2111 | return 0; | 2113 | return 0; |
diff --git a/drivers/scsi/ses.c b/drivers/scsi/ses.c index 4f618f487356..55b034b72708 100644 --- a/drivers/scsi/ses.c +++ b/drivers/scsi/ses.c | |||
@@ -347,6 +347,97 @@ static int ses_enclosure_find_by_addr(struct enclosure_device *edev, | |||
347 | return 0; | 347 | return 0; |
348 | } | 348 | } |
349 | 349 | ||
350 | #define INIT_ALLOC_SIZE 32 | ||
351 | |||
352 | static void ses_enclosure_data_process(struct enclosure_device *edev, | ||
353 | struct scsi_device *sdev, | ||
354 | int create) | ||
355 | { | ||
356 | u32 result; | ||
357 | unsigned char *buf = NULL, *type_ptr, *desc_ptr, *addl_desc_ptr = NULL; | ||
358 | int i, j, page7_len, len, components; | ||
359 | struct ses_device *ses_dev = edev->scratch; | ||
360 | int types = ses_dev->page1[10]; | ||
361 | unsigned char *hdr_buf = kzalloc(INIT_ALLOC_SIZE, GFP_KERNEL); | ||
362 | |||
363 | if (!hdr_buf) | ||
364 | goto simple_populate; | ||
365 | |||
366 | /* re-read page 10 */ | ||
367 | if (ses_dev->page10) | ||
368 | ses_recv_diag(sdev, 10, ses_dev->page10, ses_dev->page10_len); | ||
369 | /* Page 7 for the descriptors is optional */ | ||
370 | result = ses_recv_diag(sdev, 7, hdr_buf, INIT_ALLOC_SIZE); | ||
371 | if (result) | ||
372 | goto simple_populate; | ||
373 | |||
374 | page7_len = len = (hdr_buf[2] << 8) + hdr_buf[3] + 4; | ||
375 | /* add 1 for trailing '\0' we'll use */ | ||
376 | buf = kzalloc(len + 1, GFP_KERNEL); | ||
377 | if (!buf) | ||
378 | goto simple_populate; | ||
379 | result = ses_recv_diag(sdev, 7, buf, len); | ||
380 | if (result) { | ||
381 | simple_populate: | ||
382 | kfree(buf); | ||
383 | buf = NULL; | ||
384 | desc_ptr = NULL; | ||
385 | len = 0; | ||
386 | page7_len = 0; | ||
387 | } else { | ||
388 | desc_ptr = buf + 8; | ||
389 | len = (desc_ptr[2] << 8) + desc_ptr[3]; | ||
390 | /* skip past overall descriptor */ | ||
391 | desc_ptr += len + 4; | ||
392 | if (ses_dev->page10) | ||
393 | addl_desc_ptr = ses_dev->page10 + 8; | ||
394 | } | ||
395 | type_ptr = ses_dev->page1 + 12 + ses_dev->page1[11]; | ||
396 | components = 0; | ||
397 | for (i = 0; i < types; i++, type_ptr += 4) { | ||
398 | for (j = 0; j < type_ptr[1]; j++) { | ||
399 | char *name = NULL; | ||
400 | struct enclosure_component *ecomp; | ||
401 | |||
402 | if (desc_ptr) { | ||
403 | if (desc_ptr >= buf + page7_len) { | ||
404 | desc_ptr = NULL; | ||
405 | } else { | ||
406 | len = (desc_ptr[2] << 8) + desc_ptr[3]; | ||
407 | desc_ptr += 4; | ||
408 | /* Add trailing zero - pushes into | ||
409 | * reserved space */ | ||
410 | desc_ptr[len] = '\0'; | ||
411 | name = desc_ptr; | ||
412 | } | ||
413 | } | ||
414 | if (type_ptr[0] == ENCLOSURE_COMPONENT_DEVICE || | ||
415 | type_ptr[0] == ENCLOSURE_COMPONENT_ARRAY_DEVICE) { | ||
416 | |||
417 | if (create) | ||
418 | ecomp = enclosure_component_register(edev, | ||
419 | components++, | ||
420 | type_ptr[0], | ||
421 | name); | ||
422 | else | ||
423 | ecomp = &edev->component[components++]; | ||
424 | |||
425 | if (!IS_ERR(ecomp) && addl_desc_ptr) | ||
426 | ses_process_descriptor(ecomp, | ||
427 | addl_desc_ptr); | ||
428 | } | ||
429 | if (desc_ptr) | ||
430 | desc_ptr += len; | ||
431 | |||
432 | if (addl_desc_ptr) | ||
433 | addl_desc_ptr += addl_desc_ptr[1] + 2; | ||
434 | |||
435 | } | ||
436 | } | ||
437 | kfree(buf); | ||
438 | kfree(hdr_buf); | ||
439 | } | ||
440 | |||
350 | static void ses_match_to_enclosure(struct enclosure_device *edev, | 441 | static void ses_match_to_enclosure(struct enclosure_device *edev, |
351 | struct scsi_device *sdev) | 442 | struct scsi_device *sdev) |
352 | { | 443 | { |
@@ -361,6 +452,8 @@ static void ses_match_to_enclosure(struct enclosure_device *edev, | |||
361 | if (!buf) | 452 | if (!buf) |
362 | return; | 453 | return; |
363 | 454 | ||
455 | ses_enclosure_data_process(edev, to_scsi_device(edev->edev.parent), 0); | ||
456 | |||
364 | vpd_len = ((buf[2] << 8) | buf[3]) + 4; | 457 | vpd_len = ((buf[2] << 8) | buf[3]) + 4; |
365 | 458 | ||
366 | desc = buf + 4; | 459 | desc = buf + 4; |
@@ -395,28 +488,26 @@ static void ses_match_to_enclosure(struct enclosure_device *edev, | |||
395 | kfree(buf); | 488 | kfree(buf); |
396 | } | 489 | } |
397 | 490 | ||
398 | #define INIT_ALLOC_SIZE 32 | ||
399 | |||
400 | static int ses_intf_add(struct device *cdev, | 491 | static int ses_intf_add(struct device *cdev, |
401 | struct class_interface *intf) | 492 | struct class_interface *intf) |
402 | { | 493 | { |
403 | struct scsi_device *sdev = to_scsi_device(cdev->parent); | 494 | struct scsi_device *sdev = to_scsi_device(cdev->parent); |
404 | struct scsi_device *tmp_sdev; | 495 | struct scsi_device *tmp_sdev; |
405 | unsigned char *buf = NULL, *hdr_buf, *type_ptr, *desc_ptr = NULL, | 496 | unsigned char *buf = NULL, *hdr_buf, *type_ptr; |
406 | *addl_desc_ptr = NULL; | ||
407 | struct ses_device *ses_dev; | 497 | struct ses_device *ses_dev; |
408 | u32 result; | 498 | u32 result; |
409 | int i, j, types, len, page7_len = 0, components = 0; | 499 | int i, types, len, components = 0; |
410 | int err = -ENOMEM; | 500 | int err = -ENOMEM; |
411 | struct enclosure_device *edev; | 501 | struct enclosure_device *edev; |
412 | struct ses_component *scomp = NULL; | 502 | struct ses_component *scomp = NULL; |
413 | 503 | ||
414 | if (!scsi_device_enclosure(sdev)) { | 504 | if (!scsi_device_enclosure(sdev)) { |
415 | /* not an enclosure, but might be in one */ | 505 | /* not an enclosure, but might be in one */ |
416 | edev = enclosure_find(&sdev->host->shost_gendev); | 506 | struct enclosure_device *prev = NULL; |
417 | if (edev) { | 507 | |
508 | while ((edev = enclosure_find(&sdev->host->shost_gendev, prev)) != NULL) { | ||
418 | ses_match_to_enclosure(edev, sdev); | 509 | ses_match_to_enclosure(edev, sdev); |
419 | put_device(&edev->edev); | 510 | prev = edev; |
420 | } | 511 | } |
421 | return -ENODEV; | 512 | return -ENODEV; |
422 | } | 513 | } |
@@ -500,6 +591,7 @@ static int ses_intf_add(struct device *cdev, | |||
500 | ses_dev->page10_len = len; | 591 | ses_dev->page10_len = len; |
501 | buf = NULL; | 592 | buf = NULL; |
502 | } | 593 | } |
594 | kfree(hdr_buf); | ||
503 | 595 | ||
504 | scomp = kzalloc(sizeof(struct ses_component) * components, GFP_KERNEL); | 596 | scomp = kzalloc(sizeof(struct ses_component) * components, GFP_KERNEL); |
505 | if (!scomp) | 597 | if (!scomp) |
@@ -516,72 +608,7 @@ static int ses_intf_add(struct device *cdev, | |||
516 | for (i = 0; i < components; i++) | 608 | for (i = 0; i < components; i++) |
517 | edev->component[i].scratch = scomp + i; | 609 | edev->component[i].scratch = scomp + i; |
518 | 610 | ||
519 | /* Page 7 for the descriptors is optional */ | 611 | ses_enclosure_data_process(edev, sdev, 1); |
520 | result = ses_recv_diag(sdev, 7, hdr_buf, INIT_ALLOC_SIZE); | ||
521 | if (result) | ||
522 | goto simple_populate; | ||
523 | |||
524 | page7_len = len = (hdr_buf[2] << 8) + hdr_buf[3] + 4; | ||
525 | /* add 1 for trailing '\0' we'll use */ | ||
526 | buf = kzalloc(len + 1, GFP_KERNEL); | ||
527 | if (!buf) | ||
528 | goto simple_populate; | ||
529 | result = ses_recv_diag(sdev, 7, buf, len); | ||
530 | if (result) { | ||
531 | simple_populate: | ||
532 | kfree(buf); | ||
533 | buf = NULL; | ||
534 | desc_ptr = NULL; | ||
535 | addl_desc_ptr = NULL; | ||
536 | } else { | ||
537 | desc_ptr = buf + 8; | ||
538 | len = (desc_ptr[2] << 8) + desc_ptr[3]; | ||
539 | /* skip past overall descriptor */ | ||
540 | desc_ptr += len + 4; | ||
541 | if (ses_dev->page10) | ||
542 | addl_desc_ptr = ses_dev->page10 + 8; | ||
543 | } | ||
544 | type_ptr = ses_dev->page1 + 12 + ses_dev->page1[11]; | ||
545 | components = 0; | ||
546 | for (i = 0; i < types; i++, type_ptr += 4) { | ||
547 | for (j = 0; j < type_ptr[1]; j++) { | ||
548 | char *name = NULL; | ||
549 | struct enclosure_component *ecomp; | ||
550 | |||
551 | if (desc_ptr) { | ||
552 | if (desc_ptr >= buf + page7_len) { | ||
553 | desc_ptr = NULL; | ||
554 | } else { | ||
555 | len = (desc_ptr[2] << 8) + desc_ptr[3]; | ||
556 | desc_ptr += 4; | ||
557 | /* Add trailing zero - pushes into | ||
558 | * reserved space */ | ||
559 | desc_ptr[len] = '\0'; | ||
560 | name = desc_ptr; | ||
561 | } | ||
562 | } | ||
563 | if (type_ptr[0] == ENCLOSURE_COMPONENT_DEVICE || | ||
564 | type_ptr[0] == ENCLOSURE_COMPONENT_ARRAY_DEVICE) { | ||
565 | |||
566 | ecomp = enclosure_component_register(edev, | ||
567 | components++, | ||
568 | type_ptr[0], | ||
569 | name); | ||
570 | |||
571 | if (!IS_ERR(ecomp) && addl_desc_ptr) | ||
572 | ses_process_descriptor(ecomp, | ||
573 | addl_desc_ptr); | ||
574 | } | ||
575 | if (desc_ptr) | ||
576 | desc_ptr += len; | ||
577 | |||
578 | if (addl_desc_ptr) | ||
579 | addl_desc_ptr += addl_desc_ptr[1] + 2; | ||
580 | |||
581 | } | ||
582 | } | ||
583 | kfree(buf); | ||
584 | kfree(hdr_buf); | ||
585 | 612 | ||
586 | /* see if there are any devices matching before | 613 | /* see if there are any devices matching before |
587 | * we found the enclosure */ | 614 | * we found the enclosure */ |
@@ -615,17 +642,26 @@ static int ses_remove(struct device *dev) | |||
615 | return 0; | 642 | return 0; |
616 | } | 643 | } |
617 | 644 | ||
618 | static void ses_intf_remove(struct device *cdev, | 645 | static void ses_intf_remove_component(struct scsi_device *sdev) |
619 | struct class_interface *intf) | 646 | { |
647 | struct enclosure_device *edev, *prev = NULL; | ||
648 | |||
649 | while ((edev = enclosure_find(&sdev->host->shost_gendev, prev)) != NULL) { | ||
650 | prev = edev; | ||
651 | if (!enclosure_remove_device(edev, &sdev->sdev_gendev)) | ||
652 | break; | ||
653 | } | ||
654 | if (edev) | ||
655 | put_device(&edev->edev); | ||
656 | } | ||
657 | |||
658 | static void ses_intf_remove_enclosure(struct scsi_device *sdev) | ||
620 | { | 659 | { |
621 | struct scsi_device *sdev = to_scsi_device(cdev->parent); | ||
622 | struct enclosure_device *edev; | 660 | struct enclosure_device *edev; |
623 | struct ses_device *ses_dev; | 661 | struct ses_device *ses_dev; |
624 | 662 | ||
625 | if (!scsi_device_enclosure(sdev)) | 663 | /* exact match to this enclosure */ |
626 | return; | 664 | edev = enclosure_find(&sdev->sdev_gendev, NULL); |
627 | |||
628 | edev = enclosure_find(cdev->parent); | ||
629 | if (!edev) | 665 | if (!edev) |
630 | return; | 666 | return; |
631 | 667 | ||
@@ -643,6 +679,17 @@ static void ses_intf_remove(struct device *cdev, | |||
643 | enclosure_unregister(edev); | 679 | enclosure_unregister(edev); |
644 | } | 680 | } |
645 | 681 | ||
682 | static void ses_intf_remove(struct device *cdev, | ||
683 | struct class_interface *intf) | ||
684 | { | ||
685 | struct scsi_device *sdev = to_scsi_device(cdev->parent); | ||
686 | |||
687 | if (!scsi_device_enclosure(sdev)) | ||
688 | ses_intf_remove_component(sdev); | ||
689 | else | ||
690 | ses_intf_remove_enclosure(sdev); | ||
691 | } | ||
692 | |||
646 | static struct class_interface ses_interface = { | 693 | static struct class_interface ses_interface = { |
647 | .add_dev = ses_intf_add, | 694 | .add_dev = ses_intf_add, |
648 | .remove_dev = ses_intf_remove, | 695 | .remove_dev = ses_intf_remove, |
diff --git a/drivers/scsi/sg.c b/drivers/scsi/sg.c index 9230402c45af..4968c4ced385 100644 --- a/drivers/scsi/sg.c +++ b/drivers/scsi/sg.c | |||
@@ -1811,7 +1811,7 @@ retry: | |||
1811 | return 0; | 1811 | return 0; |
1812 | out: | 1812 | out: |
1813 | for (i = 0; i < k; i++) | 1813 | for (i = 0; i < k; i++) |
1814 | __free_pages(schp->pages[k], order); | 1814 | __free_pages(schp->pages[i], order); |
1815 | 1815 | ||
1816 | if (--order >= 0) | 1816 | if (--order >= 0) |
1817 | goto retry; | 1817 | goto retry; |
diff --git a/drivers/scsi/stex.c b/drivers/scsi/stex.c index 8d2a95c4e5b5..09fa8861fc58 100644 --- a/drivers/scsi/stex.c +++ b/drivers/scsi/stex.c | |||
@@ -55,6 +55,7 @@ enum { | |||
55 | OIS = 0x30, /* MU_OUTBOUND_INTERRUPT_STATUS */ | 55 | OIS = 0x30, /* MU_OUTBOUND_INTERRUPT_STATUS */ |
56 | OIM = 0x3c, /* MU_OUTBOUND_INTERRUPT_MASK */ | 56 | OIM = 0x3c, /* MU_OUTBOUND_INTERRUPT_MASK */ |
57 | 57 | ||
58 | YIOA_STATUS = 0x00, | ||
58 | YH2I_INT = 0x20, | 59 | YH2I_INT = 0x20, |
59 | YINT_EN = 0x34, | 60 | YINT_EN = 0x34, |
60 | YI2H_INT = 0x9c, | 61 | YI2H_INT = 0x9c, |
@@ -108,6 +109,10 @@ enum { | |||
108 | 109 | ||
109 | SS_HEAD_HANDSHAKE = 0x80, | 110 | SS_HEAD_HANDSHAKE = 0x80, |
110 | 111 | ||
112 | SS_H2I_INT_RESET = 0x100, | ||
113 | |||
114 | SS_MU_OPERATIONAL = 0x80000000, | ||
115 | |||
111 | STEX_CDB_LENGTH = 16, | 116 | STEX_CDB_LENGTH = 16, |
112 | STATUS_VAR_LEN = 128, | 117 | STATUS_VAR_LEN = 128, |
113 | 118 | ||
@@ -884,7 +889,7 @@ static void stex_ss_mu_intr(struct st_hba *hba) | |||
884 | tag = (u16)value; | 889 | tag = (u16)value; |
885 | if (unlikely(tag >= hba->host->can_queue)) { | 890 | if (unlikely(tag >= hba->host->can_queue)) { |
886 | printk(KERN_WARNING DRV_NAME | 891 | printk(KERN_WARNING DRV_NAME |
887 | "(%s): invalid tag\n", pci_name(hba->pdev)); | 892 | "(%s): invalid tag\n", pci_name(hba->pdev)); |
888 | continue; | 893 | continue; |
889 | } | 894 | } |
890 | 895 | ||
@@ -1040,16 +1045,27 @@ static int stex_ss_handshake(struct st_hba *hba) | |||
1040 | void __iomem *base = hba->mmio_base; | 1045 | void __iomem *base = hba->mmio_base; |
1041 | struct st_msg_header *msg_h; | 1046 | struct st_msg_header *msg_h; |
1042 | struct handshake_frame *h; | 1047 | struct handshake_frame *h; |
1043 | __le32 *scratch = hba->scratch; | 1048 | __le32 *scratch; |
1044 | u32 data; | 1049 | u32 data; |
1045 | unsigned long before; | 1050 | unsigned long before; |
1046 | int ret = 0; | 1051 | int ret = 0; |
1047 | 1052 | ||
1048 | h = (struct handshake_frame *)(hba->alloc_rq(hba)); | 1053 | before = jiffies; |
1049 | msg_h = (struct st_msg_header *)h - 1; | 1054 | while ((readl(base + YIOA_STATUS) & SS_MU_OPERATIONAL) == 0) { |
1055 | if (time_after(jiffies, before + MU_MAX_DELAY * HZ)) { | ||
1056 | printk(KERN_ERR DRV_NAME | ||
1057 | "(%s): firmware not operational\n", | ||
1058 | pci_name(hba->pdev)); | ||
1059 | return -1; | ||
1060 | } | ||
1061 | msleep(1); | ||
1062 | } | ||
1063 | |||
1064 | msg_h = (struct st_msg_header *)hba->dma_mem; | ||
1050 | msg_h->handle = cpu_to_le64(hba->dma_handle); | 1065 | msg_h->handle = cpu_to_le64(hba->dma_handle); |
1051 | msg_h->flag = SS_HEAD_HANDSHAKE; | 1066 | msg_h->flag = SS_HEAD_HANDSHAKE; |
1052 | 1067 | ||
1068 | h = (struct handshake_frame *)(msg_h + 1); | ||
1053 | h->rb_phy = cpu_to_le64(hba->dma_handle); | 1069 | h->rb_phy = cpu_to_le64(hba->dma_handle); |
1054 | h->req_sz = cpu_to_le16(hba->rq_size); | 1070 | h->req_sz = cpu_to_le16(hba->rq_size); |
1055 | h->req_cnt = cpu_to_le16(hba->rq_count+1); | 1071 | h->req_cnt = cpu_to_le16(hba->rq_count+1); |
@@ -1205,6 +1221,13 @@ static void stex_hard_reset(struct st_hba *hba) | |||
1205 | hba->pdev->saved_config_space[i]); | 1221 | hba->pdev->saved_config_space[i]); |
1206 | } | 1222 | } |
1207 | 1223 | ||
1224 | static void stex_ss_reset(struct st_hba *hba) | ||
1225 | { | ||
1226 | writel(SS_H2I_INT_RESET, hba->mmio_base + YH2I_INT); | ||
1227 | readl(hba->mmio_base + YH2I_INT); | ||
1228 | ssleep(5); | ||
1229 | } | ||
1230 | |||
1208 | static int stex_reset(struct scsi_cmnd *cmd) | 1231 | static int stex_reset(struct scsi_cmnd *cmd) |
1209 | { | 1232 | { |
1210 | struct st_hba *hba; | 1233 | struct st_hba *hba; |
@@ -1221,6 +1244,8 @@ static int stex_reset(struct scsi_cmnd *cmd) | |||
1221 | 1244 | ||
1222 | if (hba->cardtype == st_shasta) | 1245 | if (hba->cardtype == st_shasta) |
1223 | stex_hard_reset(hba); | 1246 | stex_hard_reset(hba); |
1247 | else if (hba->cardtype == st_yel) | ||
1248 | stex_ss_reset(hba); | ||
1224 | 1249 | ||
1225 | if (hba->cardtype != st_yosemite) { | 1250 | if (hba->cardtype != st_yosemite) { |
1226 | if (stex_handshake(hba)) { | 1251 | if (stex_handshake(hba)) { |