aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--drivers/s390/scsi/zfcp_aux.c45
-rw-r--r--drivers/s390/scsi/zfcp_dbf.c3
-rw-r--r--drivers/s390/scsi/zfcp_def.h26
-rw-r--r--drivers/s390/scsi/zfcp_erp.c114
-rw-r--r--drivers/s390/scsi/zfcp_ext.h18
-rw-r--r--drivers/s390/scsi/zfcp_fsf.c178
-rw-r--r--drivers/s390/scsi/zfcp_qdio.c815
7 files changed, 412 insertions, 787 deletions
diff --git a/drivers/s390/scsi/zfcp_aux.c b/drivers/s390/scsi/zfcp_aux.c
index 735f7af43d6a..7084a6ae1096 100644
--- a/drivers/s390/scsi/zfcp_aux.c
+++ b/drivers/s390/scsi/zfcp_aux.c
@@ -606,7 +606,6 @@ static void _zfcp_status_read_scheduler(struct work_struct *work)
606struct zfcp_adapter * 606struct zfcp_adapter *
607zfcp_adapter_enqueue(struct ccw_device *ccw_device) 607zfcp_adapter_enqueue(struct ccw_device *ccw_device)
608{ 608{
609 int retval = 0;
610 struct zfcp_adapter *adapter; 609 struct zfcp_adapter *adapter;
611 610
612 /* 611 /*
@@ -627,19 +626,11 @@ zfcp_adapter_enqueue(struct ccw_device *ccw_device)
627 /* save ccw_device pointer */ 626 /* save ccw_device pointer */
628 adapter->ccw_device = ccw_device; 627 adapter->ccw_device = ccw_device;
629 628
630 retval = zfcp_qdio_allocate_queues(adapter); 629 if (zfcp_qdio_allocate(adapter))
631 if (retval)
632 goto queues_alloc_failed;
633
634 retval = zfcp_qdio_allocate(adapter);
635 if (retval)
636 goto qdio_allocate_failed; 630 goto qdio_allocate_failed;
637 631
638 retval = zfcp_allocate_low_mem_buffers(adapter); 632 if (zfcp_allocate_low_mem_buffers(adapter))
639 if (retval) {
640 ZFCP_LOG_INFO("error: pool allocation failed\n");
641 goto failed_low_mem_buffers; 633 goto failed_low_mem_buffers;
642 }
643 634
644 /* initialise reference count stuff */ 635 /* initialise reference count stuff */
645 atomic_set(&adapter->refcount, 0); 636 atomic_set(&adapter->refcount, 0);
@@ -653,11 +644,8 @@ zfcp_adapter_enqueue(struct ccw_device *ccw_device)
653 644
654 /* initialize list of fsf requests */ 645 /* initialize list of fsf requests */
655 spin_lock_init(&adapter->req_list_lock); 646 spin_lock_init(&adapter->req_list_lock);
656 retval = zfcp_reqlist_alloc(adapter); 647 if (zfcp_reqlist_alloc(adapter))
657 if (retval) {
658 ZFCP_LOG_INFO("request list initialization failed\n");
659 goto failed_low_mem_buffers; 648 goto failed_low_mem_buffers;
660 }
661 649
662 /* initialize debug locks */ 650 /* initialize debug locks */
663 651
@@ -666,8 +654,7 @@ zfcp_adapter_enqueue(struct ccw_device *ccw_device)
666 spin_lock_init(&adapter->scsi_dbf_lock); 654 spin_lock_init(&adapter->scsi_dbf_lock);
667 spin_lock_init(&adapter->rec_dbf_lock); 655 spin_lock_init(&adapter->rec_dbf_lock);
668 656
669 retval = zfcp_adapter_debug_register(adapter); 657 if (zfcp_adapter_debug_register(adapter))
670 if (retval)
671 goto debug_register_failed; 658 goto debug_register_failed;
672 659
673 /* initialize error recovery stuff */ 660 /* initialize error recovery stuff */
@@ -685,7 +672,7 @@ zfcp_adapter_enqueue(struct ccw_device *ccw_device)
685 init_waitqueue_head(&adapter->erp_done_wqh); 672 init_waitqueue_head(&adapter->erp_done_wqh);
686 673
687 /* initialize lock of associated request queue */ 674 /* initialize lock of associated request queue */
688 rwlock_init(&adapter->request_queue.queue_lock); 675 rwlock_init(&adapter->req_q.lock);
689 INIT_WORK(&adapter->stat_work, _zfcp_status_read_scheduler); 676 INIT_WORK(&adapter->stat_work, _zfcp_status_read_scheduler);
690 677
691 /* mark adapter unusable as long as sysfs registration is not complete */ 678 /* mark adapter unusable as long as sysfs registration is not complete */
@@ -723,12 +710,8 @@ zfcp_adapter_enqueue(struct ccw_device *ccw_device)
723 zfcp_reqlist_free(adapter); 710 zfcp_reqlist_free(adapter);
724 failed_low_mem_buffers: 711 failed_low_mem_buffers:
725 zfcp_free_low_mem_buffers(adapter); 712 zfcp_free_low_mem_buffers(adapter);
726 if (qdio_free(ccw_device) != 0)
727 ZFCP_LOG_NORMAL("bug: qdio_free for adapter %s failed\n",
728 zfcp_get_busid_by_adapter(adapter));
729 qdio_allocate_failed: 713 qdio_allocate_failed:
730 zfcp_qdio_free_queues(adapter); 714 zfcp_qdio_free(adapter);
731 queues_alloc_failed:
732 kfree(adapter); 715 kfree(adapter);
733 adapter = NULL; 716 adapter = NULL;
734 out: 717 out:
@@ -757,10 +740,6 @@ zfcp_adapter_dequeue(struct zfcp_adapter *adapter)
757 retval = zfcp_reqlist_isempty(adapter); 740 retval = zfcp_reqlist_isempty(adapter);
758 spin_unlock_irqrestore(&adapter->req_list_lock, flags); 741 spin_unlock_irqrestore(&adapter->req_list_lock, flags);
759 if (!retval) { 742 if (!retval) {
760 ZFCP_LOG_NORMAL("bug: adapter %s (%p) still in use, "
761 "%i requests outstanding\n",
762 zfcp_get_busid_by_adapter(adapter), adapter,
763 atomic_read(&adapter->reqs_active));
764 retval = -EBUSY; 743 retval = -EBUSY;
765 goto out; 744 goto out;
766 } 745 }
@@ -775,19 +754,9 @@ zfcp_adapter_dequeue(struct zfcp_adapter *adapter)
775 /* decrease number of adapters in list */ 754 /* decrease number of adapters in list */
776 zfcp_data.adapters--; 755 zfcp_data.adapters--;
777 756
778 ZFCP_LOG_TRACE("adapter %s (%p) removed from list, " 757 zfcp_qdio_free(adapter);
779 "%i adapters still in list\n",
780 zfcp_get_busid_by_adapter(adapter),
781 adapter, zfcp_data.adapters);
782
783 retval = qdio_free(adapter->ccw_device);
784 if (retval)
785 ZFCP_LOG_NORMAL("bug: qdio_free for adapter %s failed\n",
786 zfcp_get_busid_by_adapter(adapter));
787 758
788 zfcp_free_low_mem_buffers(adapter); 759 zfcp_free_low_mem_buffers(adapter);
789 /* free memory of adapter data structure and queues */
790 zfcp_qdio_free_queues(adapter);
791 zfcp_reqlist_free(adapter); 760 zfcp_reqlist_free(adapter);
792 kfree(adapter->fc_stats); 761 kfree(adapter->fc_stats);
793 kfree(adapter->stats_reset_data); 762 kfree(adapter->stats_reset_data);
diff --git a/drivers/s390/scsi/zfcp_dbf.c b/drivers/s390/scsi/zfcp_dbf.c
index 01e817abe0a5..c47c23a01c7f 100644
--- a/drivers/s390/scsi/zfcp_dbf.c
+++ b/drivers/s390/scsi/zfcp_dbf.c
@@ -603,13 +603,14 @@ static const char *zfcp_rec_dbf_ids[] = {
603 [137] = "hbaapi port open", 603 [137] = "hbaapi port open",
604 [138] = "hbaapi unit open", 604 [138] = "hbaapi unit open",
605 [139] = "hbaapi unit shutdown", 605 [139] = "hbaapi unit shutdown",
606 [140] = "qdio error", 606 [140] = "qdio error outbound",
607 [141] = "scsi host reset", 607 [141] = "scsi host reset",
608 [142] = "dismissing fsf request for recovery action", 608 [142] = "dismissing fsf request for recovery action",
609 [143] = "recovery action timed out", 609 [143] = "recovery action timed out",
610 [144] = "recovery action gone", 610 [144] = "recovery action gone",
611 [145] = "recovery action being processed", 611 [145] = "recovery action being processed",
612 [146] = "recovery action ready for next step", 612 [146] = "recovery action ready for next step",
613 [147] = "qdio error inbound",
613}; 614};
614 615
615static int zfcp_rec_dbf_view_format(debug_info_t *id, struct debug_view *view, 616static int zfcp_rec_dbf_view_format(debug_info_t *id, struct debug_view *view,
diff --git a/drivers/s390/scsi/zfcp_def.h b/drivers/s390/scsi/zfcp_def.h
index 72f225817ebd..5fcb555e1484 100644
--- a/drivers/s390/scsi/zfcp_def.h
+++ b/drivers/s390/scsi/zfcp_def.h
@@ -112,21 +112,10 @@ zfcp_address_to_sg(void *address, struct scatterlist *list, unsigned int size)
112 /* max. number of (data buffer) SBALEs in largest SBAL chain 112 /* max. number of (data buffer) SBALEs in largest SBAL chain
113 multiplied with number of sectors per 4k block */ 113 multiplied with number of sectors per 4k block */
114 114
115/* FIXME(tune): free space should be one max. SBAL chain plus what? */
116#define ZFCP_QDIO_PCI_INTERVAL (QDIO_MAX_BUFFERS_PER_Q \
117 - (ZFCP_MAX_SBALS_PER_REQ + 4))
118
119#define ZFCP_SBAL_TIMEOUT (5*HZ) 115#define ZFCP_SBAL_TIMEOUT (5*HZ)
120 116
121#define ZFCP_TYPE2_RECOVERY_TIME 8 /* seconds */ 117#define ZFCP_TYPE2_RECOVERY_TIME 8 /* seconds */
122 118
123/* queue polling (values in microseconds) */
124#define ZFCP_MAX_INPUT_THRESHOLD 5000 /* FIXME: tune */
125#define ZFCP_MAX_OUTPUT_THRESHOLD 1000 /* FIXME: tune */
126#define ZFCP_MIN_INPUT_THRESHOLD 1 /* ignored by QDIO layer */
127#define ZFCP_MIN_OUTPUT_THRESHOLD 1 /* ignored by QDIO layer */
128
129#define QDIO_SCSI_QFMT 1 /* 1 for FSF */
130#define QBUFF_PER_PAGE (PAGE_SIZE / sizeof(struct qdio_buffer)) 119#define QBUFF_PER_PAGE (PAGE_SIZE / sizeof(struct qdio_buffer))
131 120
132/********************* FSF SPECIFIC DEFINES *********************************/ 121/********************* FSF SPECIFIC DEFINES *********************************/
@@ -649,13 +638,13 @@ struct zfcp_send_els {
649}; 638};
650 639
651struct zfcp_qdio_queue { 640struct zfcp_qdio_queue {
652 struct qdio_buffer *buffer[QDIO_MAX_BUFFERS_PER_Q]; /* SBALs */ 641 struct qdio_buffer *sbal[QDIO_MAX_BUFFERS_PER_Q]; /* SBALs */
653 u8 free_index; /* index of next free bfr 642 u8 first; /* index of next free bfr
654 in queue (free_count>0) */ 643 in queue (free_count>0) */
655 atomic_t free_count; /* number of free buffers 644 atomic_t count; /* number of free buffers
656 in queue */ 645 in queue */
657 rwlock_t queue_lock; /* lock for operations on queue */ 646 rwlock_t lock; /* lock for operations on queue */
658 int distance_from_int; /* SBALs used since PCI indication 647 int pci_batch; /* SBALs since PCI indication
659 was last set */ 648 was last set */
660}; 649};
661 650
@@ -711,15 +700,14 @@ struct zfcp_adapter {
711 struct list_head port_remove_lh; /* head of ports to be 700 struct list_head port_remove_lh; /* head of ports to be
712 removed */ 701 removed */
713 u32 ports; /* number of remote ports */ 702 u32 ports; /* number of remote ports */
714 atomic_t reqs_active; /* # active FSF reqs */
715 unsigned long req_no; /* unique FSF req number */ 703 unsigned long req_no; /* unique FSF req number */
716 struct list_head *req_list; /* list of pending reqs */ 704 struct list_head *req_list; /* list of pending reqs */
717 spinlock_t req_list_lock; /* request list lock */ 705 spinlock_t req_list_lock; /* request list lock */
718 struct zfcp_qdio_queue request_queue; /* request queue */ 706 struct zfcp_qdio_queue req_q; /* request queue */
719 u32 fsf_req_seq_no; /* FSF cmnd seq number */ 707 u32 fsf_req_seq_no; /* FSF cmnd seq number */
720 wait_queue_head_t request_wq; /* can be used to wait for 708 wait_queue_head_t request_wq; /* can be used to wait for
721 more avaliable SBALs */ 709 more avaliable SBALs */
722 struct zfcp_qdio_queue response_queue; /* response queue */ 710 struct zfcp_qdio_queue resp_q; /* response queue */
723 rwlock_t abort_lock; /* Protects against SCSI 711 rwlock_t abort_lock; /* Protects against SCSI
724 stack abort/command 712 stack abort/command
725 completion races */ 713 completion races */
diff --git a/drivers/s390/scsi/zfcp_erp.c b/drivers/s390/scsi/zfcp_erp.c
index ee19be13e708..4a6d08363d4b 100644
--- a/drivers/s390/scsi/zfcp_erp.c
+++ b/drivers/s390/scsi/zfcp_erp.c
@@ -114,41 +114,6 @@ static void zfcp_erp_action_to_running(struct zfcp_erp_action *);
114static void zfcp_erp_memwait_handler(unsigned long); 114static void zfcp_erp_memwait_handler(unsigned long);
115 115
116/** 116/**
117 * zfcp_close_qdio - close qdio queues for an adapter
118 */
119static void zfcp_close_qdio(struct zfcp_adapter *adapter)
120{
121 struct zfcp_qdio_queue *req_queue;
122 int first, count;
123
124 if (!atomic_test_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status))
125 return;
126
127 /* clear QDIOUP flag, thus do_QDIO is not called during qdio_shutdown */
128 req_queue = &adapter->request_queue;
129 write_lock_irq(&req_queue->queue_lock);
130 atomic_clear_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status);
131 write_unlock_irq(&req_queue->queue_lock);
132
133 while (qdio_shutdown(adapter->ccw_device,
134 QDIO_FLAG_CLEANUP_USING_CLEAR) == -EINPROGRESS)
135 ssleep(1);
136
137 /* cleanup used outbound sbals */
138 count = atomic_read(&req_queue->free_count);
139 if (count < QDIO_MAX_BUFFERS_PER_Q) {
140 first = (req_queue->free_index+count) % QDIO_MAX_BUFFERS_PER_Q;
141 count = QDIO_MAX_BUFFERS_PER_Q - count;
142 zfcp_qdio_zero_sbals(req_queue->buffer, first, count);
143 }
144 req_queue->free_index = 0;
145 atomic_set(&req_queue->free_count, 0);
146 req_queue->distance_from_int = 0;
147 adapter->response_queue.free_index = 0;
148 atomic_set(&adapter->response_queue.free_count, 0);
149}
150
151/**
152 * zfcp_close_fsf - stop FSF operations for an adapter 117 * zfcp_close_fsf - stop FSF operations for an adapter
153 * 118 *
154 * Dismiss and cleanup all pending fsf_reqs (this wakes up all initiators of 119 * Dismiss and cleanup all pending fsf_reqs (this wakes up all initiators of
@@ -158,7 +123,7 @@ static void zfcp_close_qdio(struct zfcp_adapter *adapter)
158static void zfcp_close_fsf(struct zfcp_adapter *adapter) 123static void zfcp_close_fsf(struct zfcp_adapter *adapter)
159{ 124{
160 /* close queues to ensure that buffers are not accessed by adapter */ 125 /* close queues to ensure that buffers are not accessed by adapter */
161 zfcp_close_qdio(adapter); 126 zfcp_qdio_close(adapter);
162 zfcp_fsf_req_dismiss_all(adapter); 127 zfcp_fsf_req_dismiss_all(adapter);
163 /* reset FSF request sequence number */ 128 /* reset FSF request sequence number */
164 adapter->fsf_req_seq_no = 0; 129 adapter->fsf_req_seq_no = 0;
@@ -1735,88 +1700,17 @@ zfcp_erp_adapter_strategy_generic(struct zfcp_erp_action *erp_action, int close)
1735static int 1700static int
1736zfcp_erp_adapter_strategy_open_qdio(struct zfcp_erp_action *erp_action) 1701zfcp_erp_adapter_strategy_open_qdio(struct zfcp_erp_action *erp_action)
1737{ 1702{
1738 int retval;
1739 int i;
1740 volatile struct qdio_buffer_element *sbale;
1741 struct zfcp_adapter *adapter = erp_action->adapter; 1703 struct zfcp_adapter *adapter = erp_action->adapter;
1742 1704
1743 if (atomic_test_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status)) { 1705 if (zfcp_qdio_open(adapter))
1744 ZFCP_LOG_NORMAL("bug: second attempt to set up QDIO on " 1706 return ZFCP_ERP_FAILED;
1745 "adapter %s\n",
1746 zfcp_get_busid_by_adapter(adapter));
1747 goto failed_sanity;
1748 }
1749
1750 if (qdio_establish(&adapter->qdio_init_data) != 0) {
1751 ZFCP_LOG_INFO("error: establishment of QDIO queues failed "
1752 "on adapter %s\n",
1753 zfcp_get_busid_by_adapter(adapter));
1754 goto failed_qdio_establish;
1755 }
1756
1757 if (qdio_activate(adapter->ccw_device, 0) != 0) {
1758 ZFCP_LOG_INFO("error: activation of QDIO queues failed "
1759 "on adapter %s\n",
1760 zfcp_get_busid_by_adapter(adapter));
1761 goto failed_qdio_activate;
1762 }
1763
1764 /*
1765 * put buffers into response queue,
1766 */
1767 for (i = 0; i < QDIO_MAX_BUFFERS_PER_Q; i++) {
1768 sbale = &(adapter->response_queue.buffer[i]->element[0]);
1769 sbale->length = 0;
1770 sbale->flags = SBAL_FLAGS_LAST_ENTRY;
1771 sbale->addr = NULL;
1772 }
1773
1774 ZFCP_LOG_TRACE("calling do_QDIO on adapter %s (flags=0x%x, "
1775 "queue_no=%i, index_in_queue=%i, count=%i)\n",
1776 zfcp_get_busid_by_adapter(adapter),
1777 QDIO_FLAG_SYNC_INPUT, 0, 0, QDIO_MAX_BUFFERS_PER_Q);
1778
1779 retval = do_QDIO(adapter->ccw_device,
1780 QDIO_FLAG_SYNC_INPUT,
1781 0, 0, QDIO_MAX_BUFFERS_PER_Q, NULL);
1782
1783 if (retval) {
1784 ZFCP_LOG_NORMAL("bug: setup of QDIO failed (retval=%d)\n",
1785 retval);
1786 goto failed_do_qdio;
1787 } else {
1788 adapter->response_queue.free_index = 0;
1789 atomic_set(&adapter->response_queue.free_count, 0);
1790 ZFCP_LOG_DEBUG("%i buffers successfully enqueued to "
1791 "response queue\n", QDIO_MAX_BUFFERS_PER_Q);
1792 }
1793 /* set index of first avalable SBALS / number of available SBALS */
1794 adapter->request_queue.free_index = 0;
1795 atomic_set(&adapter->request_queue.free_count, QDIO_MAX_BUFFERS_PER_Q);
1796 adapter->request_queue.distance_from_int = 0;
1797 1707
1798 /* initialize waitqueue used to wait for free SBALs in requests queue */ 1708 /* initialize waitqueue used to wait for free SBALs in requests queue */
1799 init_waitqueue_head(&adapter->request_wq); 1709 init_waitqueue_head(&adapter->request_wq);
1800 1710
1801 /* ok, we did it - skip all cleanups for different failures */ 1711 /* ok, we did it - skip all cleanups for different failures */
1802 atomic_set_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status); 1712 atomic_set_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status);
1803 retval = ZFCP_ERP_SUCCEEDED; 1713 return ZFCP_ERP_SUCCEEDED;
1804 goto out;
1805
1806 failed_do_qdio:
1807 /* NOP */
1808
1809 failed_qdio_activate:
1810 while (qdio_shutdown(adapter->ccw_device,
1811 QDIO_FLAG_CLEANUP_USING_CLEAR) == -EINPROGRESS)
1812 ssleep(1);
1813
1814 failed_qdio_establish:
1815 failed_sanity:
1816 retval = ZFCP_ERP_FAILED;
1817
1818 out:
1819 return retval;
1820} 1714}
1821 1715
1822 1716
diff --git a/drivers/s390/scsi/zfcp_ext.h b/drivers/s390/scsi/zfcp_ext.h
index 867972898cb1..6ffe2068ba8a 100644
--- a/drivers/s390/scsi/zfcp_ext.h
+++ b/drivers/s390/scsi/zfcp_ext.h
@@ -57,21 +57,17 @@ extern int zfcp_ccw_register(void);
57 57
58extern void zfcp_qdio_zero_sbals(struct qdio_buffer **, int, int); 58extern void zfcp_qdio_zero_sbals(struct qdio_buffer **, int, int);
59extern int zfcp_qdio_allocate(struct zfcp_adapter *); 59extern int zfcp_qdio_allocate(struct zfcp_adapter *);
60extern int zfcp_qdio_allocate_queues(struct zfcp_adapter *); 60extern void zfcp_qdio_free(struct zfcp_adapter *);
61extern void zfcp_qdio_free_queues(struct zfcp_adapter *); 61extern int zfcp_qdio_send(struct zfcp_fsf_req *fsf_req);
62extern int zfcp_qdio_determine_pci(struct zfcp_qdio_queue *,
63 struct zfcp_fsf_req *);
64 62
65extern volatile struct qdio_buffer_element *zfcp_qdio_sbale_req 63extern volatile struct qdio_buffer_element *zfcp_qdio_sbale_req
66 (struct zfcp_fsf_req *, int, int); 64 (struct zfcp_fsf_req *);
67extern volatile struct qdio_buffer_element *zfcp_qdio_sbale_curr 65extern volatile struct qdio_buffer_element *zfcp_qdio_sbale_curr
68 (struct zfcp_fsf_req *); 66 (struct zfcp_fsf_req *);
69extern int zfcp_qdio_sbals_from_sg 67extern int zfcp_qdio_sbals_from_sg
70 (struct zfcp_fsf_req *, unsigned long, struct scatterlist *, int, int); 68 (struct zfcp_fsf_req *, unsigned long, struct scatterlist *, int);
71extern int zfcp_qdio_sbals_from_scsicmnd 69extern int zfcp_qdio_open(struct zfcp_adapter *adapter);
72 (struct zfcp_fsf_req *, unsigned long, struct scsi_cmnd *); 70extern void zfcp_qdio_close(struct zfcp_adapter *adapter);
73
74
75/******************************** FSF ****************************************/ 71/******************************** FSF ****************************************/
76extern int zfcp_fsf_open_port(struct zfcp_erp_action *); 72extern int zfcp_fsf_open_port(struct zfcp_erp_action *);
77extern int zfcp_fsf_close_port(struct zfcp_erp_action *); 73extern int zfcp_fsf_close_port(struct zfcp_erp_action *);
@@ -95,7 +91,7 @@ extern int zfcp_fsf_status_read(struct zfcp_adapter *, int);
95extern int zfcp_status_read_refill(struct zfcp_adapter *adapter); 91extern int zfcp_status_read_refill(struct zfcp_adapter *adapter);
96extern int zfcp_fsf_req_create(struct zfcp_adapter *, u32, int, mempool_t *, 92extern int zfcp_fsf_req_create(struct zfcp_adapter *, u32, int, mempool_t *,
97 unsigned long *, struct zfcp_fsf_req **) 93 unsigned long *, struct zfcp_fsf_req **)
98 __acquires(adapter->request_queue.queue_lock); 94 __acquires(adapter->req_q.lock);
99extern int zfcp_fsf_send_ct(struct zfcp_send_ct *, mempool_t *, 95extern int zfcp_fsf_send_ct(struct zfcp_send_ct *, mempool_t *,
100 struct zfcp_erp_action *); 96 struct zfcp_erp_action *);
101extern int zfcp_fsf_send_els(struct zfcp_send_els *); 97extern int zfcp_fsf_send_els(struct zfcp_send_els *);
diff --git a/drivers/s390/scsi/zfcp_fsf.c b/drivers/s390/scsi/zfcp_fsf.c
index de42a01fc4b1..cc48a6462e6c 100644
--- a/drivers/s390/scsi/zfcp_fsf.c
+++ b/drivers/s390/scsi/zfcp_fsf.c
@@ -171,7 +171,6 @@ void zfcp_fsf_req_dismiss_all(struct zfcp_adapter *adapter)
171 171
172 BUG_ON(atomic_test_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status)); 172 BUG_ON(atomic_test_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status));
173 spin_lock_irqsave(&adapter->req_list_lock, flags); 173 spin_lock_irqsave(&adapter->req_list_lock, flags);
174 atomic_set(&adapter->reqs_active, 0);
175 for (i = 0; i < REQUEST_LIST_SIZE; i++) 174 for (i = 0; i < REQUEST_LIST_SIZE; i++)
176 list_splice_init(&adapter->req_list[i], &remove_queue); 175 list_splice_init(&adapter->req_list[i], &remove_queue);
177 spin_unlock_irqrestore(&adapter->req_list_lock, flags); 176 spin_unlock_irqrestore(&adapter->req_list_lock, flags);
@@ -726,7 +725,7 @@ zfcp_fsf_status_read(struct zfcp_adapter *adapter, int req_flags)
726 goto failed_req_create; 725 goto failed_req_create;
727 } 726 }
728 727
729 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 728 sbale = zfcp_qdio_sbale_req(fsf_req);
730 sbale[0].flags |= SBAL_FLAGS0_TYPE_STATUS; 729 sbale[0].flags |= SBAL_FLAGS0_TYPE_STATUS;
731 sbale[2].flags |= SBAL_FLAGS_LAST_ENTRY; 730 sbale[2].flags |= SBAL_FLAGS_LAST_ENTRY;
732 fsf_req->sbale_curr = 2; 731 fsf_req->sbale_curr = 2;
@@ -763,7 +762,7 @@ zfcp_fsf_status_read(struct zfcp_adapter *adapter, int req_flags)
763 failed_req_create: 762 failed_req_create:
764 zfcp_hba_dbf_event_fsf_unsol("fail", adapter, NULL); 763 zfcp_hba_dbf_event_fsf_unsol("fail", adapter, NULL);
765 out: 764 out:
766 write_unlock_irqrestore(&adapter->request_queue.queue_lock, lock_flags); 765 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
767 return retval; 766 return retval;
768} 767}
769 768
@@ -1075,7 +1074,7 @@ zfcp_fsf_abort_fcp_command(unsigned long old_req_id,
1075 &unit->status))) 1074 &unit->status)))
1076 goto unit_blocked; 1075 goto unit_blocked;
1077 1076
1078 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 1077 sbale = zfcp_qdio_sbale_req(fsf_req);
1079 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; 1078 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ;
1080 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; 1079 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY;
1081 1080
@@ -1098,7 +1097,7 @@ zfcp_fsf_abort_fcp_command(unsigned long old_req_id,
1098 fsf_req = NULL; 1097 fsf_req = NULL;
1099 1098
1100 out: 1099 out:
1101 write_unlock_irqrestore(&adapter->request_queue.queue_lock, lock_flags); 1100 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
1102 return fsf_req; 1101 return fsf_req;
1103} 1102}
1104 1103
@@ -1295,7 +1294,7 @@ zfcp_fsf_send_ct(struct zfcp_send_ct *ct, mempool_t *pool,
1295 goto failed_req; 1294 goto failed_req;
1296 } 1295 }
1297 1296
1298 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 1297 sbale = zfcp_qdio_sbale_req(fsf_req);
1299 if (zfcp_use_one_sbal(ct->req, ct->req_count, 1298 if (zfcp_use_one_sbal(ct->req, ct->req_count,
1300 ct->resp, ct->resp_count)){ 1299 ct->resp, ct->resp_count)){
1301 /* both request buffer and response buffer 1300 /* both request buffer and response buffer
@@ -1311,7 +1310,7 @@ zfcp_fsf_send_ct(struct zfcp_send_ct *ct, mempool_t *pool,
1311 /* try to use chained SBALs */ 1310 /* try to use chained SBALs */
1312 bytes = zfcp_qdio_sbals_from_sg(fsf_req, 1311 bytes = zfcp_qdio_sbals_from_sg(fsf_req,
1313 SBAL_FLAGS0_TYPE_WRITE_READ, 1312 SBAL_FLAGS0_TYPE_WRITE_READ,
1314 ct->req, ct->req_count, 1313 ct->req,
1315 ZFCP_MAX_SBALS_PER_CT_REQ); 1314 ZFCP_MAX_SBALS_PER_CT_REQ);
1316 if (bytes <= 0) { 1315 if (bytes <= 0) {
1317 ZFCP_LOG_INFO("error: creation of CT request failed " 1316 ZFCP_LOG_INFO("error: creation of CT request failed "
@@ -1328,7 +1327,7 @@ zfcp_fsf_send_ct(struct zfcp_send_ct *ct, mempool_t *pool,
1328 fsf_req->sbale_curr = ZFCP_LAST_SBALE_PER_SBAL; 1327 fsf_req->sbale_curr = ZFCP_LAST_SBALE_PER_SBAL;
1329 bytes = zfcp_qdio_sbals_from_sg(fsf_req, 1328 bytes = zfcp_qdio_sbals_from_sg(fsf_req,
1330 SBAL_FLAGS0_TYPE_WRITE_READ, 1329 SBAL_FLAGS0_TYPE_WRITE_READ,
1331 ct->resp, ct->resp_count, 1330 ct->resp,
1332 ZFCP_MAX_SBALS_PER_CT_REQ); 1331 ZFCP_MAX_SBALS_PER_CT_REQ);
1333 if (bytes <= 0) { 1332 if (bytes <= 0) {
1334 ZFCP_LOG_INFO("error: creation of CT request failed " 1333 ZFCP_LOG_INFO("error: creation of CT request failed "
@@ -1387,8 +1386,7 @@ zfcp_fsf_send_ct(struct zfcp_send_ct *ct, mempool_t *pool,
1387 } 1386 }
1388 failed_req: 1387 failed_req:
1389 out: 1388 out:
1390 write_unlock_irqrestore(&adapter->request_queue.queue_lock, 1389 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
1391 lock_flags);
1392 return ret; 1390 return ret;
1393} 1391}
1394 1392
@@ -1593,7 +1591,7 @@ zfcp_fsf_send_els(struct zfcp_send_els *els)
1593 goto port_blocked; 1591 goto port_blocked;
1594 } 1592 }
1595 1593
1596 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 1594 sbale = zfcp_qdio_sbale_req(fsf_req);
1597 if (zfcp_use_one_sbal(els->req, els->req_count, 1595 if (zfcp_use_one_sbal(els->req, els->req_count,
1598 els->resp, els->resp_count)){ 1596 els->resp, els->resp_count)){
1599 /* both request buffer and response buffer 1597 /* both request buffer and response buffer
@@ -1609,7 +1607,7 @@ zfcp_fsf_send_els(struct zfcp_send_els *els)
1609 /* try to use chained SBALs */ 1607 /* try to use chained SBALs */
1610 bytes = zfcp_qdio_sbals_from_sg(fsf_req, 1608 bytes = zfcp_qdio_sbals_from_sg(fsf_req,
1611 SBAL_FLAGS0_TYPE_WRITE_READ, 1609 SBAL_FLAGS0_TYPE_WRITE_READ,
1612 els->req, els->req_count, 1610 els->req,
1613 ZFCP_MAX_SBALS_PER_ELS_REQ); 1611 ZFCP_MAX_SBALS_PER_ELS_REQ);
1614 if (bytes <= 0) { 1612 if (bytes <= 0) {
1615 ZFCP_LOG_INFO("error: creation of ELS request failed " 1613 ZFCP_LOG_INFO("error: creation of ELS request failed "
@@ -1626,7 +1624,7 @@ zfcp_fsf_send_els(struct zfcp_send_els *els)
1626 fsf_req->sbale_curr = ZFCP_LAST_SBALE_PER_SBAL; 1624 fsf_req->sbale_curr = ZFCP_LAST_SBALE_PER_SBAL;
1627 bytes = zfcp_qdio_sbals_from_sg(fsf_req, 1625 bytes = zfcp_qdio_sbals_from_sg(fsf_req,
1628 SBAL_FLAGS0_TYPE_WRITE_READ, 1626 SBAL_FLAGS0_TYPE_WRITE_READ,
1629 els->resp, els->resp_count, 1627 els->resp,
1630 ZFCP_MAX_SBALS_PER_ELS_REQ); 1628 ZFCP_MAX_SBALS_PER_ELS_REQ);
1631 if (bytes <= 0) { 1629 if (bytes <= 0) {
1632 ZFCP_LOG_INFO("error: creation of ELS request failed " 1630 ZFCP_LOG_INFO("error: creation of ELS request failed "
@@ -1657,7 +1655,7 @@ zfcp_fsf_send_els(struct zfcp_send_els *els)
1657 fsf_req->qtcb->bottom.support.timeout = ZFCP_ELS_TIMEOUT; 1655 fsf_req->qtcb->bottom.support.timeout = ZFCP_ELS_TIMEOUT;
1658 fsf_req->data = (unsigned long) els; 1656 fsf_req->data = (unsigned long) els;
1659 1657
1660 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 1658 sbale = zfcp_qdio_sbale_req(fsf_req);
1661 1659
1662 zfcp_san_dbf_event_els_request(fsf_req); 1660 zfcp_san_dbf_event_els_request(fsf_req);
1663 1661
@@ -1680,8 +1678,7 @@ zfcp_fsf_send_els(struct zfcp_send_els *els)
1680 1678
1681 failed_req: 1679 failed_req:
1682 out: 1680 out:
1683 write_unlock_irqrestore(&adapter->request_queue.queue_lock, 1681 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
1684 lock_flags);
1685 1682
1686 return ret; 1683 return ret;
1687} 1684}
@@ -1863,12 +1860,11 @@ zfcp_fsf_exchange_config_data(struct zfcp_erp_action *erp_action)
1863 ZFCP_LOG_INFO("error: Could not create exchange configuration " 1860 ZFCP_LOG_INFO("error: Could not create exchange configuration "
1864 "data request for adapter %s.\n", 1861 "data request for adapter %s.\n",
1865 zfcp_get_busid_by_adapter(adapter)); 1862 zfcp_get_busid_by_adapter(adapter));
1866 write_unlock_irqrestore(&adapter->request_queue.queue_lock, 1863 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
1867 lock_flags);
1868 return retval; 1864 return retval;
1869 } 1865 }
1870 1866
1871 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 1867 sbale = zfcp_qdio_sbale_req(fsf_req);
1872 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; 1868 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ;
1873 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; 1869 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY;
1874 1870
@@ -1882,8 +1878,7 @@ zfcp_fsf_exchange_config_data(struct zfcp_erp_action *erp_action)
1882 1878
1883 zfcp_erp_start_timer(fsf_req); 1879 zfcp_erp_start_timer(fsf_req);
1884 retval = zfcp_fsf_req_send(fsf_req); 1880 retval = zfcp_fsf_req_send(fsf_req);
1885 write_unlock_irqrestore(&adapter->request_queue.queue_lock, 1881 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
1886 lock_flags);
1887 if (retval) { 1882 if (retval) {
1888 ZFCP_LOG_INFO("error: Could not send exchange configuration " 1883 ZFCP_LOG_INFO("error: Could not send exchange configuration "
1889 "data command on the adapter %s\n", 1884 "data command on the adapter %s\n",
@@ -1916,12 +1911,11 @@ zfcp_fsf_exchange_config_data_sync(struct zfcp_adapter *adapter,
1916 ZFCP_LOG_INFO("error: Could not create exchange configuration " 1911 ZFCP_LOG_INFO("error: Could not create exchange configuration "
1917 "data request for adapter %s.\n", 1912 "data request for adapter %s.\n",
1918 zfcp_get_busid_by_adapter(adapter)); 1913 zfcp_get_busid_by_adapter(adapter));
1919 write_unlock_irqrestore(&adapter->request_queue.queue_lock, 1914 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
1920 lock_flags);
1921 return retval; 1915 return retval;
1922 } 1916 }
1923 1917
1924 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 1918 sbale = zfcp_qdio_sbale_req(fsf_req);
1925 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; 1919 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ;
1926 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; 1920 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY;
1927 1921
@@ -1936,8 +1930,7 @@ zfcp_fsf_exchange_config_data_sync(struct zfcp_adapter *adapter,
1936 1930
1937 zfcp_fsf_start_timer(fsf_req, ZFCP_FSF_REQUEST_TIMEOUT); 1931 zfcp_fsf_start_timer(fsf_req, ZFCP_FSF_REQUEST_TIMEOUT);
1938 retval = zfcp_fsf_req_send(fsf_req); 1932 retval = zfcp_fsf_req_send(fsf_req);
1939 write_unlock_irqrestore(&adapter->request_queue.queue_lock, 1933 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
1940 lock_flags);
1941 if (retval) 1934 if (retval)
1942 ZFCP_LOG_INFO("error: Could not send exchange configuration " 1935 ZFCP_LOG_INFO("error: Could not send exchange configuration "
1943 "data command on the adapter %s\n", 1936 "data command on the adapter %s\n",
@@ -2178,12 +2171,11 @@ zfcp_fsf_exchange_port_data(struct zfcp_erp_action *erp_action)
2178 "exchange port data request for " 2171 "exchange port data request for "
2179 "the adapter %s.\n", 2172 "the adapter %s.\n",
2180 zfcp_get_busid_by_adapter(adapter)); 2173 zfcp_get_busid_by_adapter(adapter));
2181 write_unlock_irqrestore(&adapter->request_queue.queue_lock, 2174 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
2182 lock_flags);
2183 return retval; 2175 return retval;
2184 } 2176 }
2185 2177
2186 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 2178 sbale = zfcp_qdio_sbale_req(fsf_req);
2187 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; 2179 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ;
2188 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; 2180 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY;
2189 2181
@@ -2192,7 +2184,7 @@ zfcp_fsf_exchange_port_data(struct zfcp_erp_action *erp_action)
2192 zfcp_erp_start_timer(fsf_req); 2184 zfcp_erp_start_timer(fsf_req);
2193 2185
2194 retval = zfcp_fsf_req_send(fsf_req); 2186 retval = zfcp_fsf_req_send(fsf_req);
2195 write_unlock_irqrestore(&adapter->request_queue.queue_lock, lock_flags); 2187 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
2196 2188
2197 if (retval) { 2189 if (retval) {
2198 ZFCP_LOG_INFO("error: Could not send an exchange port data " 2190 ZFCP_LOG_INFO("error: Could not send an exchange port data "
@@ -2237,21 +2229,20 @@ zfcp_fsf_exchange_port_data_sync(struct zfcp_adapter *adapter,
2237 "exchange port data request for " 2229 "exchange port data request for "
2238 "the adapter %s.\n", 2230 "the adapter %s.\n",
2239 zfcp_get_busid_by_adapter(adapter)); 2231 zfcp_get_busid_by_adapter(adapter));
2240 write_unlock_irqrestore(&adapter->request_queue.queue_lock, 2232 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
2241 lock_flags);
2242 return retval; 2233 return retval;
2243 } 2234 }
2244 2235
2245 if (data) 2236 if (data)
2246 fsf_req->data = (unsigned long) data; 2237 fsf_req->data = (unsigned long) data;
2247 2238
2248 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 2239 sbale = zfcp_qdio_sbale_req(fsf_req);
2249 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; 2240 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ;
2250 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; 2241 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY;
2251 2242
2252 zfcp_fsf_start_timer(fsf_req, ZFCP_FSF_REQUEST_TIMEOUT); 2243 zfcp_fsf_start_timer(fsf_req, ZFCP_FSF_REQUEST_TIMEOUT);
2253 retval = zfcp_fsf_req_send(fsf_req); 2244 retval = zfcp_fsf_req_send(fsf_req);
2254 write_unlock_irqrestore(&adapter->request_queue.queue_lock, lock_flags); 2245 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
2255 2246
2256 if (retval) 2247 if (retval)
2257 ZFCP_LOG_INFO("error: Could not send an exchange port data " 2248 ZFCP_LOG_INFO("error: Could not send an exchange port data "
@@ -2355,7 +2346,7 @@ zfcp_fsf_open_port(struct zfcp_erp_action *erp_action)
2355 goto out; 2346 goto out;
2356 } 2347 }
2357 2348
2358 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 2349 sbale = zfcp_qdio_sbale_req(fsf_req);
2359 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; 2350 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ;
2360 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; 2351 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY;
2361 2352
@@ -2382,8 +2373,7 @@ zfcp_fsf_open_port(struct zfcp_erp_action *erp_action)
2382 zfcp_get_busid_by_adapter(erp_action->adapter), 2373 zfcp_get_busid_by_adapter(erp_action->adapter),
2383 erp_action->port->wwpn); 2374 erp_action->port->wwpn);
2384 out: 2375 out:
2385 write_unlock_irqrestore(&erp_action->adapter->request_queue.queue_lock, 2376 write_unlock_irqrestore(&erp_action->adapter->req_q.lock, lock_flags);
2386 lock_flags);
2387 return retval; 2377 return retval;
2388} 2378}
2389 2379
@@ -2587,7 +2577,7 @@ zfcp_fsf_close_port(struct zfcp_erp_action *erp_action)
2587 goto out; 2577 goto out;
2588 } 2578 }
2589 2579
2590 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 2580 sbale = zfcp_qdio_sbale_req(fsf_req);
2591 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; 2581 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ;
2592 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; 2582 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY;
2593 2583
@@ -2615,8 +2605,7 @@ zfcp_fsf_close_port(struct zfcp_erp_action *erp_action)
2615 zfcp_get_busid_by_adapter(erp_action->adapter), 2605 zfcp_get_busid_by_adapter(erp_action->adapter),
2616 erp_action->port->wwpn); 2606 erp_action->port->wwpn);
2617 out: 2607 out:
2618 write_unlock_irqrestore(&erp_action->adapter->request_queue.queue_lock, 2608 write_unlock_irqrestore(&erp_action->adapter->req_q.lock, lock_flags);
2619 lock_flags);
2620 return retval; 2609 return retval;
2621} 2610}
2622 2611
@@ -2716,7 +2705,7 @@ zfcp_fsf_close_physical_port(struct zfcp_erp_action *erp_action)
2716 goto out; 2705 goto out;
2717 } 2706 }
2718 2707
2719 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 2708 sbale = zfcp_qdio_sbale_req(fsf_req);
2720 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; 2709 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ;
2721 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; 2710 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY;
2722 2711
@@ -2746,8 +2735,7 @@ zfcp_fsf_close_physical_port(struct zfcp_erp_action *erp_action)
2746 zfcp_get_busid_by_adapter(erp_action->adapter), 2735 zfcp_get_busid_by_adapter(erp_action->adapter),
2747 erp_action->port->wwpn); 2736 erp_action->port->wwpn);
2748 out: 2737 out:
2749 write_unlock_irqrestore(&erp_action->adapter->request_queue.queue_lock, 2738 write_unlock_irqrestore(&erp_action->adapter->req_q.lock, lock_flags);
2750 lock_flags);
2751 return retval; 2739 return retval;
2752} 2740}
2753 2741
@@ -2911,7 +2899,7 @@ zfcp_fsf_open_unit(struct zfcp_erp_action *erp_action)
2911 goto out; 2899 goto out;
2912 } 2900 }
2913 2901
2914 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 2902 sbale = zfcp_qdio_sbale_req(fsf_req);
2915 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; 2903 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ;
2916 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; 2904 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY;
2917 2905
@@ -2944,8 +2932,7 @@ zfcp_fsf_open_unit(struct zfcp_erp_action *erp_action)
2944 zfcp_get_busid_by_adapter(erp_action->adapter), 2932 zfcp_get_busid_by_adapter(erp_action->adapter),
2945 erp_action->port->wwpn, erp_action->unit->fcp_lun); 2933 erp_action->port->wwpn, erp_action->unit->fcp_lun);
2946 out: 2934 out:
2947 write_unlock_irqrestore(&erp_action->adapter->request_queue.queue_lock, 2935 write_unlock_irqrestore(&erp_action->adapter->req_q.lock, lock_flags);
2948 lock_flags);
2949 return retval; 2936 return retval;
2950} 2937}
2951 2938
@@ -3226,7 +3213,7 @@ zfcp_fsf_close_unit(struct zfcp_erp_action *erp_action)
3226 goto out; 3213 goto out;
3227 } 3214 }
3228 3215
3229 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 3216 sbale = zfcp_qdio_sbale_req(fsf_req);
3230 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ; 3217 sbale[0].flags |= SBAL_FLAGS0_TYPE_READ;
3231 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; 3218 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY;
3232 3219
@@ -3255,8 +3242,7 @@ zfcp_fsf_close_unit(struct zfcp_erp_action *erp_action)
3255 zfcp_get_busid_by_adapter(erp_action->adapter), 3242 zfcp_get_busid_by_adapter(erp_action->adapter),
3256 erp_action->port->wwpn, erp_action->unit->fcp_lun); 3243 erp_action->port->wwpn, erp_action->unit->fcp_lun);
3257 out: 3244 out:
3258 write_unlock_irqrestore(&erp_action->adapter->request_queue.queue_lock, 3245 write_unlock_irqrestore(&erp_action->adapter->req_q.lock, lock_flags);
3259 lock_flags);
3260 return retval; 3246 return retval;
3261} 3247}
3262 3248
@@ -3498,7 +3484,9 @@ zfcp_fsf_send_fcp_command_task(struct zfcp_adapter *adapter,
3498 fcp_cmnd_iu->add_fcp_cdb_length + sizeof (fcp_dl_t); 3484 fcp_cmnd_iu->add_fcp_cdb_length + sizeof (fcp_dl_t);
3499 3485
3500 /* generate SBALEs from data buffer */ 3486 /* generate SBALEs from data buffer */
3501 real_bytes = zfcp_qdio_sbals_from_scsicmnd(fsf_req, sbtype, scsi_cmnd); 3487 real_bytes = zfcp_qdio_sbals_from_sg(fsf_req, sbtype,
3488 scsi_sglist(scsi_cmnd),
3489 ZFCP_MAX_SBALS_PER_REQ);
3502 if (unlikely(real_bytes < 0)) { 3490 if (unlikely(real_bytes < 0)) {
3503 if (fsf_req->sbal_number < ZFCP_MAX_SBALS_PER_REQ) { 3491 if (fsf_req->sbal_number < ZFCP_MAX_SBALS_PER_REQ) {
3504 ZFCP_LOG_DEBUG( 3492 ZFCP_LOG_DEBUG(
@@ -3556,7 +3544,7 @@ zfcp_fsf_send_fcp_command_task(struct zfcp_adapter *adapter,
3556 scsi_cmnd->host_scribble = NULL; 3544 scsi_cmnd->host_scribble = NULL;
3557 success: 3545 success:
3558 failed_req_create: 3546 failed_req_create:
3559 write_unlock_irqrestore(&adapter->request_queue.queue_lock, lock_flags); 3547 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
3560 return retval; 3548 return retval;
3561} 3549}
3562 3550
@@ -3609,7 +3597,7 @@ zfcp_fsf_send_fcp_command_task_management(struct zfcp_adapter *adapter,
3609 fsf_req->qtcb->bottom.io.fcp_cmnd_length = 3597 fsf_req->qtcb->bottom.io.fcp_cmnd_length =
3610 sizeof (struct fcp_cmnd_iu) + sizeof (fcp_dl_t); 3598 sizeof (struct fcp_cmnd_iu) + sizeof (fcp_dl_t);
3611 3599
3612 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 3600 sbale = zfcp_qdio_sbale_req(fsf_req);
3613 sbale[0].flags |= SBAL_FLAGS0_TYPE_WRITE; 3601 sbale[0].flags |= SBAL_FLAGS0_TYPE_WRITE;
3614 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY; 3602 sbale[1].flags |= SBAL_FLAGS_LAST_ENTRY;
3615 3603
@@ -3629,7 +3617,7 @@ zfcp_fsf_send_fcp_command_task_management(struct zfcp_adapter *adapter,
3629 fsf_req = NULL; 3617 fsf_req = NULL;
3630 3618
3631 out: 3619 out:
3632 write_unlock_irqrestore(&adapter->request_queue.queue_lock, lock_flags); 3620 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
3633 return fsf_req; 3621 return fsf_req;
3634} 3622}
3635 3623
@@ -4216,7 +4204,7 @@ struct zfcp_fsf_req *zfcp_fsf_control_file(struct zfcp_adapter *adapter,
4216 goto unlock_queue_lock; 4204 goto unlock_queue_lock;
4217 } 4205 }
4218 4206
4219 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 4207 sbale = zfcp_qdio_sbale_req(fsf_req);
4220 sbale[0].flags |= direction; 4208 sbale[0].flags |= direction;
4221 4209
4222 bottom = &fsf_req->qtcb->bottom.support; 4210 bottom = &fsf_req->qtcb->bottom.support;
@@ -4224,7 +4212,7 @@ struct zfcp_fsf_req *zfcp_fsf_control_file(struct zfcp_adapter *adapter,
4224 bottom->option = fsf_cfdc->option; 4212 bottom->option = fsf_cfdc->option;
4225 4213
4226 bytes = zfcp_qdio_sbals_from_sg(fsf_req, direction, 4214 bytes = zfcp_qdio_sbals_from_sg(fsf_req, direction,
4227 fsf_cfdc->sg, ZFCP_CFDC_PAGES, 4215 fsf_cfdc->sg,
4228 ZFCP_MAX_SBALS_PER_REQ); 4216 ZFCP_MAX_SBALS_PER_REQ);
4229 if (bytes != ZFCP_CFDC_MAX_SIZE) { 4217 if (bytes != ZFCP_CFDC_MAX_SIZE) {
4230 retval = -ENOMEM; 4218 retval = -ENOMEM;
@@ -4237,7 +4225,7 @@ struct zfcp_fsf_req *zfcp_fsf_control_file(struct zfcp_adapter *adapter,
4237 retval = -EPERM; 4225 retval = -EPERM;
4238 goto free_fsf_req; 4226 goto free_fsf_req;
4239 } 4227 }
4240 write_unlock_irqrestore(&adapter->request_queue.queue_lock, lock_flags); 4228 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
4241 4229
4242 wait_event(fsf_req->completion_wq, 4230 wait_event(fsf_req->completion_wq,
4243 fsf_req->status & ZFCP_STATUS_FSFREQ_COMPLETED); 4231 fsf_req->status & ZFCP_STATUS_FSFREQ_COMPLETED);
@@ -4247,7 +4235,7 @@ struct zfcp_fsf_req *zfcp_fsf_control_file(struct zfcp_adapter *adapter,
4247 free_fsf_req: 4235 free_fsf_req:
4248 zfcp_fsf_req_free(fsf_req); 4236 zfcp_fsf_req_free(fsf_req);
4249 unlock_queue_lock: 4237 unlock_queue_lock:
4250 write_unlock_irqrestore(&adapter->request_queue.queue_lock, lock_flags); 4238 write_unlock_irqrestore(&adapter->req_q.lock, lock_flags);
4251 return ERR_PTR(retval); 4239 return ERR_PTR(retval);
4252} 4240}
4253 4241
@@ -4261,10 +4249,10 @@ static inline int
4261zfcp_fsf_req_sbal_check(unsigned long *flags, 4249zfcp_fsf_req_sbal_check(unsigned long *flags,
4262 struct zfcp_qdio_queue *queue, int needed) 4250 struct zfcp_qdio_queue *queue, int needed)
4263{ 4251{
4264 write_lock_irqsave(&queue->queue_lock, *flags); 4252 write_lock_irqsave(&queue->lock, *flags);
4265 if (likely(atomic_read(&queue->free_count) >= needed)) 4253 if (likely(atomic_read(&queue->count) >= needed))
4266 return 1; 4254 return 1;
4267 write_unlock_irqrestore(&queue->queue_lock, *flags); 4255 write_unlock_irqrestore(&queue->lock, *flags);
4268 return 0; 4256 return 0;
4269} 4257}
4270 4258
@@ -4293,24 +4281,24 @@ zfcp_fsf_req_qtcb_init(struct zfcp_fsf_req *fsf_req)
4293 * @req_flags: flags indicating whether to wait for needed SBAL or not 4281 * @req_flags: flags indicating whether to wait for needed SBAL or not
4294 * @lock_flags: lock_flags if queue_lock is taken 4282 * @lock_flags: lock_flags if queue_lock is taken
4295 * Return: 0 on success, otherwise -EIO, or -ERESTARTSYS 4283 * Return: 0 on success, otherwise -EIO, or -ERESTARTSYS
4296 * Locks: lock adapter->request_queue->queue_lock on success 4284 * Locks: lock adapter->req_q->lock on success
4297 */ 4285 */
4298static int 4286static int
4299zfcp_fsf_req_sbal_get(struct zfcp_adapter *adapter, int req_flags, 4287zfcp_fsf_req_sbal_get(struct zfcp_adapter *adapter, int req_flags,
4300 unsigned long *lock_flags) 4288 unsigned long *lock_flags)
4301{ 4289{
4302 long ret; 4290 long ret;
4303 struct zfcp_qdio_queue *req_queue = &adapter->request_queue; 4291 struct zfcp_qdio_queue *req_q = &adapter->req_q;
4304 4292
4305 if (unlikely(req_flags & ZFCP_WAIT_FOR_SBAL)) { 4293 if (unlikely(req_flags & ZFCP_WAIT_FOR_SBAL)) {
4306 ret = wait_event_interruptible_timeout(adapter->request_wq, 4294 ret = wait_event_interruptible_timeout(adapter->request_wq,
4307 zfcp_fsf_req_sbal_check(lock_flags, req_queue, 1), 4295 zfcp_fsf_req_sbal_check(lock_flags, req_q, 1),
4308 ZFCP_SBAL_TIMEOUT); 4296 ZFCP_SBAL_TIMEOUT);
4309 if (ret < 0) 4297 if (ret < 0)
4310 return ret; 4298 return ret;
4311 if (!ret) 4299 if (!ret)
4312 return -EIO; 4300 return -EIO;
4313 } else if (!zfcp_fsf_req_sbal_check(lock_flags, req_queue, 1)) 4301 } else if (!zfcp_fsf_req_sbal_check(lock_flags, req_q, 1))
4314 return -EIO; 4302 return -EIO;
4315 4303
4316 return 0; 4304 return 0;
@@ -4340,7 +4328,7 @@ zfcp_fsf_req_create(struct zfcp_adapter *adapter, u32 fsf_cmd, int req_flags,
4340 volatile struct qdio_buffer_element *sbale; 4328 volatile struct qdio_buffer_element *sbale;
4341 struct zfcp_fsf_req *fsf_req = NULL; 4329 struct zfcp_fsf_req *fsf_req = NULL;
4342 int ret = 0; 4330 int ret = 0;
4343 struct zfcp_qdio_queue *req_queue = &adapter->request_queue; 4331 struct zfcp_qdio_queue *req_q = &adapter->req_q;
4344 4332
4345 /* allocate new FSF request */ 4333 /* allocate new FSF request */
4346 fsf_req = zfcp_fsf_req_alloc(pool, req_flags); 4334 fsf_req = zfcp_fsf_req_alloc(pool, req_flags);
@@ -4377,7 +4365,7 @@ zfcp_fsf_req_create(struct zfcp_adapter *adapter, u32 fsf_cmd, int req_flags,
4377 */ 4365 */
4378 4366
4379 if (!atomic_test_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status)) { 4367 if (!atomic_test_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status)) {
4380 write_unlock_irqrestore(&req_queue->queue_lock, *lock_flags); 4368 write_unlock_irqrestore(&req_q->lock, *lock_flags);
4381 ret = -EIO; 4369 ret = -EIO;
4382 goto failed_sbals; 4370 goto failed_sbals;
4383 } 4371 }
@@ -4387,15 +4375,15 @@ zfcp_fsf_req_create(struct zfcp_adapter *adapter, u32 fsf_cmd, int req_flags,
4387 fsf_req->qtcb->prefix.req_seq_no = adapter->fsf_req_seq_no; 4375 fsf_req->qtcb->prefix.req_seq_no = adapter->fsf_req_seq_no;
4388 } 4376 }
4389 fsf_req->sbal_number = 1; 4377 fsf_req->sbal_number = 1;
4390 fsf_req->sbal_first = req_queue->free_index; 4378 fsf_req->sbal_first = req_q->first;
4391 fsf_req->sbal_last = req_queue->free_index; 4379 fsf_req->sbal_last = req_q->first;
4392 fsf_req->sbale_curr = 1; 4380 fsf_req->sbale_curr = 1;
4393 4381
4394 if (likely(req_flags & ZFCP_REQ_AUTO_CLEANUP)) { 4382 if (likely(req_flags & ZFCP_REQ_AUTO_CLEANUP)) {
4395 fsf_req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 4383 fsf_req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
4396 } 4384 }
4397 4385
4398 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 4386 sbale = zfcp_qdio_sbale_req(fsf_req);
4399 4387
4400 /* setup common SBALE fields */ 4388 /* setup common SBALE fields */
4401 sbale[0].addr = (void *) fsf_req->req_id; 4389 sbale[0].addr = (void *) fsf_req->req_id;
@@ -4416,7 +4404,7 @@ zfcp_fsf_req_create(struct zfcp_adapter *adapter, u32 fsf_cmd, int req_flags,
4416 fsf_req = NULL; 4404 fsf_req = NULL;
4417 4405
4418 failed_fsf_req: 4406 failed_fsf_req:
4419 write_lock_irqsave(&req_queue->queue_lock, *lock_flags); 4407 write_lock_irqsave(&req_q->lock, *lock_flags);
4420 success: 4408 success:
4421 *fsf_req_p = fsf_req; 4409 *fsf_req_p = fsf_req;
4422 return ret; 4410 return ret;
@@ -4433,18 +4421,17 @@ zfcp_fsf_req_create(struct zfcp_adapter *adapter, u32 fsf_cmd, int req_flags,
4433static int zfcp_fsf_req_send(struct zfcp_fsf_req *fsf_req) 4421static int zfcp_fsf_req_send(struct zfcp_fsf_req *fsf_req)
4434{ 4422{
4435 struct zfcp_adapter *adapter; 4423 struct zfcp_adapter *adapter;
4436 struct zfcp_qdio_queue *req_queue; 4424 struct zfcp_qdio_queue *req_q;
4437 volatile struct qdio_buffer_element *sbale; 4425 volatile struct qdio_buffer_element *sbale;
4438 int inc_seq_no; 4426 int inc_seq_no;
4439 int new_distance_from_int;
4440 int retval = 0; 4427 int retval = 0;
4441 4428
4442 adapter = fsf_req->adapter; 4429 adapter = fsf_req->adapter;
4443 req_queue = &adapter->request_queue, 4430 req_q = &adapter->req_q;
4444 4431
4445 4432
4446 /* FIXME(debug): remove it later */ 4433 /* FIXME(debug): remove it later */
4447 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_first, 0); 4434 sbale = zfcp_qdio_sbale_req(fsf_req);
4448 ZFCP_LOG_DEBUG("SBALE0 flags=0x%x\n", sbale[0].flags); 4435 ZFCP_LOG_DEBUG("SBALE0 flags=0x%x\n", sbale[0].flags);
4449 ZFCP_LOG_TRACE("HEX DUMP OF SBALE1 PAYLOAD:\n"); 4436 ZFCP_LOG_TRACE("HEX DUMP OF SBALE1 PAYLOAD:\n");
4450 ZFCP_HEX_DUMP(ZFCP_LOG_LEVEL_TRACE, (char *) sbale[1].addr, 4437 ZFCP_HEX_DUMP(ZFCP_LOG_LEVEL_TRACE, (char *) sbale[1].addr,
@@ -4457,52 +4444,24 @@ static int zfcp_fsf_req_send(struct zfcp_fsf_req *fsf_req)
4457 4444
4458 inc_seq_no = (fsf_req->qtcb != NULL); 4445 inc_seq_no = (fsf_req->qtcb != NULL);
4459 4446
4460 ZFCP_LOG_TRACE("request queue of adapter %s: "
4461 "next free SBAL is %i, %i free SBALs\n",
4462 zfcp_get_busid_by_adapter(adapter),
4463 req_queue->free_index,
4464 atomic_read(&req_queue->free_count));
4465
4466 ZFCP_LOG_DEBUG("calling do_QDIO adapter %s, flags=0x%x, queue_no=%i, "
4467 "index_in_queue=%i, count=%i, buffers=%p\n",
4468 zfcp_get_busid_by_adapter(adapter),
4469 QDIO_FLAG_SYNC_OUTPUT,
4470 0, fsf_req->sbal_first, fsf_req->sbal_number,
4471 &req_queue->buffer[fsf_req->sbal_first]);
4472
4473 /*
4474 * adjust the number of free SBALs in request queue as well as
4475 * position of first one
4476 */
4477 atomic_sub(fsf_req->sbal_number, &req_queue->free_count);
4478 ZFCP_LOG_TRACE("free_count=%d\n", atomic_read(&req_queue->free_count));
4479 req_queue->free_index += fsf_req->sbal_number; /* increase */
4480 req_queue->free_index %= QDIO_MAX_BUFFERS_PER_Q; /* wrap if needed */
4481 new_distance_from_int = zfcp_qdio_determine_pci(req_queue, fsf_req);
4482
4483 fsf_req->issued = get_clock(); 4447 fsf_req->issued = get_clock();
4484 4448
4485 retval = do_QDIO(adapter->ccw_device, 4449 retval = zfcp_qdio_send(fsf_req);
4486 QDIO_FLAG_SYNC_OUTPUT,
4487 0, fsf_req->sbal_first, fsf_req->sbal_number, NULL);
4488 4450
4489 if (unlikely(retval)) { 4451 if (unlikely(retval)) {
4490 /* Queues are down..... */ 4452 /* Queues are down..... */
4491 retval = -EIO;
4492 del_timer(&fsf_req->timer); 4453 del_timer(&fsf_req->timer);
4493 spin_lock(&adapter->req_list_lock); 4454 spin_lock(&adapter->req_list_lock);
4494 zfcp_reqlist_remove(adapter, fsf_req); 4455 zfcp_reqlist_remove(adapter, fsf_req);
4495 spin_unlock(&adapter->req_list_lock); 4456 spin_unlock(&adapter->req_list_lock);
4496 /* undo changes in request queue made for this request */ 4457 /* undo changes in request queue made for this request */
4497 zfcp_qdio_zero_sbals(req_queue->buffer, 4458 atomic_add(fsf_req->sbal_number, &req_q->count);
4498 fsf_req->sbal_first, fsf_req->sbal_number); 4459 req_q->first -= fsf_req->sbal_number;
4499 atomic_add(fsf_req->sbal_number, &req_queue->free_count); 4460 req_q->first += QDIO_MAX_BUFFERS_PER_Q;
4500 req_queue->free_index -= fsf_req->sbal_number; 4461 req_q->first %= QDIO_MAX_BUFFERS_PER_Q;
4501 req_queue->free_index += QDIO_MAX_BUFFERS_PER_Q;
4502 req_queue->free_index %= QDIO_MAX_BUFFERS_PER_Q; /* wrap */
4503 zfcp_erp_adapter_reopen(adapter, 0, 116, fsf_req); 4462 zfcp_erp_adapter_reopen(adapter, 0, 116, fsf_req);
4463 retval = -EIO;
4504 } else { 4464 } else {
4505 req_queue->distance_from_int = new_distance_from_int;
4506 /* 4465 /*
4507 * increase FSF sequence counter - 4466 * increase FSF sequence counter -
4508 * this must only be done for request successfully enqueued to 4467 * this must only be done for request successfully enqueued to
@@ -4514,9 +4473,6 @@ static int zfcp_fsf_req_send(struct zfcp_fsf_req *fsf_req)
4514 /* Don't increase for unsolicited status */ 4473 /* Don't increase for unsolicited status */
4515 if (inc_seq_no) 4474 if (inc_seq_no)
4516 adapter->fsf_req_seq_no++; 4475 adapter->fsf_req_seq_no++;
4517
4518 /* count FSF requests pending */
4519 atomic_inc(&adapter->reqs_active);
4520 } 4476 }
4521 return retval; 4477 return retval;
4522} 4478}
diff --git a/drivers/s390/scsi/zfcp_qdio.c b/drivers/s390/scsi/zfcp_qdio.c
index e71547357f62..bd6561d53589 100644
--- a/drivers/s390/scsi/zfcp_qdio.c
+++ b/drivers/s390/scsi/zfcp_qdio.c
@@ -1,241 +1,92 @@
1/* 1/*
2 * This file is part of the zfcp device driver for 2 * zfcp device driver
3 * FCP adapters for IBM System z9 and zSeries.
4 * 3 *
5 * (C) Copyright IBM Corp. 2002, 2006 4 * Setup and helper functions to access QDIO.
6 * 5 *
7 * This program is free software; you can redistribute it and/or modify 6 * Copyright IBM Corporation 2002, 2008
8 * it under the terms of the GNU General Public License as published by
9 * the Free Software Foundation; either version 2, or (at your option)
10 * any later version.
11 *
12 * This program is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 * GNU General Public License for more details.
16 *
17 * You should have received a copy of the GNU General Public License
18 * along with this program; if not, write to the Free Software
19 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
20 */ 7 */
21 8
22#include "zfcp_ext.h" 9#include "zfcp_ext.h"
23 10
24static void zfcp_qdio_sbal_limit(struct zfcp_fsf_req *, int); 11/* FIXME(tune): free space should be one max. SBAL chain plus what? */
25static inline volatile struct qdio_buffer_element *zfcp_qdio_sbale_get 12#define ZFCP_QDIO_PCI_INTERVAL (QDIO_MAX_BUFFERS_PER_Q \
26 (struct zfcp_qdio_queue *, int, int); 13 - (ZFCP_MAX_SBALS_PER_REQ + 4))
27static inline volatile struct qdio_buffer_element *zfcp_qdio_sbale_resp
28 (struct zfcp_fsf_req *, int, int);
29static volatile struct qdio_buffer_element *zfcp_qdio_sbal_chain
30 (struct zfcp_fsf_req *, unsigned long);
31static volatile struct qdio_buffer_element *zfcp_qdio_sbale_next
32 (struct zfcp_fsf_req *, unsigned long);
33static int zfcp_qdio_sbals_zero(struct zfcp_qdio_queue *, int, int);
34static inline int zfcp_qdio_sbals_wipe(struct zfcp_fsf_req *);
35static void zfcp_qdio_sbale_fill
36 (struct zfcp_fsf_req *, unsigned long, void *, int);
37static int zfcp_qdio_sbals_from_segment
38 (struct zfcp_fsf_req *, unsigned long, void *, unsigned long);
39
40static qdio_handler_t zfcp_qdio_request_handler;
41static qdio_handler_t zfcp_qdio_response_handler;
42static int zfcp_qdio_handler_error_check(struct zfcp_adapter *,
43 unsigned int, unsigned int, unsigned int, int, int);
44
45#define ZFCP_LOG_AREA ZFCP_LOG_AREA_QDIO
46 14
47/* 15static int zfcp_qdio_buffers_enqueue(struct qdio_buffer **sbal)
48 * Frees BUFFER memory for each of the pointers of the struct qdio_buffer array
49 * in the adapter struct sbuf is the pointer array.
50 *
51 * locks: must only be called with zfcp_data.config_sema taken
52 */
53static void
54zfcp_qdio_buffers_dequeue(struct qdio_buffer **sbuf)
55{
56 int pos;
57
58 for (pos = 0; pos < QDIO_MAX_BUFFERS_PER_Q; pos += QBUFF_PER_PAGE)
59 free_page((unsigned long) sbuf[pos]);
60}
61
62/*
63 * Allocates BUFFER memory to each of the pointers of the qdio_buffer_t
64 * array in the adapter struct.
65 * Cur_buf is the pointer array
66 *
67 * returns: zero on success else -ENOMEM
68 * locks: must only be called with zfcp_data.config_sema taken
69 */
70static int
71zfcp_qdio_buffers_enqueue(struct qdio_buffer **sbuf)
72{ 16{
73 int pos; 17 int pos;
74 18
75 for (pos = 0; pos < QDIO_MAX_BUFFERS_PER_Q; pos += QBUFF_PER_PAGE) { 19 for (pos = 0; pos < QDIO_MAX_BUFFERS_PER_Q; pos += QBUFF_PER_PAGE) {
76 sbuf[pos] = (struct qdio_buffer *) get_zeroed_page(GFP_KERNEL); 20 sbal[pos] = (struct qdio_buffer *) get_zeroed_page(GFP_KERNEL);
77 if (!sbuf[pos]) { 21 if (!sbal[pos])
78 zfcp_qdio_buffers_dequeue(sbuf);
79 return -ENOMEM; 22 return -ENOMEM;
80 }
81 } 23 }
82 for (pos = 0; pos < QDIO_MAX_BUFFERS_PER_Q; pos++) 24 for (pos = 0; pos < QDIO_MAX_BUFFERS_PER_Q; pos++)
83 if (pos % QBUFF_PER_PAGE) 25 if (pos % QBUFF_PER_PAGE)
84 sbuf[pos] = sbuf[pos - 1] + 1; 26 sbal[pos] = sbal[pos - 1] + 1;
85 return 0; 27 return 0;
86} 28}
87 29
88/* locks: must only be called with zfcp_data.config_sema taken */ 30static volatile struct qdio_buffer_element *
89int 31zfcp_qdio_sbale(struct zfcp_qdio_queue *q, int sbal_idx, int sbale_idx)
90zfcp_qdio_allocate_queues(struct zfcp_adapter *adapter)
91{ 32{
92 int ret; 33 return &q->sbal[sbal_idx]->element[sbale_idx];
93
94 ret = zfcp_qdio_buffers_enqueue(adapter->request_queue.buffer);
95 if (ret)
96 return ret;
97 return zfcp_qdio_buffers_enqueue(adapter->response_queue.buffer);
98} 34}
99 35
100/* locks: must only be called with zfcp_data.config_sema taken */ 36/**
101void 37 * zfcp_qdio_free - free memory used by request- and resposne queue
102zfcp_qdio_free_queues(struct zfcp_adapter *adapter) 38 * @adapter: pointer to the zfcp_adapter structure
39 */
40void zfcp_qdio_free(struct zfcp_adapter *adapter)
103{ 41{
104 ZFCP_LOG_TRACE("freeing request_queue buffers\n"); 42 struct qdio_buffer **sbal_req, **sbal_resp;
105 zfcp_qdio_buffers_dequeue(adapter->request_queue.buffer); 43 int p;
106
107 ZFCP_LOG_TRACE("freeing response_queue buffers\n");
108 zfcp_qdio_buffers_dequeue(adapter->response_queue.buffer);
109}
110 44
111int 45 if (adapter->ccw_device)
112zfcp_qdio_allocate(struct zfcp_adapter *adapter) 46 qdio_free(adapter->ccw_device);
113{
114 struct qdio_initialize *init_data;
115 47
116 init_data = &adapter->qdio_init_data; 48 sbal_req = adapter->req_q.sbal;
49 sbal_resp = adapter->resp_q.sbal;
117 50
118 init_data->cdev = adapter->ccw_device; 51 for (p = 0; p < QDIO_MAX_BUFFERS_PER_Q; p += QBUFF_PER_PAGE) {
119 init_data->q_format = QDIO_SCSI_QFMT; 52 free_page((unsigned long) sbal_req[p]);
120 memcpy(init_data->adapter_name, zfcp_get_busid_by_adapter(adapter), 8); 53 free_page((unsigned long) sbal_resp[p]);
121 ASCEBC(init_data->adapter_name, 8); 54 }
122 init_data->qib_param_field_format = 0;
123 init_data->qib_param_field = NULL;
124 init_data->input_slib_elements = NULL;
125 init_data->output_slib_elements = NULL;
126 init_data->min_input_threshold = ZFCP_MIN_INPUT_THRESHOLD;
127 init_data->max_input_threshold = ZFCP_MAX_INPUT_THRESHOLD;
128 init_data->min_output_threshold = ZFCP_MIN_OUTPUT_THRESHOLD;
129 init_data->max_output_threshold = ZFCP_MAX_OUTPUT_THRESHOLD;
130 init_data->no_input_qs = 1;
131 init_data->no_output_qs = 1;
132 init_data->input_handler = zfcp_qdio_response_handler;
133 init_data->output_handler = zfcp_qdio_request_handler;
134 init_data->int_parm = (unsigned long) adapter;
135 init_data->flags = QDIO_INBOUND_0COPY_SBALS |
136 QDIO_OUTBOUND_0COPY_SBALS | QDIO_USE_OUTBOUND_PCIS;
137 init_data->input_sbal_addr_array =
138 (void **) (adapter->response_queue.buffer);
139 init_data->output_sbal_addr_array =
140 (void **) (adapter->request_queue.buffer);
141
142 return qdio_allocate(init_data);
143} 55}
144 56
145/* 57static void zfcp_qdio_handler_error(struct zfcp_adapter *adapter, u8 id)
146 * function: zfcp_qdio_handler_error_check
147 *
148 * purpose: called by the response handler to determine error condition
149 *
150 * returns: error flag
151 *
152 */
153static int
154zfcp_qdio_handler_error_check(struct zfcp_adapter *adapter, unsigned int status,
155 unsigned int qdio_error, unsigned int siga_error,
156 int first_element, int elements_processed)
157{ 58{
158 int retval = 0; 59 dev_warn(&adapter->ccw_device->dev, "QDIO problem occurred.\n");
159 60
160 if (unlikely(status & QDIO_STATUS_LOOK_FOR_ERROR)) { 61 zfcp_erp_adapter_reopen(adapter,
161 retval = -EIO; 62 ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED |
162 63 ZFCP_STATUS_COMMON_ERP_FAILED, id, NULL);
163 ZFCP_LOG_INFO("QDIO problem occurred (status=0x%x, "
164 "qdio_error=0x%x, siga_error=0x%x)\n",
165 status, qdio_error, siga_error);
166
167 zfcp_hba_dbf_event_qdio(adapter, status, qdio_error, siga_error,
168 first_element, elements_processed);
169 /*
170 * Restarting IO on the failed adapter from scratch.
171 * Since we have been using this adapter, it is save to assume
172 * that it is not failed but recoverable. The card seems to
173 * report link-up events by self-initiated queue shutdown.
174 * That is why we need to clear the link-down flag
175 * which is set again in case we have missed by a mile.
176 */
177 zfcp_erp_adapter_reopen(adapter,
178 ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED |
179 ZFCP_STATUS_COMMON_ERP_FAILED, 140,
180 NULL);
181 }
182 return retval;
183} 64}
184 65
185/* 66static void zfcp_qdio_int_req(struct ccw_device *cdev, unsigned int status,
186 * function: zfcp_qdio_request_handler 67 unsigned int qdio_err, unsigned int siga_err,
187 * 68 unsigned int queue_no, int first, int count,
188 * purpose: is called by QDIO layer for completed SBALs in request queue 69 unsigned long parm)
189 *
190 * returns: (void)
191 */
192static void
193zfcp_qdio_request_handler(struct ccw_device *ccw_device,
194 unsigned int status,
195 unsigned int qdio_error,
196 unsigned int siga_error,
197 unsigned int queue_number,
198 int first_element,
199 int elements_processed,
200 unsigned long int_parm)
201{ 70{
202 struct zfcp_adapter *adapter; 71 struct zfcp_adapter *adapter = (struct zfcp_adapter *) parm;
203 struct zfcp_qdio_queue *queue; 72 struct zfcp_qdio_queue *queue = &adapter->req_q;
204
205 adapter = (struct zfcp_adapter *) int_parm;
206 queue = &adapter->request_queue;
207 73
208 ZFCP_LOG_DEBUG("adapter %s, first=%d, elements_processed=%d\n", 74 if (unlikely(status & QDIO_STATUS_LOOK_FOR_ERROR)) {
209 zfcp_get_busid_by_adapter(adapter), 75 zfcp_hba_dbf_event_qdio(adapter, status, qdio_err, siga_err,
210 first_element, elements_processed); 76 first, count);
211 77 zfcp_qdio_handler_error(adapter, 140);
212 if (unlikely(zfcp_qdio_handler_error_check(adapter, status, qdio_error, 78 return;
213 siga_error, first_element, 79 }
214 elements_processed)))
215 goto out;
216 /*
217 * we stored address of struct zfcp_adapter data structure
218 * associated with irq in int_parm
219 */
220 80
221 /* cleanup all SBALs being program-owned now */ 81 /* cleanup all SBALs being program-owned now */
222 zfcp_qdio_zero_sbals(queue->buffer, first_element, elements_processed); 82 zfcp_qdio_zero_sbals(queue->sbal, first, count);
223 83
224 /* increase free space in outbound queue */ 84 atomic_add(count, &queue->count);
225 atomic_add(elements_processed, &queue->free_count);
226 ZFCP_LOG_DEBUG("free_count=%d\n", atomic_read(&queue->free_count));
227 wake_up(&adapter->request_wq); 85 wake_up(&adapter->request_wq);
228 ZFCP_LOG_DEBUG("elements_processed=%d, free count=%d\n",
229 elements_processed, atomic_read(&queue->free_count));
230 out:
231 return;
232} 86}
233 87
234/**
235 * zfcp_qdio_reqid_check - checks for valid reqids.
236 */
237static void zfcp_qdio_reqid_check(struct zfcp_adapter *adapter, 88static void zfcp_qdio_reqid_check(struct zfcp_adapter *adapter,
238 unsigned long req_id, int sbal) 89 unsigned long req_id, int sbal_idx)
239{ 90{
240 struct zfcp_fsf_req *fsf_req; 91 struct zfcp_fsf_req *fsf_req;
241 unsigned long flags; 92 unsigned long flags;
@@ -248,204 +99,117 @@ static void zfcp_qdio_reqid_check(struct zfcp_adapter *adapter,
248 * Unknown request means that we have potentially memory 99 * Unknown request means that we have potentially memory
249 * corruption and must stop the machine immediatly. 100 * corruption and must stop the machine immediatly.
250 */ 101 */
251 panic("error: unknown request id (%ld) on adapter %s.\n", 102 panic("error: unknown request id (%lx) on adapter %s.\n",
252 req_id, zfcp_get_busid_by_adapter(adapter)); 103 req_id, zfcp_get_busid_by_adapter(adapter));
253 104
254 zfcp_reqlist_remove(adapter, fsf_req); 105 zfcp_reqlist_remove(adapter, fsf_req);
255 atomic_dec(&adapter->reqs_active);
256 spin_unlock_irqrestore(&adapter->req_list_lock, flags); 106 spin_unlock_irqrestore(&adapter->req_list_lock, flags);
257 107
258 fsf_req->sbal_response = sbal; 108 fsf_req->sbal_response = sbal_idx;
259 /* finish the FSF request */
260 zfcp_fsf_req_complete(fsf_req); 109 zfcp_fsf_req_complete(fsf_req);
261} 110}
262 111
263/* 112static void zfcp_qdio_resp_put_back(struct zfcp_adapter *adapter, int processed)
264 * function: zfcp_qdio_response_handler
265 *
266 * purpose: is called by QDIO layer for completed SBALs in response queue
267 *
268 * returns: (void)
269 */
270static void
271zfcp_qdio_response_handler(struct ccw_device *ccw_device,
272 unsigned int status,
273 unsigned int qdio_error,
274 unsigned int siga_error,
275 unsigned int queue_number,
276 int first_element,
277 int elements_processed,
278 unsigned long int_parm)
279{ 113{
280 struct zfcp_adapter *adapter; 114 struct zfcp_qdio_queue *queue = &adapter->resp_q;
281 struct zfcp_qdio_queue *queue; 115 struct ccw_device *cdev = adapter->ccw_device;
282 int buffer_index; 116 u8 count, start = queue->first;
283 int i; 117 unsigned int retval;
284 struct qdio_buffer *buffer;
285 int retval = 0;
286 u8 count;
287 u8 start;
288 volatile struct qdio_buffer_element *buffere = NULL;
289 int buffere_index;
290
291 adapter = (struct zfcp_adapter *) int_parm;
292 queue = &adapter->response_queue;
293
294 if (unlikely(zfcp_qdio_handler_error_check(adapter, status, qdio_error,
295 siga_error, first_element,
296 elements_processed)))
297 goto out;
298 118
299 /* 119 count = atomic_read(&queue->count) + processed;
300 * we stored address of struct zfcp_adapter data structure 120
301 * associated with irq in int_parm 121 retval = do_QDIO(cdev, QDIO_FLAG_SYNC_INPUT | QDIO_FLAG_UNDER_INTERRUPT,
302 */ 122 0, start, count, NULL);
123
124 if (unlikely(retval)) {
125 atomic_set(&queue->count, count);
126 /* FIXME: Recover this with an adapter reopen? */
127 } else {
128 queue->first += count;
129 queue->first %= QDIO_MAX_BUFFERS_PER_Q;
130 atomic_set(&queue->count, 0);
131 }
132}
133
134static void zfcp_qdio_int_resp(struct ccw_device *cdev, unsigned int status,
135 unsigned int qdio_err, unsigned int siga_err,
136 unsigned int queue_no, int first, int count,
137 unsigned long parm)
138{
139 struct zfcp_adapter *adapter = (struct zfcp_adapter *) parm;
140 struct zfcp_qdio_queue *queue = &adapter->resp_q;
141 volatile struct qdio_buffer_element *sbale;
142 int sbal_idx, sbale_idx, sbal_no;
143
144 if (unlikely(status & QDIO_STATUS_LOOK_FOR_ERROR)) {
145 zfcp_hba_dbf_event_qdio(adapter, status, qdio_err, siga_err,
146 first, count);
147 zfcp_qdio_handler_error(adapter, 147);
148 return;
149 }
303 150
304 buffere = &(queue->buffer[first_element]->element[0]);
305 ZFCP_LOG_DEBUG("first BUFFERE flags=0x%x\n", buffere->flags);
306 /* 151 /*
307 * go through all SBALs from input queue currently 152 * go through all SBALs from input queue currently
308 * returned by QDIO layer 153 * returned by QDIO layer
309 */ 154 */
310 155 for (sbal_no = 0; sbal_no < count; sbal_no++) {
311 for (i = 0; i < elements_processed; i++) { 156 sbal_idx = (first + sbal_no) % QDIO_MAX_BUFFERS_PER_Q;
312
313 buffer_index = first_element + i;
314 buffer_index %= QDIO_MAX_BUFFERS_PER_Q;
315 buffer = queue->buffer[buffer_index];
316 157
317 /* go through all SBALEs of SBAL */ 158 /* go through all SBALEs of SBAL */
318 for (buffere_index = 0; 159 for (sbale_idx = 0; sbale_idx < QDIO_MAX_ELEMENTS_PER_BUFFER;
319 buffere_index < QDIO_MAX_ELEMENTS_PER_BUFFER; 160 sbale_idx++) {
320 buffere_index++) { 161 sbale = zfcp_qdio_sbale(queue, sbal_idx, sbale_idx);
321
322 /* look for QDIO request identifiers in SB */
323 buffere = &buffer->element[buffere_index];
324 zfcp_qdio_reqid_check(adapter, 162 zfcp_qdio_reqid_check(adapter,
325 (unsigned long) buffere->addr, i); 163 (unsigned long) sbale->addr,
326 164 sbal_idx);
327 /* 165 if (likely(sbale->flags & SBAL_FLAGS_LAST_ENTRY))
328 * A single used SBALE per inbound SBALE has been
329 * implemented by QDIO so far. Hope they will
330 * do some optimisation. Will need to change to
331 * unlikely() then.
332 */
333 if (likely(buffere->flags & SBAL_FLAGS_LAST_ENTRY))
334 break; 166 break;
335 }; 167 };
336 168
337 if (unlikely(!(buffere->flags & SBAL_FLAGS_LAST_ENTRY))) { 169 if (unlikely(!(sbale->flags & SBAL_FLAGS_LAST_ENTRY)))
338 ZFCP_LOG_NORMAL("bug: End of inbound data " 170 dev_warn(&adapter->ccw_device->dev,
339 "not marked!\n"); 171 "Protocol violation by adapter. "
340 } 172 "Continuing operations.\n");
341 } 173 }
342 174
343 /* 175 /*
344 * put range of SBALs back to response queue 176 * put range of SBALs back to response queue
345 * (including SBALs which have already been free before) 177 * (including SBALs which have already been free before)
346 */ 178 */
347 count = atomic_read(&queue->free_count) + elements_processed; 179 zfcp_qdio_resp_put_back(adapter, count);
348 start = queue->free_index;
349
350 ZFCP_LOG_TRACE("calling do_QDIO on adapter %s (flags=0x%x, "
351 "queue_no=%i, index_in_queue=%i, count=%i, "
352 "buffers=0x%lx\n",
353 zfcp_get_busid_by_adapter(adapter),
354 QDIO_FLAG_SYNC_INPUT | QDIO_FLAG_UNDER_INTERRUPT,
355 0, start, count, (unsigned long) &queue->buffer[start]);
356
357 retval = do_QDIO(ccw_device,
358 QDIO_FLAG_SYNC_INPUT | QDIO_FLAG_UNDER_INTERRUPT,
359 0, start, count, NULL);
360
361 if (unlikely(retval)) {
362 atomic_set(&queue->free_count, count);
363 ZFCP_LOG_DEBUG("clearing of inbound data regions failed, "
364 "queues may be down "
365 "(count=%d, start=%d, retval=%d)\n",
366 count, start, retval);
367 } else {
368 queue->free_index += count;
369 queue->free_index %= QDIO_MAX_BUFFERS_PER_Q;
370 atomic_set(&queue->free_count, 0);
371 ZFCP_LOG_TRACE("%i buffers enqueued to response "
372 "queue at position %i\n", count, start);
373 }
374 out:
375 return;
376} 180}
377 181
378/** 182/**
379 * zfcp_qdio_sbale_get - return pointer to SBALE of qdio_queue 183 * zfcp_qdio_sbale_req - return ptr to SBALE of req_q for a struct zfcp_fsf_req
380 * @queue: queue from which SBALE should be returned 184 * @fsf_req: pointer to struct fsf_req
381 * @sbal: specifies number of SBAL in queue 185 * Returns: pointer to qdio_buffer_element (SBALE) structure
382 * @sbale: specifes number of SBALE in SBAL
383 */
384static inline volatile struct qdio_buffer_element *
385zfcp_qdio_sbale_get(struct zfcp_qdio_queue *queue, int sbal, int sbale)
386{
387 return &queue->buffer[sbal]->element[sbale];
388}
389
390/**
391 * zfcp_qdio_sbale_req - return pointer to SBALE of request_queue for
392 * a struct zfcp_fsf_req
393 */ 186 */
394volatile struct qdio_buffer_element * 187volatile struct qdio_buffer_element *
395zfcp_qdio_sbale_req(struct zfcp_fsf_req *fsf_req, int sbal, int sbale) 188zfcp_qdio_sbale_req(struct zfcp_fsf_req *req)
396{
397 return zfcp_qdio_sbale_get(&fsf_req->adapter->request_queue,
398 sbal, sbale);
399}
400
401/**
402 * zfcp_qdio_sbale_resp - return pointer to SBALE of response_queue for
403 * a struct zfcp_fsf_req
404 */
405static inline volatile struct qdio_buffer_element *
406zfcp_qdio_sbale_resp(struct zfcp_fsf_req *fsf_req, int sbal, int sbale)
407{ 189{
408 return zfcp_qdio_sbale_get(&fsf_req->adapter->response_queue, 190 return zfcp_qdio_sbale(&req->adapter->req_q, req->sbal_last, 0);
409 sbal, sbale);
410} 191}
411 192
412/** 193/**
413 * zfcp_qdio_sbale_curr - return current SBALE on request_queue for 194 * zfcp_qdio_sbale_curr - return curr SBALE on req_q for a struct zfcp_fsf_req
414 * a struct zfcp_fsf_req 195 * @fsf_req: pointer to struct fsf_req
196 * Returns: pointer to qdio_buffer_element (SBALE) structure
415 */ 197 */
416volatile struct qdio_buffer_element * 198volatile struct qdio_buffer_element *
417zfcp_qdio_sbale_curr(struct zfcp_fsf_req *fsf_req) 199zfcp_qdio_sbale_curr(struct zfcp_fsf_req *req)
418{ 200{
419 return zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 201 return zfcp_qdio_sbale(&req->adapter->req_q, req->sbal_last,
420 fsf_req->sbale_curr); 202 req->sbale_curr);
421} 203}
422 204
423/** 205static void zfcp_qdio_sbal_limit(struct zfcp_fsf_req *fsf_req, int max_sbals)
424 * zfcp_qdio_sbal_limit - determine maximum number of SBALs that can be used
425 * on the request_queue for a struct zfcp_fsf_req
426 * @fsf_req: the number of the last SBAL that can be used is stored herein
427 * @max_sbals: used to pass an upper limit for the number of SBALs
428 *
429 * Note: We can assume at least one free SBAL in the request_queue when called.
430 */
431static void
432zfcp_qdio_sbal_limit(struct zfcp_fsf_req *fsf_req, int max_sbals)
433{ 206{
434 int count = atomic_read(&fsf_req->adapter->request_queue.free_count); 207 int count = atomic_read(&fsf_req->adapter->req_q.count);
435 count = min(count, max_sbals); 208 count = min(count, max_sbals);
436 fsf_req->sbal_limit = fsf_req->sbal_first; 209 fsf_req->sbal_limit = (fsf_req->sbal_first + count - 1)
437 fsf_req->sbal_limit += (count - 1); 210 % QDIO_MAX_BUFFERS_PER_Q;
438 fsf_req->sbal_limit %= QDIO_MAX_BUFFERS_PER_Q;
439} 211}
440 212
441/**
442 * zfcp_qdio_sbal_chain - chain SBALs if more than one SBAL is needed for a
443 * request
444 * @fsf_req: zfcp_fsf_req to be processed
445 * @sbtype: SBAL flags which have to be set in first SBALE of new SBAL
446 *
447 * This function changes sbal_last, sbale_curr, sbal_number of fsf_req.
448 */
449static volatile struct qdio_buffer_element * 213static volatile struct qdio_buffer_element *
450zfcp_qdio_sbal_chain(struct zfcp_fsf_req *fsf_req, unsigned long sbtype) 214zfcp_qdio_sbal_chain(struct zfcp_fsf_req *fsf_req, unsigned long sbtype)
451{ 215{
@@ -460,7 +224,7 @@ zfcp_qdio_sbal_chain(struct zfcp_fsf_req *fsf_req, unsigned long sbtype)
460 return NULL; 224 return NULL;
461 225
462 /* set chaining flag in first SBALE of current SBAL */ 226 /* set chaining flag in first SBALE of current SBAL */
463 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 227 sbale = zfcp_qdio_sbale_req(fsf_req);
464 sbale->flags |= SBAL_FLAGS0_MORE_SBALS; 228 sbale->flags |= SBAL_FLAGS0_MORE_SBALS;
465 229
466 /* calculate index of next SBAL */ 230 /* calculate index of next SBAL */
@@ -480,214 +244,271 @@ zfcp_qdio_sbal_chain(struct zfcp_fsf_req *fsf_req, unsigned long sbtype)
480 return sbale; 244 return sbale;
481} 245}
482 246
483/**
484 * zfcp_qdio_sbale_next - switch to next SBALE, chain SBALs if needed
485 */
486static volatile struct qdio_buffer_element * 247static volatile struct qdio_buffer_element *
487zfcp_qdio_sbale_next(struct zfcp_fsf_req *fsf_req, unsigned long sbtype) 248zfcp_qdio_sbale_next(struct zfcp_fsf_req *fsf_req, unsigned long sbtype)
488{ 249{
489 if (fsf_req->sbale_curr == ZFCP_LAST_SBALE_PER_SBAL) 250 if (fsf_req->sbale_curr == ZFCP_LAST_SBALE_PER_SBAL)
490 return zfcp_qdio_sbal_chain(fsf_req, sbtype); 251 return zfcp_qdio_sbal_chain(fsf_req, sbtype);
491
492 fsf_req->sbale_curr++; 252 fsf_req->sbale_curr++;
493
494 return zfcp_qdio_sbale_curr(fsf_req); 253 return zfcp_qdio_sbale_curr(fsf_req);
495} 254}
496 255
497/** 256static void zfcp_qdio_undo_sbals(struct zfcp_fsf_req *fsf_req)
498 * zfcp_qdio_sbals_zero - initialize SBALs between first and last in queue
499 * with zero from
500 */
501static int
502zfcp_qdio_sbals_zero(struct zfcp_qdio_queue *queue, int first, int last)
503{ 257{
504 struct qdio_buffer **buf = queue->buffer; 258 struct qdio_buffer **sbal = fsf_req->adapter->req_q.sbal;
505 int curr = first; 259 int first = fsf_req->sbal_first;
506 int count = 0; 260 int last = fsf_req->sbal_last;
507 261 int count = (last - first + QDIO_MAX_BUFFERS_PER_Q) %
508 for(;;) { 262 QDIO_MAX_BUFFERS_PER_Q + 1;
509 curr %= QDIO_MAX_BUFFERS_PER_Q; 263 zfcp_qdio_zero_sbals(sbal, first, count);
510 count++;
511 memset(buf[curr], 0, sizeof(struct qdio_buffer));
512 if (curr == last)
513 break;
514 curr++;
515 }
516 return count;
517} 264}
518 265
519 266static int zfcp_qdio_fill_sbals(struct zfcp_fsf_req *fsf_req,
520/** 267 unsigned int sbtype, void *start_addr,
521 * zfcp_qdio_sbals_wipe - reset all changes in SBALs for an fsf_req 268 unsigned int total_length)
522 */
523static inline int
524zfcp_qdio_sbals_wipe(struct zfcp_fsf_req *fsf_req)
525{
526 return zfcp_qdio_sbals_zero(&fsf_req->adapter->request_queue,
527 fsf_req->sbal_first, fsf_req->sbal_last);
528}
529
530
531/**
532 * zfcp_qdio_sbale_fill - set address and length in current SBALE
533 * on request_queue
534 */
535static void
536zfcp_qdio_sbale_fill(struct zfcp_fsf_req *fsf_req, unsigned long sbtype,
537 void *addr, int length)
538{ 269{
539 volatile struct qdio_buffer_element *sbale; 270 volatile struct qdio_buffer_element *sbale;
540
541 sbale = zfcp_qdio_sbale_curr(fsf_req);
542 sbale->addr = addr;
543 sbale->length = length;
544}
545
546/**
547 * zfcp_qdio_sbals_from_segment - map memory segment to SBALE(s)
548 * @fsf_req: request to be processed
549 * @sbtype: SBALE flags
550 * @start_addr: address of memory segment
551 * @total_length: length of memory segment
552 *
553 * Alignment and length of the segment determine how many SBALEs are needed
554 * for the memory segment.
555 */
556static int
557zfcp_qdio_sbals_from_segment(struct zfcp_fsf_req *fsf_req, unsigned long sbtype,
558 void *start_addr, unsigned long total_length)
559{
560 unsigned long remaining, length; 271 unsigned long remaining, length;
561 void *addr; 272 void *addr;
562 273
563 /* split segment up heeding page boundaries */ 274 /* split segment up */
564 for (addr = start_addr, remaining = total_length; remaining > 0; 275 for (addr = start_addr, remaining = total_length; remaining > 0;
565 addr += length, remaining -= length) { 276 addr += length, remaining -= length) {
566 /* get next free SBALE for new piece */ 277 sbale = zfcp_qdio_sbale_next(fsf_req, sbtype);
567 if (NULL == zfcp_qdio_sbale_next(fsf_req, sbtype)) { 278 if (!sbale) {
568 /* no SBALE left, clean up and leave */ 279 zfcp_qdio_undo_sbals(fsf_req);
569 zfcp_qdio_sbals_wipe(fsf_req);
570 return -EINVAL; 280 return -EINVAL;
571 } 281 }
572 /* calculate length of new piece */ 282
283 /* new piece must not exceed next page boundary */
573 length = min(remaining, 284 length = min(remaining,
574 (PAGE_SIZE - ((unsigned long) addr & 285 (PAGE_SIZE - ((unsigned long)addr &
575 (PAGE_SIZE - 1)))); 286 (PAGE_SIZE - 1))));
576 /* fill current SBALE with calculated piece */ 287 sbale->addr = addr;
577 zfcp_qdio_sbale_fill(fsf_req, sbtype, addr, length); 288 sbale->length = length;
578 } 289 }
579 return total_length; 290 return 0;
580} 291}
581 292
582
583/** 293/**
584 * zfcp_qdio_sbals_from_sg - fill SBALs from scatter-gather list 294 * zfcp_qdio_sbals_from_sg - fill SBALs from scatter-gather list
585 * @fsf_req: request to be processed 295 * @fsf_req: request to be processed
586 * @sbtype: SBALE flags 296 * @sbtype: SBALE flags
587 * @sg: scatter-gather list 297 * @sg: scatter-gather list
588 * @sg_count: number of elements in scatter-gather list
589 * @max_sbals: upper bound for number of SBALs to be used 298 * @max_sbals: upper bound for number of SBALs to be used
299 * Returns: number of bytes, or error (negativ)
590 */ 300 */
591int 301int zfcp_qdio_sbals_from_sg(struct zfcp_fsf_req *fsf_req, unsigned long sbtype,
592zfcp_qdio_sbals_from_sg(struct zfcp_fsf_req *fsf_req, unsigned long sbtype, 302 struct scatterlist *sg, int max_sbals)
593 struct scatterlist *sgl, int sg_count, int max_sbals)
594{ 303{
595 int sg_index;
596 struct scatterlist *sg_segment;
597 int retval;
598 volatile struct qdio_buffer_element *sbale; 304 volatile struct qdio_buffer_element *sbale;
599 int bytes = 0; 305 int retval, bytes = 0;
600 306
601 /* figure out last allowed SBAL */ 307 /* figure out last allowed SBAL */
602 zfcp_qdio_sbal_limit(fsf_req, max_sbals); 308 zfcp_qdio_sbal_limit(fsf_req, max_sbals);
603 309
604 /* set storage-block type for current SBAL */ 310 /* set storage-block type for this request */
605 sbale = zfcp_qdio_sbale_req(fsf_req, fsf_req->sbal_last, 0); 311 sbale = zfcp_qdio_sbale_req(fsf_req);
606 sbale->flags |= sbtype; 312 sbale->flags |= sbtype;
607 313
608 /* process all segements of scatter-gather list */ 314 for (; sg; sg = sg_next(sg)) {
609 for_each_sg(sgl, sg_segment, sg_count, sg_index) { 315 retval = zfcp_qdio_fill_sbals(fsf_req, sbtype, sg_virt(sg),
610 retval = zfcp_qdio_sbals_from_segment( 316 sg->length);
611 fsf_req, 317 if (retval < 0)
612 sbtype, 318 return retval;
613 zfcp_sg_to_address(sg_segment), 319 bytes += sg->length;
614 sg_segment->length);
615 if (retval < 0) {
616 bytes = retval;
617 goto out;
618 } else
619 bytes += retval;
620 } 320 }
321
621 /* assume that no other SBALEs are to follow in the same SBAL */ 322 /* assume that no other SBALEs are to follow in the same SBAL */
622 sbale = zfcp_qdio_sbale_curr(fsf_req); 323 sbale = zfcp_qdio_sbale_curr(fsf_req);
623 sbale->flags |= SBAL_FLAGS_LAST_ENTRY; 324 sbale->flags |= SBAL_FLAGS_LAST_ENTRY;
624out: 325
625 return bytes; 326 return bytes;
626} 327}
627 328
628
629/** 329/**
630 * zfcp_qdio_sbals_from_scsicmnd - fill SBALs from scsi command 330 * zfcp_qdio_send - set PCI flag in first SBALE and send req to QDIO
631 * @fsf_req: request to be processed 331 * @fsf_req: pointer to struct zfcp_fsf_req
632 * @sbtype: SBALE flags 332 * Returns: 0 on success, error otherwise
633 * @scsi_cmnd: either scatter-gather list or buffer contained herein is used
634 * to fill SBALs
635 */ 333 */
636int 334int zfcp_qdio_send(struct zfcp_fsf_req *fsf_req)
637zfcp_qdio_sbals_from_scsicmnd(struct zfcp_fsf_req *fsf_req,
638 unsigned long sbtype, struct scsi_cmnd *scsi_cmnd)
639{ 335{
640 return zfcp_qdio_sbals_from_sg(fsf_req, sbtype, scsi_sglist(scsi_cmnd), 336 struct zfcp_adapter *adapter = fsf_req->adapter;
641 scsi_sg_count(scsi_cmnd), 337 struct zfcp_qdio_queue *req_q = &adapter->req_q;
642 ZFCP_MAX_SBALS_PER_REQ); 338 int first = fsf_req->sbal_first;
339 int count = fsf_req->sbal_number;
340 int retval, pci, pci_batch;
341 volatile struct qdio_buffer_element *sbale;
342
343 /* acknowledgements for transferred buffers */
344 pci_batch = req_q->pci_batch + count;
345 if (unlikely(pci_batch >= ZFCP_QDIO_PCI_INTERVAL)) {
346 pci_batch %= ZFCP_QDIO_PCI_INTERVAL;
347 pci = first + count - (pci_batch + 1);
348 pci %= QDIO_MAX_BUFFERS_PER_Q;
349 sbale = zfcp_qdio_sbale(req_q, pci, 0);
350 sbale->flags |= SBAL_FLAGS0_PCI;
351 }
352
353 retval = do_QDIO(adapter->ccw_device, QDIO_FLAG_SYNC_OUTPUT, 0, first,
354 count, NULL);
355 if (unlikely(retval)) {
356 zfcp_qdio_zero_sbals(req_q->sbal, first, count);
357 return retval;
358 }
359
360 /* account for transferred buffers */
361 atomic_sub(count, &req_q->count);
362 req_q->first += count;
363 req_q->first %= QDIO_MAX_BUFFERS_PER_Q;
364 req_q->pci_batch = pci_batch;
365 return 0;
643} 366}
644 367
645/** 368/**
646 * zfcp_qdio_determine_pci - set PCI flag in first SBALE on qdio queue if needed 369 * zfcp_qdio_zero_sbals - zero all sbals of the specified area and queue
370 * @buf: pointer to array of SBALS
371 * @first: integer specifying the SBAL number to start
372 * @count: integer specifying the number of SBALS to process
647 */ 373 */
648int 374void zfcp_qdio_zero_sbals(struct qdio_buffer *sbal[], int first, int count)
649zfcp_qdio_determine_pci(struct zfcp_qdio_queue *req_queue,
650 struct zfcp_fsf_req *fsf_req)
651{ 375{
652 int new_distance_from_int; 376 int i, sbal_idx;
653 int pci_pos;
654 volatile struct qdio_buffer_element *sbale;
655 377
656 new_distance_from_int = req_queue->distance_from_int + 378 for (i = first; i < first + count; i++) {
657 fsf_req->sbal_number; 379 sbal_idx = i % QDIO_MAX_BUFFERS_PER_Q;
658 380 memset(sbal[sbal_idx], 0, sizeof(struct qdio_buffer));
659 if (unlikely(new_distance_from_int >= ZFCP_QDIO_PCI_INTERVAL)) {
660 new_distance_from_int %= ZFCP_QDIO_PCI_INTERVAL;
661 pci_pos = fsf_req->sbal_first;
662 pci_pos += fsf_req->sbal_number;
663 pci_pos -= new_distance_from_int;
664 pci_pos -= 1;
665 pci_pos %= QDIO_MAX_BUFFERS_PER_Q;
666 sbale = zfcp_qdio_sbale_req(fsf_req, pci_pos, 0);
667 sbale->flags |= SBAL_FLAGS0_PCI;
668 } 381 }
669 return new_distance_from_int;
670} 382}
671 383
672/* 384/**
673 * function: zfcp_zero_sbals 385 * zfcp_qdio_allocate - allocate queue memory and initialize QDIO data
674 * 386 * @adapter: pointer to struct zfcp_adapter
675 * purpose: zeros specified range of SBALs 387 * Returns: -ENOMEM on memory allocation error or return value from
676 * 388 * qdio_allocate
677 * returns: 389 */
390int zfcp_qdio_allocate(struct zfcp_adapter *adapter)
391{
392 struct qdio_initialize *init_data;
393
394 if (zfcp_qdio_buffers_enqueue(adapter->req_q.sbal) ||
395 zfcp_qdio_buffers_enqueue(adapter->resp_q.sbal))
396 return -ENOMEM;
397
398 init_data = &adapter->qdio_init_data;
399
400 init_data->cdev = adapter->ccw_device;
401 init_data->q_format = QDIO_ZFCP_QFMT;
402 memcpy(init_data->adapter_name, zfcp_get_busid_by_adapter(adapter), 8);
403 ASCEBC(init_data->adapter_name, 8);
404 init_data->qib_param_field_format = 0;
405 init_data->qib_param_field = NULL;
406 init_data->input_slib_elements = NULL;
407 init_data->output_slib_elements = NULL;
408 init_data->min_input_threshold = 1;
409 init_data->max_input_threshold = 5000;
410 init_data->min_output_threshold = 1;
411 init_data->max_output_threshold = 1000;
412 init_data->no_input_qs = 1;
413 init_data->no_output_qs = 1;
414 init_data->input_handler = zfcp_qdio_int_resp;
415 init_data->output_handler = zfcp_qdio_int_req;
416 init_data->int_parm = (unsigned long) adapter;
417 init_data->flags = QDIO_INBOUND_0COPY_SBALS |
418 QDIO_OUTBOUND_0COPY_SBALS | QDIO_USE_OUTBOUND_PCIS;
419 init_data->input_sbal_addr_array =
420 (void **) (adapter->resp_q.sbal);
421 init_data->output_sbal_addr_array =
422 (void **) (adapter->req_q.sbal);
423
424 return qdio_allocate(init_data);
425}
426
427/**
428 * zfcp_close_qdio - close qdio queues for an adapter
678 */ 429 */
679void 430void zfcp_qdio_close(struct zfcp_adapter *adapter)
680zfcp_qdio_zero_sbals(struct qdio_buffer *buf[], int first, int clean_count)
681{ 431{
682 int cur_pos; 432 struct zfcp_qdio_queue *req_q;
683 int index; 433 int first, count;
684 434
685 for (cur_pos = first; cur_pos < (first + clean_count); cur_pos++) { 435 if (!atomic_test_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status))
686 index = cur_pos % QDIO_MAX_BUFFERS_PER_Q; 436 return;
687 memset(buf[index], 0, sizeof (struct qdio_buffer)); 437
688 ZFCP_LOG_TRACE("zeroing BUFFER %d at address %p\n", 438 /* clear QDIOUP flag, thus do_QDIO is not called during qdio_shutdown */
689 index, buf[index]); 439 req_q = &adapter->req_q;
440 write_lock_irq(&req_q->lock);
441 atomic_clear_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status);
442 write_unlock_irq(&req_q->lock);
443
444 while (qdio_shutdown(adapter->ccw_device, QDIO_FLAG_CLEANUP_USING_CLEAR)
445 == -EINPROGRESS)
446 ssleep(1);
447
448 /* cleanup used outbound sbals */
449 count = atomic_read(&req_q->count);
450 if (count < QDIO_MAX_BUFFERS_PER_Q) {
451 first = (req_q->first + count) % QDIO_MAX_BUFFERS_PER_Q;
452 count = QDIO_MAX_BUFFERS_PER_Q - count;
453 zfcp_qdio_zero_sbals(req_q->sbal, first, count);
690 } 454 }
455 req_q->first = 0;
456 atomic_set(&req_q->count, 0);
457 req_q->pci_batch = 0;
458 adapter->resp_q.first = 0;
459 atomic_set(&adapter->resp_q.count, 0);
691} 460}
692 461
693#undef ZFCP_LOG_AREA 462/**
463 * zfcp_qdio_open - prepare and initialize response queue
464 * @adapter: pointer to struct zfcp_adapter
465 * Returns: 0 on success, otherwise -EIO
466 */
467int zfcp_qdio_open(struct zfcp_adapter *adapter)
468{
469 volatile struct qdio_buffer_element *sbale;
470 int cc;
471
472 if (atomic_test_mask(ZFCP_STATUS_ADAPTER_QDIOUP, &adapter->status))
473 return -EIO;
474
475 if (qdio_establish(&adapter->qdio_init_data)) {
476 dev_err(&adapter->ccw_device->dev,
477 "Establish of QDIO queues failed.\n");
478 return -EIO;
479 }
480
481 if (qdio_activate(adapter->ccw_device, 0)) {
482 dev_err(&adapter->ccw_device->dev,
483 "Activate of QDIO queues failed.\n");
484 goto failed_qdio;
485 }
486
487 for (cc = 0; cc < QDIO_MAX_BUFFERS_PER_Q; cc++) {
488 sbale = &(adapter->resp_q.sbal[cc]->element[0]);
489 sbale->length = 0;
490 sbale->flags = SBAL_FLAGS_LAST_ENTRY;
491 sbale->addr = NULL;
492 }
493
494 if (do_QDIO(adapter->ccw_device, QDIO_FLAG_SYNC_INPUT, 0, 0,
495 QDIO_MAX_BUFFERS_PER_Q, NULL)) {
496 dev_err(&adapter->ccw_device->dev,
497 "Init of QDIO response queue failed.\n");
498 goto failed_qdio;
499 }
500
501 /* set index of first avalable SBALS / number of available SBALS */
502 adapter->req_q.first = 0;
503 atomic_set(&adapter->req_q.count, QDIO_MAX_BUFFERS_PER_Q);
504 adapter->req_q.pci_batch = 0;
505
506 return 0;
507
508failed_qdio:
509 while (qdio_shutdown(adapter->ccw_device, QDIO_FLAG_CLEANUP_USING_CLEAR)
510 == -EINPROGRESS)
511 ssleep(1);
512
513 return -EIO;
514}