aboutsummaryrefslogtreecommitdiffstats
path: root/drivers
diff options
context:
space:
mode:
authorVasu Dev <vasu.dev@intel.com>2009-08-25 16:58:47 -0400
committerJames Bottomley <James.Bottomley@suse.de>2009-09-05 10:47:36 -0400
commite4bc50bedf0dd6c63f20a7bc0a2b46667664fba1 (patch)
tree43f4e7a1d1fb5c042330c07cd0209f1b0eb92fb5 /drivers
parenta69b06bc5e7b153043db8984564b731f99e014fc (diff)
[SCSI] fcoe, libfc: adds per cpu exch pool within exchange manager(EM)
Adds per cpu exch pool for these reasons:- 1. Currently an EM instance is shared across all cpus to manage all exches for all cpus. This required em_lock across all cpus for an exch alloc, free, lookup and reset each frame and that made em_lock expensive, so instead having per cpu exch pool with their own per cpu pool lock will likely reduce locking contention in fast path for an exch alloc, free and lookup. 2. Per cpu exch pool will likely improve cache hit ratio since all frames of an exch will be processed on the same cpu on which exch originated. This patch is only prep work to help in keeping complexity of next patch low, so this patch only sets up per cpu exch pool and related helper funcs to be used by next patch. The next patch fully makes use of per cpu exch pool in all code paths ie. tx, rx and reset. Divides per EM exch id range equally across all cpus to setup per cpu exch pool. This division is such that lower bits of exch id carries cpu number info on which exch originated, later a simple bitwise AND operation on exch id of incoming frame with fc_cpu_mask retrieves cpu number info to direct all frames to same cpu on which exch originated. This required a global fc_cpu_mask and fc_cpu_order initialized to max possible cpus number nr_cpu_ids rounded up to 2's power, this will be used in mapping exch id and exch ptr array index in pool during exch allocation, find or reset code paths. Adds a check in fc_exch_mgr_alloc() to ensure specified min_xid lower bits are zero since these bits are used to carry cpu info. Adds and initializes struct fc_exch_pool with all required fields to manage exches in pool. Allocates per cpu struct fc_exch_pool with memory for exches array for range of exches per pool. The exches array memory is followed by struct fc_exch_pool. Adds fc_exch_ptr_get/set() helper functions to get/set exch ptr in pool exches array at specified array index. Increases default FCOE_MAX_XID to 0x0FFF from 0x07EF, so that more exches are available per cpu after above described exch id range division across all cpus to each pool. Signed-off-by: Vasu Dev <vasu.dev@intel.com> Signed-off-by: Robert Love <robert.w.love@intel.com> Signed-off-by: James Bottomley <James.Bottomley@suse.de>
Diffstat (limited to 'drivers')
-rw-r--r--drivers/scsi/fcoe/fcoe.h2
-rw-r--r--drivers/scsi/libfc/fc_exch.c89
2 files changed, 87 insertions, 4 deletions
diff --git a/drivers/scsi/fcoe/fcoe.h b/drivers/scsi/fcoe/fcoe.h
index 65120e21f5c4..550d1e49d1a3 100644
--- a/drivers/scsi/fcoe/fcoe.h
+++ b/drivers/scsi/fcoe/fcoe.h
@@ -38,7 +38,7 @@
38#define FCOE_MAX_OUTSTANDING_COMMANDS 1024 38#define FCOE_MAX_OUTSTANDING_COMMANDS 1024
39 39
40#define FCOE_MIN_XID 0x0000 /* the min xid supported by fcoe_sw */ 40#define FCOE_MIN_XID 0x0000 /* the min xid supported by fcoe_sw */
41#define FCOE_MAX_XID 0x07ef /* the max xid supported by fcoe_sw */ 41#define FCOE_MAX_XID 0x0FFF /* the max xid supported by fcoe_sw */
42 42
43unsigned int fcoe_debug_logging; 43unsigned int fcoe_debug_logging;
44module_param_named(debug_logging, fcoe_debug_logging, int, S_IRUGO|S_IWUSR); 44module_param_named(debug_logging, fcoe_debug_logging, int, S_IRUGO|S_IWUSR);
diff --git a/drivers/scsi/libfc/fc_exch.c b/drivers/scsi/libfc/fc_exch.c
index 40c34274bd81..9cbe8d66eb25 100644
--- a/drivers/scsi/libfc/fc_exch.c
+++ b/drivers/scsi/libfc/fc_exch.c
@@ -32,6 +32,9 @@
32#include <scsi/libfc.h> 32#include <scsi/libfc.h>
33#include <scsi/fc_encode.h> 33#include <scsi/fc_encode.h>
34 34
35u16 fc_cpu_mask; /* cpu mask for possible cpus */
36EXPORT_SYMBOL(fc_cpu_mask);
37static u16 fc_cpu_order; /* 2's power to represent total possible cpus */
35static struct kmem_cache *fc_em_cachep; /* cache for exchanges */ 38static struct kmem_cache *fc_em_cachep; /* cache for exchanges */
36 39
37/* 40/*
@@ -48,6 +51,20 @@ static struct kmem_cache *fc_em_cachep; /* cache for exchanges */
48 */ 51 */
49 52
50/* 53/*
54 * Per cpu exchange pool
55 *
56 * This structure manages per cpu exchanges in array of exchange pointers.
57 * This array is allocated followed by struct fc_exch_pool memory for
58 * assigned range of exchanges to per cpu pool.
59 */
60struct fc_exch_pool {
61 u16 next_index; /* next possible free exchange index */
62 u16 total_exches; /* total allocated exchanges */
63 spinlock_t lock; /* exch pool lock */
64 struct list_head ex_list; /* allocated exchanges list */
65};
66
67/*
51 * Exchange manager. 68 * Exchange manager.
52 * 69 *
53 * This structure is the center for creating exchanges and sequences. 70 * This structure is the center for creating exchanges and sequences.
@@ -66,6 +83,8 @@ struct fc_exch_mgr {
66 u32 total_exches; /* total allocated exchanges */ 83 u32 total_exches; /* total allocated exchanges */
67 struct list_head ex_list; /* allocated exchanges list */ 84 struct list_head ex_list; /* allocated exchanges list */
68 mempool_t *ep_pool; /* reserve ep's */ 85 mempool_t *ep_pool; /* reserve ep's */
86 u16 pool_max_index; /* max exch array index in exch pool */
87 struct fc_exch_pool *pool; /* per cpu exch pool */
69 88
70 /* 89 /*
71 * currently exchange mgr stats are updated but not used. 90 * currently exchange mgr stats are updated but not used.
@@ -303,6 +322,19 @@ static int fc_exch_done_locked(struct fc_exch *ep)
303 return rc; 322 return rc;
304} 323}
305 324
325static inline struct fc_exch *fc_exch_ptr_get(struct fc_exch_pool *pool,
326 u16 index)
327{
328 struct fc_exch **exches = (struct fc_exch **)(pool + 1);
329 return exches[index];
330}
331
332static inline void fc_exch_ptr_set(struct fc_exch_pool *pool, u16 index,
333 struct fc_exch *ep)
334{
335 ((struct fc_exch **)(pool + 1))[index] = ep;
336}
337
306static void fc_exch_mgr_delete_ep(struct fc_exch *ep) 338static void fc_exch_mgr_delete_ep(struct fc_exch *ep)
307{ 339{
308 struct fc_exch_mgr *mp; 340 struct fc_exch_mgr *mp;
@@ -1751,6 +1783,7 @@ static void fc_exch_mgr_destroy(struct kref *kref)
1751 */ 1783 */
1752 WARN_ON(mp->total_exches != 0); 1784 WARN_ON(mp->total_exches != 0);
1753 mempool_destroy(mp->ep_pool); 1785 mempool_destroy(mp->ep_pool);
1786 free_percpu(mp->pool);
1754 kfree(mp); 1787 kfree(mp);
1755} 1788}
1756 1789
@@ -1770,8 +1803,13 @@ struct fc_exch_mgr *fc_exch_mgr_alloc(struct fc_lport *lp,
1770{ 1803{
1771 struct fc_exch_mgr *mp; 1804 struct fc_exch_mgr *mp;
1772 size_t len; 1805 size_t len;
1806 u16 pool_exch_range;
1807 size_t pool_size;
1808 unsigned int cpu;
1809 struct fc_exch_pool *pool;
1773 1810
1774 if (max_xid <= min_xid || max_xid == FC_XID_UNKNOWN) { 1811 if (max_xid <= min_xid || max_xid == FC_XID_UNKNOWN ||
1812 (min_xid & fc_cpu_mask) != 0) {
1775 FC_LPORT_DBG(lp, "Invalid min_xid 0x:%x and max_xid 0x:%x\n", 1813 FC_LPORT_DBG(lp, "Invalid min_xid 0x:%x and max_xid 0x:%x\n",
1776 min_xid, max_xid); 1814 min_xid, max_xid);
1777 return NULL; 1815 return NULL;
@@ -1802,10 +1840,31 @@ struct fc_exch_mgr *fc_exch_mgr_alloc(struct fc_lport *lp,
1802 if (!mp->ep_pool) 1840 if (!mp->ep_pool)
1803 goto free_mp; 1841 goto free_mp;
1804 1842
1843 /*
1844 * Setup per cpu exch pool with entire exchange id range equally
1845 * divided across all cpus. The exch pointers array memory is
1846 * allocated for exch range per pool.
1847 */
1848 pool_exch_range = (mp->max_xid - mp->min_xid + 1) / (fc_cpu_mask + 1);
1849 mp->pool_max_index = pool_exch_range - 1;
1850
1851 /*
1852 * Allocate and initialize per cpu exch pool
1853 */
1854 pool_size = sizeof(*pool) + pool_exch_range * sizeof(struct fc_exch *);
1855 mp->pool = __alloc_percpu(pool_size, __alignof__(struct fc_exch_pool));
1856 if (!mp->pool)
1857 goto free_mempool;
1858 for_each_possible_cpu(cpu) {
1859 pool = per_cpu_ptr(mp->pool, cpu);
1860 spin_lock_init(&pool->lock);
1861 INIT_LIST_HEAD(&pool->ex_list);
1862 }
1863
1805 kref_init(&mp->kref); 1864 kref_init(&mp->kref);
1806 if (!fc_exch_mgr_add(lp, mp, match)) { 1865 if (!fc_exch_mgr_add(lp, mp, match)) {
1807 mempool_destroy(mp->ep_pool); 1866 free_percpu(mp->pool);
1808 goto free_mp; 1867 goto free_mempool;
1809 } 1868 }
1810 1869
1811 /* 1870 /*
@@ -1816,6 +1875,8 @@ struct fc_exch_mgr *fc_exch_mgr_alloc(struct fc_lport *lp,
1816 kref_put(&mp->kref, fc_exch_mgr_destroy); 1875 kref_put(&mp->kref, fc_exch_mgr_destroy);
1817 return mp; 1876 return mp;
1818 1877
1878free_mempool:
1879 mempool_destroy(mp->ep_pool);
1819free_mp: 1880free_mp:
1820 kfree(mp); 1881 kfree(mp);
1821 return NULL; 1882 return NULL;
@@ -1975,6 +2036,28 @@ int fc_exch_init(struct fc_lport *lp)
1975 if (!lp->tt.seq_exch_abort) 2036 if (!lp->tt.seq_exch_abort)
1976 lp->tt.seq_exch_abort = fc_seq_exch_abort; 2037 lp->tt.seq_exch_abort = fc_seq_exch_abort;
1977 2038
2039 /*
2040 * Initialize fc_cpu_mask and fc_cpu_order. The
2041 * fc_cpu_mask is set for nr_cpu_ids rounded up
2042 * to order of 2's * power and order is stored
2043 * in fc_cpu_order as this is later required in
2044 * mapping between an exch id and exch array index
2045 * in per cpu exch pool.
2046 *
2047 * This round up is required to align fc_cpu_mask
2048 * to exchange id's lower bits such that all incoming
2049 * frames of an exchange gets delivered to the same
2050 * cpu on which exchange originated by simple bitwise
2051 * AND operation between fc_cpu_mask and exchange id.
2052 */
2053 fc_cpu_mask = 1;
2054 fc_cpu_order = 0;
2055 while (fc_cpu_mask < nr_cpu_ids) {
2056 fc_cpu_mask <<= 1;
2057 fc_cpu_order++;
2058 }
2059 fc_cpu_mask--;
2060
1978 return 0; 2061 return 0;
1979} 2062}
1980EXPORT_SYMBOL(fc_exch_init); 2063EXPORT_SYMBOL(fc_exch_init);