aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorDavid S. Miller <davem@davemloft.net>2014-10-10 15:37:36 -0400
committerDavid S. Miller <davem@davemloft.net>2014-10-10 15:37:36 -0400
commit35b7a1915aa33da812074744647db0d9262a555c (patch)
tree85ca97083049de5acf67f1a1b467b1db209e7cdc
parent1fadee0c364572f2b2e098b34001fbaa82ee2e00 (diff)
parent4c450583d9d0a8241f0f62b80038ac47b43ff843 (diff)
Merge branch 'net-drivers-pgcnt'
Eric Dumazet says: ==================== net: fix races accessing page->_count This is illegal to use atomic_set(&page->_count, ...) even if we 'own' the page. Other entities in the kernel need to use get_page_unless_zero() to get a reference to the page before testing page properties, so we could loose a refcount increment. The only case it is valid is when page->_count is 0, we can use this in __netdev_alloc_frag() Note that I never seen crashes caused by these races, the issue was reported by Andres Lagar-Cavilla and Hugh Dickins. ==================== Signed-off-by: David S. Miller <davem@davemloft.net>
-rw-r--r--drivers/net/ethernet/intel/fm10k/fm10k_main.c7
-rw-r--r--drivers/net/ethernet/intel/igb/igb_main.c7
-rw-r--r--drivers/net/ethernet/intel/ixgbe/ixgbe_main.c8
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/en_rx.c6
-rw-r--r--net/core/skbuff.c25
5 files changed, 30 insertions, 23 deletions
diff --git a/drivers/net/ethernet/intel/fm10k/fm10k_main.c b/drivers/net/ethernet/intel/fm10k/fm10k_main.c
index 6c800a330d66..9d7118a0d67a 100644
--- a/drivers/net/ethernet/intel/fm10k/fm10k_main.c
+++ b/drivers/net/ethernet/intel/fm10k/fm10k_main.c
@@ -219,11 +219,10 @@ static bool fm10k_can_reuse_rx_page(struct fm10k_rx_buffer *rx_buffer,
219 /* flip page offset to other buffer */ 219 /* flip page offset to other buffer */
220 rx_buffer->page_offset ^= FM10K_RX_BUFSZ; 220 rx_buffer->page_offset ^= FM10K_RX_BUFSZ;
221 221
222 /* since we are the only owner of the page and we need to 222 /* Even if we own the page, we are not allowed to use atomic_set()
223 * increment it, just set the value to 2 in order to avoid 223 * This would break get_page_unless_zero() users.
224 * an unnecessary locked operation
225 */ 224 */
226 atomic_set(&page->_count, 2); 225 atomic_inc(&page->_count);
227#else 226#else
228 /* move offset up to the next cache line */ 227 /* move offset up to the next cache line */
229 rx_buffer->page_offset += truesize; 228 rx_buffer->page_offset += truesize;
diff --git a/drivers/net/ethernet/intel/igb/igb_main.c b/drivers/net/ethernet/intel/igb/igb_main.c
index ae59c0b108c5..a21b14495ebd 100644
--- a/drivers/net/ethernet/intel/igb/igb_main.c
+++ b/drivers/net/ethernet/intel/igb/igb_main.c
@@ -6545,11 +6545,10 @@ static bool igb_can_reuse_rx_page(struct igb_rx_buffer *rx_buffer,
6545 /* flip page offset to other buffer */ 6545 /* flip page offset to other buffer */
6546 rx_buffer->page_offset ^= IGB_RX_BUFSZ; 6546 rx_buffer->page_offset ^= IGB_RX_BUFSZ;
6547 6547
6548 /* since we are the only owner of the page and we need to 6548 /* Even if we own the page, we are not allowed to use atomic_set()
6549 * increment it, just set the value to 2 in order to avoid 6549 * This would break get_page_unless_zero() users.
6550 * an unnecessary locked operation
6551 */ 6550 */
6552 atomic_set(&page->_count, 2); 6551 atomic_inc(&page->_count);
6553#else 6552#else
6554 /* move offset up to the next cache line */ 6553 /* move offset up to the next cache line */
6555 rx_buffer->page_offset += truesize; 6554 rx_buffer->page_offset += truesize;
diff --git a/drivers/net/ethernet/intel/ixgbe/ixgbe_main.c b/drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
index d677b5a23b58..fec5212d4337 100644
--- a/drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
+++ b/drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
@@ -1865,12 +1865,10 @@ static bool ixgbe_add_rx_frag(struct ixgbe_ring *rx_ring,
1865 /* flip page offset to other buffer */ 1865 /* flip page offset to other buffer */
1866 rx_buffer->page_offset ^= truesize; 1866 rx_buffer->page_offset ^= truesize;
1867 1867
1868 /* 1868 /* Even if we own the page, we are not allowed to use atomic_set()
1869 * since we are the only owner of the page and we need to 1869 * This would break get_page_unless_zero() users.
1870 * increment it, just set the value to 2 in order to avoid
1871 * an unecessary locked operation
1872 */ 1870 */
1873 atomic_set(&page->_count, 2); 1871 atomic_inc(&page->_count);
1874#else 1872#else
1875 /* move offset up to the next cache line */ 1873 /* move offset up to the next cache line */
1876 rx_buffer->page_offset += truesize; 1874 rx_buffer->page_offset += truesize;
diff --git a/drivers/net/ethernet/mellanox/mlx4/en_rx.c b/drivers/net/ethernet/mellanox/mlx4/en_rx.c
index a33048ee9621..01660c595f5c 100644
--- a/drivers/net/ethernet/mellanox/mlx4/en_rx.c
+++ b/drivers/net/ethernet/mellanox/mlx4/en_rx.c
@@ -76,10 +76,10 @@ static int mlx4_alloc_pages(struct mlx4_en_priv *priv,
76 page_alloc->dma = dma; 76 page_alloc->dma = dma;
77 page_alloc->page_offset = frag_info->frag_align; 77 page_alloc->page_offset = frag_info->frag_align;
78 /* Not doing get_page() for each frag is a big win 78 /* Not doing get_page() for each frag is a big win
79 * on asymetric workloads. 79 * on asymetric workloads. Note we can not use atomic_set().
80 */ 80 */
81 atomic_set(&page->_count, 81 atomic_add(page_alloc->page_size / frag_info->frag_stride - 1,
82 page_alloc->page_size / frag_info->frag_stride); 82 &page->_count);
83 return 0; 83 return 0;
84} 84}
85 85
diff --git a/net/core/skbuff.c b/net/core/skbuff.c
index a30d750647e7..829d013745ab 100644
--- a/net/core/skbuff.c
+++ b/net/core/skbuff.c
@@ -360,18 +360,29 @@ refill:
360 goto end; 360 goto end;
361 } 361 }
362 nc->frag.size = PAGE_SIZE << order; 362 nc->frag.size = PAGE_SIZE << order;
363recycle: 363 /* Even if we own the page, we do not use atomic_set().
364 atomic_set(&nc->frag.page->_count, NETDEV_PAGECNT_MAX_BIAS); 364 * This would break get_page_unless_zero() users.
365 */
366 atomic_add(NETDEV_PAGECNT_MAX_BIAS - 1,
367 &nc->frag.page->_count);
365 nc->pagecnt_bias = NETDEV_PAGECNT_MAX_BIAS; 368 nc->pagecnt_bias = NETDEV_PAGECNT_MAX_BIAS;
366 nc->frag.offset = 0; 369 nc->frag.offset = 0;
367 } 370 }
368 371
369 if (nc->frag.offset + fragsz > nc->frag.size) { 372 if (nc->frag.offset + fragsz > nc->frag.size) {
370 /* avoid unnecessary locked operations if possible */ 373 if (atomic_read(&nc->frag.page->_count) != nc->pagecnt_bias) {
371 if ((atomic_read(&nc->frag.page->_count) == nc->pagecnt_bias) || 374 if (!atomic_sub_and_test(nc->pagecnt_bias,
372 atomic_sub_and_test(nc->pagecnt_bias, &nc->frag.page->_count)) 375 &nc->frag.page->_count))
373 goto recycle; 376 goto refill;
374 goto refill; 377 /* OK, page count is 0, we can safely set it */
378 atomic_set(&nc->frag.page->_count,
379 NETDEV_PAGECNT_MAX_BIAS);
380 } else {
381 atomic_add(NETDEV_PAGECNT_MAX_BIAS - nc->pagecnt_bias,
382 &nc->frag.page->_count);
383 }
384 nc->pagecnt_bias = NETDEV_PAGECNT_MAX_BIAS;
385 nc->frag.offset = 0;
375 } 386 }
376 387
377 data = page_address(nc->frag.page) + nc->frag.offset; 388 data = page_address(nc->frag.page) + nc->frag.offset;