diff options
-rw-r--r-- | drivers/net/xen-netfront.c | 37 |
1 files changed, 2 insertions, 35 deletions
diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c index 70e551c19e3a..b4de126825e4 100644 --- a/drivers/net/xen-netfront.c +++ b/drivers/net/xen-netfront.c | |||
@@ -214,11 +214,9 @@ static void xennet_alloc_rx_buffers(struct net_device *dev) | |||
214 | struct page *page; | 214 | struct page *page; |
215 | int i, batch_target, notify; | 215 | int i, batch_target, notify; |
216 | RING_IDX req_prod = np->rx.req_prod_pvt; | 216 | RING_IDX req_prod = np->rx.req_prod_pvt; |
217 | struct xen_memory_reservation reservation; | ||
218 | grant_ref_t ref; | 217 | grant_ref_t ref; |
219 | unsigned long pfn; | 218 | unsigned long pfn; |
220 | void *vaddr; | 219 | void *vaddr; |
221 | int nr_flips; | ||
222 | struct xen_netif_rx_request *req; | 220 | struct xen_netif_rx_request *req; |
223 | 221 | ||
224 | if (unlikely(!netif_carrier_ok(dev))) | 222 | if (unlikely(!netif_carrier_ok(dev))) |
@@ -268,7 +266,7 @@ no_skb: | |||
268 | np->rx_target = np->rx_max_target; | 266 | np->rx_target = np->rx_max_target; |
269 | 267 | ||
270 | refill: | 268 | refill: |
271 | for (nr_flips = i = 0; ; i++) { | 269 | for (i = 0; ; i++) { |
272 | skb = __skb_dequeue(&np->rx_batch); | 270 | skb = __skb_dequeue(&np->rx_batch); |
273 | if (skb == NULL) | 271 | if (skb == NULL) |
274 | break; | 272 | break; |
@@ -297,38 +295,7 @@ no_skb: | |||
297 | req->gref = ref; | 295 | req->gref = ref; |
298 | } | 296 | } |
299 | 297 | ||
300 | if (nr_flips != 0) { | 298 | wmb(); /* barrier so backend seens requests */ |
301 | reservation.extent_start = np->rx_pfn_array; | ||
302 | reservation.nr_extents = nr_flips; | ||
303 | reservation.extent_order = 0; | ||
304 | reservation.address_bits = 0; | ||
305 | reservation.domid = DOMID_SELF; | ||
306 | |||
307 | if (!xen_feature(XENFEAT_auto_translated_physmap)) { | ||
308 | /* After all PTEs have been zapped, flush the TLB. */ | ||
309 | np->rx_mcl[i-1].args[MULTI_UVMFLAGS_INDEX] = | ||
310 | UVMF_TLB_FLUSH|UVMF_ALL; | ||
311 | |||
312 | /* Give away a batch of pages. */ | ||
313 | np->rx_mcl[i].op = __HYPERVISOR_memory_op; | ||
314 | np->rx_mcl[i].args[0] = XENMEM_decrease_reservation; | ||
315 | np->rx_mcl[i].args[1] = (unsigned long)&reservation; | ||
316 | |||
317 | /* Zap PTEs and give away pages in one big | ||
318 | * multicall. */ | ||
319 | (void)HYPERVISOR_multicall(np->rx_mcl, i+1); | ||
320 | |||
321 | /* Check return status of HYPERVISOR_memory_op(). */ | ||
322 | if (unlikely(np->rx_mcl[i].result != i)) | ||
323 | panic("Unable to reduce memory reservation\n"); | ||
324 | } else { | ||
325 | if (HYPERVISOR_memory_op(XENMEM_decrease_reservation, | ||
326 | &reservation) != i) | ||
327 | panic("Unable to reduce memory reservation\n"); | ||
328 | } | ||
329 | } else { | ||
330 | wmb(); /* barrier so backend seens requests */ | ||
331 | } | ||
332 | 299 | ||
333 | /* Above is a suitable barrier to ensure backend will see requests. */ | 300 | /* Above is a suitable barrier to ensure backend will see requests. */ |
334 | np->rx.req_prod_pvt = req_prod + i; | 301 | np->rx.req_prod_pvt = req_prod + i; |