vmxnet3: Fix memory leaks in rx path (fwd)
authorShreyas Bhatewara <sbhatewara@vmware.com>
Fri, 19 Jun 2015 20:37:03 +0000 (13:37 -0700)
committerDavid S. Miller <davem@davemloft.net>
Tue, 23 Jun 2015 13:26:00 +0000 (06:26 -0700)
If rcd length was zero, the page used for frag was not being released. It
was being replaced with a newly allocated page. This change takes care
of that memory leak.

Signed-off-by: Guolin Yang <gyang@vmware.com>
Signed-off-by: Shreyas N Bhatewara <sbhatewara@vmware.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
drivers/net/vmxnet3/vmxnet3_drv.c

index bb352106fc91692270b0137d32a36b8797ca0255..ab539758bec1c6f095c16d3e1ec007f18dfeffd9 100644 (file)
@@ -861,6 +861,9 @@ vmxnet3_parse_and_copy_hdr(struct sk_buff *skb, struct vmxnet3_tx_queue *tq,
                                             , skb_headlen(skb));
                }
 
+               if (skb->len <= VMXNET3_HDR_COPY_SIZE)
+                       ctx->copy_size = skb->len;
+
                /* make sure headers are accessible directly */
                if (unlikely(!pskb_may_pull(skb, ctx->copy_size)))
                        goto err;
@@ -1273,36 +1276,36 @@ vmxnet3_rq_rx_complete(struct vmxnet3_rx_queue *rq,
                        if (skip_page_frags)
                                goto rcd_done;
 
-                       new_page = alloc_page(GFP_ATOMIC);
-                       if (unlikely(new_page == NULL)) {
+                       if (rcd->len) {
+                               new_page = alloc_page(GFP_ATOMIC);
                                /* Replacement page frag could not be allocated.
                                 * Reuse this page. Drop the pkt and free the
                                 * skb which contained this page as a frag. Skip
                                 * processing all the following non-sop frags.
                                 */
-                               rq->stats.rx_buf_alloc_failure++;
-                               dev_kfree_skb(ctx->skb);
-                               ctx->skb = NULL;
-                               skip_page_frags = true;
-                               goto rcd_done;
-                       }
+                               if (unlikely(!new_page)) {
+                                       rq->stats.rx_buf_alloc_failure++;
+                                       dev_kfree_skb(ctx->skb);
+                                       ctx->skb = NULL;
+                                       skip_page_frags = true;
+                                       goto rcd_done;
+                               }
 
-                       if (rcd->len) {
                                dma_unmap_page(&adapter->pdev->dev,
                                               rbi->dma_addr, rbi->len,
                                               PCI_DMA_FROMDEVICE);
 
                                vmxnet3_append_frag(ctx->skb, rcd, rbi);
-                       }
 
-                       /* Immediate refill */
-                       rbi->page = new_page;
-                       rbi->dma_addr = dma_map_page(&adapter->pdev->dev,
-                                                    rbi->page,
-                                                    0, PAGE_SIZE,
-                                                    PCI_DMA_FROMDEVICE);
-                       rxd->addr = cpu_to_le64(rbi->dma_addr);
-                       rxd->len = rbi->len;
+                               /* Immediate refill */
+                               rbi->page = new_page;
+                               rbi->dma_addr = dma_map_page(&adapter->pdev->dev
+                                                       , rbi->page,
+                                                       0, PAGE_SIZE,
+                                                       PCI_DMA_FROMDEVICE);
+                               rxd->addr = cpu_to_le64(rbi->dma_addr);
+                               rxd->len = rbi->len;
+                       }
                }