]> git.baikalelectronics.ru Git - kernel.git/commitdiff
net: use per task frag allocator in skb_append_datato_frags
authorEric Dumazet <edumazet@google.com>
Fri, 28 Dec 2012 06:06:37 +0000 (06:06 +0000)
committerDavid S. Miller <davem@davemloft.net>
Fri, 28 Dec 2012 23:25:19 +0000 (15:25 -0800)
Use the new per task frag allocator in skb_append_datato_frags(),
to reduce number of frags and page allocator overhead.

Tested:
 ifconfig lo mtu 16436
 perf record netperf -t UDP_STREAM ; perf report

before :
 Throughput: 32928 Mbit/s
    51.79%  netperf  [kernel.kallsyms]  [k] copy_user_generic_string
     5.98%  netperf  [kernel.kallsyms]  [k] __alloc_pages_nodemask
     5.58%  netperf  [kernel.kallsyms]  [k] get_page_from_freelist
     5.01%  netperf  [kernel.kallsyms]  [k] __rmqueue
     3.74%  netperf  [kernel.kallsyms]  [k] skb_append_datato_frags
     1.87%  netperf  [kernel.kallsyms]  [k] prep_new_page
     1.42%  netperf  [kernel.kallsyms]  [k] next_zones_zonelist
     1.28%  netperf  [kernel.kallsyms]  [k] __inc_zone_state
     1.26%  netperf  [kernel.kallsyms]  [k] alloc_pages_current
     0.78%  netperf  [kernel.kallsyms]  [k] sock_alloc_send_pskb
     0.74%  netperf  [kernel.kallsyms]  [k] udp_sendmsg
     0.72%  netperf  [kernel.kallsyms]  [k] zone_watermark_ok
     0.68%  netperf  [kernel.kallsyms]  [k] __cpuset_node_allowed_softwall
     0.67%  netperf  [kernel.kallsyms]  [k] fib_table_lookup
     0.60%  netperf  [kernel.kallsyms]  [k] memcpy_fromiovecend
     0.55%  netperf  [kernel.kallsyms]  [k] __udp4_lib_lookup

 after:
  Throughput: 47185 Mbit/s
61.74% netperf  [kernel.kallsyms] [k] copy_user_generic_string
 2.07% netperf  [kernel.kallsyms] [k] prep_new_page
 1.98% netperf  [kernel.kallsyms] [k] skb_append_datato_frags
 1.02% netperf  [kernel.kallsyms] [k] sock_alloc_send_pskb
 0.97% netperf  [kernel.kallsyms] [k] enqueue_task_fair
 0.97% netperf  [kernel.kallsyms] [k] udp_sendmsg
 0.91% netperf  [kernel.kallsyms] [k] __ip_route_output_key
 0.88% netperf  [kernel.kallsyms] [k] __netif_receive_skb
 0.87% netperf  [kernel.kallsyms] [k] fib_table_lookup
 0.85% netperf  [kernel.kallsyms] [k] resched_task
 0.78% netperf  [kernel.kallsyms] [k] __udp4_lib_lookup
 0.77% netperf  [kernel.kallsyms] [k] _raw_spin_lock_irqsave

Signed-off-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
net/core/skbuff.c

index 3ab989b0de42a0bfe7905ee99d98901ee9e9a075..ec8737ec59b590f0c691828a76144afdf262304b 100644 (file)
@@ -2686,48 +2686,37 @@ int skb_append_datato_frags(struct sock *sk, struct sk_buff *skb,
                                        int len, int odd, struct sk_buff *skb),
                        void *from, int length)
 {
-       int frg_cnt = 0;
-       skb_frag_t *frag = NULL;
-       struct page *page = NULL;
-       int copy, left;
+       int frg_cnt = skb_shinfo(skb)->nr_frags;
+       int copy;
        int offset = 0;
        int ret;
+       struct page_frag *pfrag = &current->task_frag;
 
        do {
                /* Return error if we don't have space for new frag */
-               frg_cnt = skb_shinfo(skb)->nr_frags;
                if (frg_cnt >= MAX_SKB_FRAGS)
-                       return -EFAULT;
-
-               /* allocate a new page for next frag */
-               page = alloc_pages(sk->sk_allocation, 0);
+                       return -EMSGSIZE;
 
-               /* If alloc_page fails just return failure and caller will
-                * free previous allocated pages by doing kfree_skb()
-                */
-               if (page == NULL)
+               if (!sk_page_frag_refill(sk, pfrag))
                        return -ENOMEM;
 
-               /* initialize the next frag */
-               skb_fill_page_desc(skb, frg_cnt, page, 0, 0);
-               skb->truesize += PAGE_SIZE;
-               atomic_add(PAGE_SIZE, &sk->sk_wmem_alloc);
-
-               /* get the new initialized frag */
-               frg_cnt = skb_shinfo(skb)->nr_frags;
-               frag = &skb_shinfo(skb)->frags[frg_cnt - 1];
-
                /* copy the user data to page */
-               left = PAGE_SIZE - frag->page_offset;
-               copy = (length > left)? left : length;
+               copy = min_t(int, length, pfrag->size - pfrag->offset);
 
-               ret = getfrag(from, skb_frag_address(frag) + skb_frag_size(frag),
-                           offset, copy, 0, skb);
+               ret = getfrag(from, page_address(pfrag->page) + pfrag->offset,
+                             offset, copy, 0, skb);
                if (ret < 0)
                        return -EFAULT;
 
                /* copy was successful so update the size parameters */
-               skb_frag_size_add(frag, copy);
+               skb_fill_page_desc(skb, frg_cnt, pfrag->page, pfrag->offset,
+                                  copy);
+               frg_cnt++;
+               pfrag->offset += copy;
+               get_page(pfrag->page);
+
+               skb->truesize += copy;
+               atomic_add(copy, &sk->sk_wmem_alloc);
                skb->len += copy;
                skb->data_len += copy;
                offset += copy;