aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorHerbert Xu <herbert@gondor.apana.org.au>2007-10-14 03:37:52 -0400
committerDavid S. Miller <davem@sunset.davemloft.net>2007-10-15 15:26:24 -0400
commite0053ec07e32ec94535c47b10af3377255f00836 (patch)
tree05d74698e2be2ac052fca3132e830c524887a5fd
parentdec18810c52ed564c1aedc7f93dbf278b7fdf6d5 (diff)
[SKBUFF]: Add skb_morph
This patch creates a new function skb_morph that's just like skb_clone except that it lets user provide the spare skb that will be overwritten by the one that's to be cloned. This will be used by IP fragment reassembly so that we get back the same skb that went in last (rather than the head skb that we get now which requires us to carry around double pointers all over the place). Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au> Signed-off-by: David S. Miller <davem@davemloft.net>
-rw-r--r--include/linux/skbuff.h1
-rw-r--r--net/core/skbuff.c83
2 files changed, 54 insertions, 30 deletions
diff --git a/include/linux/skbuff.h b/include/linux/skbuff.h
index a656cecd373c..be5bf0b41513 100644
--- a/include/linux/skbuff.h
+++ b/include/linux/skbuff.h
@@ -357,6 +357,7 @@ static inline struct sk_buff *alloc_skb_fclone(unsigned int size,
357} 357}
358 358
359extern void kfree_skbmem(struct sk_buff *skb); 359extern void kfree_skbmem(struct sk_buff *skb);
360extern struct sk_buff *skb_morph(struct sk_buff *dst, struct sk_buff *src);
360extern struct sk_buff *skb_clone(struct sk_buff *skb, 361extern struct sk_buff *skb_clone(struct sk_buff *skb,
361 gfp_t priority); 362 gfp_t priority);
362extern struct sk_buff *skb_copy(const struct sk_buff *skb, 363extern struct sk_buff *skb_copy(const struct sk_buff *skb,
diff --git a/net/core/skbuff.c b/net/core/skbuff.c
index 758bbef506d3..c3aa68ceed62 100644
--- a/net/core/skbuff.c
+++ b/net/core/skbuff.c
@@ -400,37 +400,8 @@ static void __copy_skb_header(struct sk_buff *new, const struct sk_buff *old)
400 skb_copy_secmark(new, old); 400 skb_copy_secmark(new, old);
401} 401}
402 402
403/** 403static struct sk_buff *__skb_clone(struct sk_buff *n, struct sk_buff *skb)
404 * skb_clone - duplicate an sk_buff
405 * @skb: buffer to clone
406 * @gfp_mask: allocation priority
407 *
408 * Duplicate an &sk_buff. The new one is not owned by a socket. Both
409 * copies share the same packet data but not structure. The new
410 * buffer has a reference count of 1. If the allocation fails the
411 * function returns %NULL otherwise the new buffer is returned.
412 *
413 * If this function is called from an interrupt gfp_mask() must be
414 * %GFP_ATOMIC.
415 */
416
417struct sk_buff *skb_clone(struct sk_buff *skb, gfp_t gfp_mask)
418{ 404{
419 struct sk_buff *n;
420
421 n = skb + 1;
422 if (skb->fclone == SKB_FCLONE_ORIG &&
423 n->fclone == SKB_FCLONE_UNAVAILABLE) {
424 atomic_t *fclone_ref = (atomic_t *) (n + 1);
425 n->fclone = SKB_FCLONE_CLONE;
426 atomic_inc(fclone_ref);
427 } else {
428 n = kmem_cache_alloc(skbuff_head_cache, gfp_mask);
429 if (!n)
430 return NULL;
431 n->fclone = SKB_FCLONE_UNAVAILABLE;
432 }
433
434#define C(x) n->x = skb->x 405#define C(x) n->x = skb->x
435 406
436 n->next = n->prev = NULL; 407 n->next = n->prev = NULL;
@@ -462,6 +433,58 @@ struct sk_buff *skb_clone(struct sk_buff *skb, gfp_t gfp_mask)
462 skb->cloned = 1; 433 skb->cloned = 1;
463 434
464 return n; 435 return n;
436#undef C
437}
438
439/**
440 * skb_morph - morph one skb into another
441 * @dst: the skb to receive the contents
442 * @src: the skb to supply the contents
443 *
444 * This is identical to skb_clone except that the target skb is
445 * supplied by the user.
446 *
447 * The target skb is returned upon exit.
448 */
449struct sk_buff *skb_morph(struct sk_buff *dst, struct sk_buff *src)
450{
451 skb_release_data(dst);
452 return __skb_clone(dst, src);
453}
454EXPORT_SYMBOL_GPL(skb_morph);
455
456/**
457 * skb_clone - duplicate an sk_buff
458 * @skb: buffer to clone
459 * @gfp_mask: allocation priority
460 *
461 * Duplicate an &sk_buff. The new one is not owned by a socket. Both
462 * copies share the same packet data but not structure. The new
463 * buffer has a reference count of 1. If the allocation fails the
464 * function returns %NULL otherwise the new buffer is returned.
465 *
466 * If this function is called from an interrupt gfp_mask() must be
467 * %GFP_ATOMIC.
468 */
469
470struct sk_buff *skb_clone(struct sk_buff *skb, gfp_t gfp_mask)
471{
472 struct sk_buff *n;
473
474 n = skb + 1;
475 if (skb->fclone == SKB_FCLONE_ORIG &&
476 n->fclone == SKB_FCLONE_UNAVAILABLE) {
477 atomic_t *fclone_ref = (atomic_t *) (n + 1);
478 n->fclone = SKB_FCLONE_CLONE;
479 atomic_inc(fclone_ref);
480 } else {
481 n = kmem_cache_alloc(skbuff_head_cache, gfp_mask);
482 if (!n)
483 return NULL;
484 n->fclone = SKB_FCLONE_UNAVAILABLE;
485 }
486
487 return __skb_clone(n, skb);
465} 488}
466 489
467static void copy_skb_header(struct sk_buff *new, const struct sk_buff *old) 490static void copy_skb_header(struct sk_buff *new, const struct sk_buff *old)