projects
/
powerpc.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
[POWERPC] 83xx: Updated mpc834x_itx_defconfig
[powerpc.git]
/
net
/
core
/
skbuff.c
diff --git
a/net/core/skbuff.c
b/net/core/skbuff.c
index
8e1c385
..
3d56468
100644
(file)
--- a/
net/core/skbuff.c
+++ b/
net/core/skbuff.c
@@
-68,8
+68,8
@@
#include "kmap_skb.h"
#include "kmap_skb.h"
-static
kmem_cache_t
*skbuff_head_cache __read_mostly;
-static
kmem_cache_t
*skbuff_fclone_cache __read_mostly;
+static
struct kmem_cache
*skbuff_head_cache __read_mostly;
+static
struct kmem_cache
*skbuff_fclone_cache __read_mostly;
/*
* Keep out-of-line to prevent kernel bloat.
/*
* Keep out-of-line to prevent kernel bloat.
@@
-88,7
+88,7
@@
static kmem_cache_t *skbuff_fclone_cache __read_mostly;
void skb_over_panic(struct sk_buff *skb, int sz, void *here)
{
printk(KERN_EMERG "skb_over_panic: text:%p len:%d put:%d head:%p "
void skb_over_panic(struct sk_buff *skb, int sz, void *here)
{
printk(KERN_EMERG "skb_over_panic: text:%p len:%d put:%d head:%p "
- "data:%p tail:%p end:%p dev:%s\n",
+ "data:%p tail:%p end:%p dev:%s\n",
here, skb->len, sz, skb->head, skb->data, skb->tail, skb->end,
skb->dev ? skb->dev->name : "<NULL>");
BUG();
here, skb->len, sz, skb->head, skb->data, skb->tail, skb->end,
skb->dev ? skb->dev->name : "<NULL>");
BUG();
@@
-106,7
+106,7
@@
void skb_over_panic(struct sk_buff *skb, int sz, void *here)
void skb_under_panic(struct sk_buff *skb, int sz, void *here)
{
printk(KERN_EMERG "skb_under_panic: text:%p len:%d put:%d head:%p "
void skb_under_panic(struct sk_buff *skb, int sz, void *here)
{
printk(KERN_EMERG "skb_under_panic: text:%p len:%d put:%d head:%p "
- "data:%p tail:%p end:%p dev:%s\n",
+ "data:%p tail:%p end:%p dev:%s\n",
here, skb->len, sz, skb->head, skb->data, skb->tail, skb->end,
skb->dev ? skb->dev->name : "<NULL>");
BUG();
here, skb->len, sz, skb->head, skb->data, skb->tail, skb->end,
skb->dev ? skb->dev->name : "<NULL>");
BUG();
@@
-132,6
+132,7
@@
EXPORT_SYMBOL(skb_truesize_bug);
* @gfp_mask: allocation mask
* @fclone: allocate from fclone cache instead of head cache
* and allocate a cloned (child) skb
* @gfp_mask: allocation mask
* @fclone: allocate from fclone cache instead of head cache
* and allocate a cloned (child) skb
+ * @node: numa node to allocate memory on
*
* Allocate a new &sk_buff. The returned buffer has no headroom and a
* tail room of size bytes. The object has a reference count of one.
*
* Allocate a new &sk_buff. The returned buffer has no headroom and a
* tail room of size bytes. The object has a reference count of one.
@@
-141,9
+142,9
@@
EXPORT_SYMBOL(skb_truesize_bug);
* %GFP_ATOMIC.
*/
struct sk_buff *__alloc_skb(unsigned int size, gfp_t gfp_mask,
* %GFP_ATOMIC.
*/
struct sk_buff *__alloc_skb(unsigned int size, gfp_t gfp_mask,
- int fclone)
+ int fclone
, int node
)
{
{
-
kmem_cache_t
*cache;
+
struct kmem_cache
*cache;
struct skb_shared_info *shinfo;
struct sk_buff *skb;
u8 *data;
struct skb_shared_info *shinfo;
struct sk_buff *skb;
u8 *data;
@@
-151,14
+152,14
@@
struct sk_buff *__alloc_skb(unsigned int size, gfp_t gfp_mask,
cache = fclone ? skbuff_fclone_cache : skbuff_head_cache;
/* Get the HEAD */
cache = fclone ? skbuff_fclone_cache : skbuff_head_cache;
/* Get the HEAD */
- skb = kmem_cache_alloc
(cache, gfp_mask & ~__GFP_DMA
);
+ skb = kmem_cache_alloc
_node(cache, gfp_mask & ~__GFP_DMA, node
);
if (!skb)
goto out;
/* Get the DATA. Size must match skb_add_mtu(). */
size = SKB_DATA_ALIGN(size);
if (!skb)
goto out;
/* Get the DATA. Size must match skb_add_mtu(). */
size = SKB_DATA_ALIGN(size);
- data = kmalloc_track_caller(size + sizeof(struct skb_shared_info),
- gfp_mask);
+ data = kmalloc_
node_
track_caller(size + sizeof(struct skb_shared_info),
+ gfp_mask
, node
);
if (!data)
goto nodata;
if (!data)
goto nodata;
@@
-210,7
+211,7
@@
nodata:
* Buffers may only be allocated from interrupts using a @gfp_mask of
* %GFP_ATOMIC.
*/
* Buffers may only be allocated from interrupts using a @gfp_mask of
* %GFP_ATOMIC.
*/
-struct sk_buff *alloc_skb_from_cache(
kmem_cache_t
*cp,
+struct sk_buff *alloc_skb_from_cache(
struct kmem_cache
*cp,
unsigned int size,
gfp_t gfp_mask)
{
unsigned int size,
gfp_t gfp_mask)
{
@@
-267,9
+268,10
@@
nodata:
struct sk_buff *__netdev_alloc_skb(struct net_device *dev,
unsigned int length, gfp_t gfp_mask)
{
struct sk_buff *__netdev_alloc_skb(struct net_device *dev,
unsigned int length, gfp_t gfp_mask)
{
+ int node = dev->dev.parent ? dev_to_node(dev->dev.parent) : -1;
struct sk_buff *skb;
struct sk_buff *skb;
- skb =
alloc_skb(length + NET_SKB_PAD, gfp_mask
);
+ skb =
__alloc_skb(length + NET_SKB_PAD, gfp_mask, 0, node
);
if (likely(skb)) {
skb_reserve(skb, NET_SKB_PAD);
skb->dev = dev;
if (likely(skb)) {
skb_reserve(skb, NET_SKB_PAD);
skb->dev = dev;
@@
-817,12
+819,12
@@
struct sk_buff *skb_copy_expand(const struct sk_buff *skb,
*
* May return error in out of memory cases. The skb is freed on error.
*/
*
* May return error in out of memory cases. The skb is freed on error.
*/
-
+
int skb_pad(struct sk_buff *skb, int pad)
{
int err;
int ntail;
int skb_pad(struct sk_buff *skb, int pad)
{
int err;
int ntail;
-
+
/* If the skbuff is non linear tailroom is always zero.. */
if (!skb_cloned(skb) && skb_tailroom(skb) >= pad) {
memset(skb->data+skb->len, 0, pad);
/* If the skbuff is non linear tailroom is always zero.. */
if (!skb_cloned(skb) && skb_tailroom(skb) >= pad) {
memset(skb->data+skb->len, 0, pad);
@@
-849,8
+851,8
@@
int skb_pad(struct sk_buff *skb, int pad)
free_skb:
kfree_skb(skb);
return err;
free_skb:
kfree_skb(skb);
return err;
-}
-
+}
+
/* Trims skb to length len. It can change skb pointers.
*/
/* Trims skb to length len. It can change skb pointers.
*/