upgrade IMQ patch to the latest one, refresh patches
SVN-Revision: 19431
This commit is contained in:
parent
d33454d86d
commit
4aadb74cc1
2 changed files with 152 additions and 76 deletions
|
@ -1,6 +1,6 @@
|
|||
--- /dev/null
|
||||
+++ b/drivers/net/imq.c
|
||||
@@ -0,0 +1,571 @@
|
||||
@@ -0,0 +1,632 @@
|
||||
+/*
|
||||
+ * Pseudo-driver for the intermediate queue device.
|
||||
+ *
|
||||
|
@ -73,6 +73,15 @@
|
|||
+ * - Use netdevice feature flags to avoid extra packet handling
|
||||
+ * by core networking layer and possibly increase performance.
|
||||
+ *
|
||||
+ * 2009/09/26 - (Jussi Kivilinna)
|
||||
+ * - Add imq_nf_reinject_lockless to fix deadlock with
|
||||
+ * imq_nf_queue/imq_nf_reinject.
|
||||
+ *
|
||||
+ * 2009/12/08 - (Jussi Kivilinna)
|
||||
+ * - Port to 2.6.32
|
||||
+ * - Add check for skb->nf_queue_entry==NULL in imq_dev_xmit()
|
||||
+ * - Also add better error checking for skb->nf_queue_entry usage
|
||||
+ *
|
||||
+ * Also, many thanks to pablo Sebastian Greco for making the initial
|
||||
+ * patch and to those who helped the testing.
|
||||
+ *
|
||||
|
@ -170,6 +179,8 @@
|
|||
+{
|
||||
+ struct nf_queue_entry *entry = skb->nf_queue_entry;
|
||||
+
|
||||
+ skb->nf_queue_entry = NULL;
|
||||
+
|
||||
+ if (entry) {
|
||||
+ nf_queue_entry_release_refs(entry);
|
||||
+ kfree(entry);
|
||||
|
@ -178,6 +189,25 @@
|
|||
+ skb_restore_cb(skb); /* kfree backup */
|
||||
+}
|
||||
+
|
||||
+/* locking not needed when called from imq_nf_queue */
|
||||
+static void imq_nf_reinject_lockless(struct nf_queue_entry *entry,
|
||||
+ unsigned int verdict)
|
||||
+{
|
||||
+ int status;
|
||||
+
|
||||
+ if (!entry->next_outfn) {
|
||||
+ nf_reinject(entry, verdict);
|
||||
+ return;
|
||||
+ }
|
||||
+
|
||||
+ status = entry->next_outfn(entry, entry->next_queuenum);
|
||||
+ if (status < 0) {
|
||||
+ nf_queue_entry_release_refs(entry);
|
||||
+ kfree_skb(entry->skb);
|
||||
+ kfree(entry);
|
||||
+ }
|
||||
+}
|
||||
+
|
||||
+static void imq_nf_reinject(struct nf_queue_entry *entry, unsigned int verdict)
|
||||
+{
|
||||
+ int status;
|
||||
|
@ -202,19 +232,48 @@
|
|||
+ rcu_read_unlock();
|
||||
+}
|
||||
+
|
||||
+static int imq_dev_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
+static netdev_tx_t imq_dev_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
+{
|
||||
+ struct nf_queue_entry *entry = skb->nf_queue_entry;
|
||||
+
|
||||
+ skb->nf_queue_entry = NULL;
|
||||
+ dev->trans_start = jiffies;
|
||||
+
|
||||
+ dev->stats.tx_bytes += skb->len;
|
||||
+ dev->stats.tx_packets++;
|
||||
+
|
||||
+ if (entry == NULL) {
|
||||
+ /* We don't know what is going on here.. packet is queued for
|
||||
+ * imq device, but (probably) not by us.
|
||||
+ *
|
||||
+ * If this packet was not send here by imq_nf_queue(), then
|
||||
+ * skb_save_cb() was not used and skb_free() should not show:
|
||||
+ * WARNING: IMQ: kfree_skb: skb->cb_next:..
|
||||
+ * and/or
|
||||
+ * WARNING: IMQ: kfree_skb: skb->nf_queue_entry...
|
||||
+ *
|
||||
+ * However if this message is shown, then IMQ is somehow broken
|
||||
+ * and you should report this to linuximq.net.
|
||||
+ */
|
||||
+
|
||||
+ /* imq_dev_xmit is black hole that eats all packets, report that
|
||||
+ * we eat this packet happily and increase dropped counters.
|
||||
+ */
|
||||
+
|
||||
+ dev->stats.tx_dropped++;
|
||||
+ dev_kfree_skb(skb);
|
||||
+
|
||||
+ return NETDEV_TX_OK;
|
||||
+ }
|
||||
+
|
||||
+ skb_restore_cb(skb); /* restore skb->cb */
|
||||
+
|
||||
+ skb->imq_flags = 0;
|
||||
+ skb->destructor = NULL;
|
||||
+
|
||||
+ skb_restore_cb(skb); /* restore skb->cb */
|
||||
+ imq_nf_reinject(entry, NF_ACCEPT);
|
||||
+
|
||||
+ dev->trans_start = jiffies;
|
||||
+ imq_nf_reinject(skb->nf_queue_entry, NF_ACCEPT);
|
||||
+ return 0;
|
||||
+ return NETDEV_TX_OK;
|
||||
+}
|
||||
+
|
||||
+static int imq_nf_queue(struct nf_queue_entry *entry, unsigned queue_num)
|
||||
|
@ -257,7 +316,7 @@
|
|||
+
|
||||
+ if (unlikely(!(dev->flags & IFF_UP))) {
|
||||
+ entry->skb->imq_flags = 0;
|
||||
+ imq_nf_reinject(entry, NF_ACCEPT);
|
||||
+ imq_nf_reinject_lockless(entry, NF_ACCEPT);
|
||||
+ retval = 0;
|
||||
+ goto out;
|
||||
+ }
|
||||
|
@ -315,6 +374,7 @@
|
|||
+ goto out;
|
||||
+ } else {
|
||||
+ skb_restore_cb(skb_shared); /* restore skb->cb */
|
||||
+ skb->nf_queue_entry = NULL;
|
||||
+ /* qdisc dropped packet and decreased skb reference count of
|
||||
+ * skb, so we don't really want to and try refree as that would
|
||||
+ * actually destroy the skb. */
|
||||
|
@ -378,6 +438,7 @@
|
|||
+ dev->features = NETIF_F_SG | NETIF_F_FRAGLIST |
|
||||
+ NETIF_F_GSO | NETIF_F_HW_CSUM |
|
||||
+ NETIF_F_HIGHDMA;
|
||||
+ dev->priv_flags &= ~IFF_XMIT_DST_RELEASE;
|
||||
+}
|
||||
+
|
||||
+static int imq_validate(struct nlattr *tb[], struct nlattr *data[])
|
||||
|
@ -730,6 +791,28 @@
|
|||
+
|
||||
+#endif /* _IMQ_H */
|
||||
+
|
||||
--- a/include/linux/netdevice.h
|
||||
+++ b/include/linux/netdevice.h
|
||||
@@ -1114,6 +1114,7 @@ extern int dev_alloc_name(struct net_de
|
||||
extern int dev_open(struct net_device *dev);
|
||||
extern int dev_close(struct net_device *dev);
|
||||
extern void dev_disable_lro(struct net_device *dev);
|
||||
+extern struct netdev_queue *dev_pick_tx(struct net_device *dev, struct sk_buff *skb);
|
||||
extern int dev_queue_xmit(struct sk_buff *skb);
|
||||
extern int register_netdevice(struct net_device *dev);
|
||||
extern void unregister_netdevice(struct net_device *dev);
|
||||
--- /dev/null
|
||||
+++ b/include/linux/netfilter/xt_IMQ.h
|
||||
@@ -0,0 +1,9 @@
|
||||
+#ifndef _XT_IMQ_H
|
||||
+#define _XT_IMQ_H
|
||||
+
|
||||
+struct xt_imq_info {
|
||||
+ unsigned int todev; /* target imq device */
|
||||
+};
|
||||
+
|
||||
+#endif /* _XT_IMQ_H */
|
||||
+
|
||||
--- /dev/null
|
||||
+++ b/include/linux/netfilter_ipv4/ipt_IMQ.h
|
||||
@@ -0,0 +1,10 @@
|
||||
|
@ -788,17 +871,18 @@
|
|||
#ifdef CONFIG_BRIDGE_NETFILTER
|
||||
struct nf_bridge_info *nf_bridge;
|
||||
#endif
|
||||
@@ -382,6 +391,9 @@ struct sk_buff {
|
||||
kmemcheck_bitfield_end(flags2);
|
||||
@@ -383,6 +392,10 @@ struct sk_buff {
|
||||
|
||||
/* 0/14 bit hole */
|
||||
|
||||
+#if defined(CONFIG_IMQ) || defined(CONFIG_IMQ_MODULE)
|
||||
+ __u8 imq_flags:IMQ_F_BITS;
|
||||
+#endif
|
||||
|
||||
+
|
||||
#ifdef CONFIG_NET_DMA
|
||||
dma_cookie_t dma_cookie;
|
||||
@@ -437,6 +449,12 @@ static inline struct rtable *skb_rtable(
|
||||
#endif
|
||||
@@ -437,6 +450,12 @@ static inline struct rtable *skb_rtable(
|
||||
return (struct rtable *)skb_dst(skb);
|
||||
}
|
||||
|
||||
|
@ -811,7 +895,7 @@
|
|||
extern void kfree_skb(struct sk_buff *skb);
|
||||
extern void consume_skb(struct sk_buff *skb);
|
||||
extern void __kfree_skb(struct sk_buff *skb);
|
||||
@@ -1972,6 +1990,10 @@ static inline void __nf_copy(struct sk_b
|
||||
@@ -1972,6 +1991,10 @@ static inline void __nf_copy(struct sk_b
|
||||
dst->nfct_reasm = src->nfct_reasm;
|
||||
nf_conntrack_get_reasm(src->nfct_reasm);
|
||||
#endif
|
||||
|
@ -822,6 +906,33 @@
|
|||
#ifdef CONFIG_BRIDGE_NETFILTER
|
||||
dst->nf_bridge = src->nf_bridge;
|
||||
nf_bridge_get(src->nf_bridge);
|
||||
--- a/include/net/netfilter/nf_queue.h
|
||||
+++ b/include/net/netfilter/nf_queue.h
|
||||
@@ -13,6 +13,12 @@ struct nf_queue_entry {
|
||||
struct net_device *indev;
|
||||
struct net_device *outdev;
|
||||
int (*okfn)(struct sk_buff *);
|
||||
+
|
||||
+#if defined(CONFIG_IMQ) || defined(CONFIG_IMQ_MODULE)
|
||||
+ int (*next_outfn)(struct nf_queue_entry *entry,
|
||||
+ unsigned int queuenum);
|
||||
+ unsigned int next_queuenum;
|
||||
+#endif
|
||||
};
|
||||
|
||||
#define nf_queue_entry_reroute(x) ((void *)x + sizeof(struct nf_queue_entry))
|
||||
@@ -30,5 +36,11 @@ extern int nf_unregister_queue_handler(u
|
||||
const struct nf_queue_handler *qh);
|
||||
extern void nf_unregister_queue_handlers(const struct nf_queue_handler *qh);
|
||||
extern void nf_reinject(struct nf_queue_entry *entry, unsigned int verdict);
|
||||
+extern void nf_queue_entry_release_refs(struct nf_queue_entry *entry);
|
||||
+
|
||||
+#if defined(CONFIG_IMQ) || defined(CONFIG_IMQ_MODULE)
|
||||
+extern void nf_register_queue_imq_handler(const struct nf_queue_handler *qh);
|
||||
+extern void nf_unregister_queue_imq_handler(void);
|
||||
+#endif
|
||||
|
||||
#endif /* _NF_QUEUE_H */
|
||||
--- a/net/core/dev.c
|
||||
+++ b/net/core/dev.c
|
||||
@@ -96,6 +96,9 @@
|
||||
|
@ -865,55 +976,6 @@
|
|||
|
||||
static inline int __dev_xmit_skb(struct sk_buff *skb, struct Qdisc *q,
|
||||
struct net_device *dev,
|
||||
--- a/include/linux/netdevice.h
|
||||
+++ b/include/linux/netdevice.h
|
||||
@@ -1114,6 +1114,7 @@ extern int dev_alloc_name(struct net_de
|
||||
extern int dev_open(struct net_device *dev);
|
||||
extern int dev_close(struct net_device *dev);
|
||||
extern void dev_disable_lro(struct net_device *dev);
|
||||
+extern struct netdev_queue *dev_pick_tx(struct net_device *dev, struct sk_buff *skb);
|
||||
extern int dev_queue_xmit(struct sk_buff *skb);
|
||||
extern int register_netdevice(struct net_device *dev);
|
||||
extern void unregister_netdevice(struct net_device *dev);
|
||||
--- /dev/null
|
||||
+++ b/include/linux/netfilter/xt_IMQ.h
|
||||
@@ -0,0 +1,9 @@
|
||||
+#ifndef _XT_IMQ_H
|
||||
+#define _XT_IMQ_H
|
||||
+
|
||||
+struct xt_imq_info {
|
||||
+ unsigned int todev; /* target imq device */
|
||||
+};
|
||||
+
|
||||
+#endif /* _XT_IMQ_H */
|
||||
+
|
||||
--- a/include/net/netfilter/nf_queue.h
|
||||
+++ b/include/net/netfilter/nf_queue.h
|
||||
@@ -13,6 +13,12 @@ struct nf_queue_entry {
|
||||
struct net_device *indev;
|
||||
struct net_device *outdev;
|
||||
int (*okfn)(struct sk_buff *);
|
||||
+
|
||||
+#if defined(CONFIG_IMQ) || defined(CONFIG_IMQ_MODULE)
|
||||
+ int (*next_outfn)(struct nf_queue_entry *entry,
|
||||
+ unsigned int queuenum);
|
||||
+ unsigned int next_queuenum;
|
||||
+#endif
|
||||
};
|
||||
|
||||
#define nf_queue_entry_reroute(x) ((void *)x + sizeof(struct nf_queue_entry))
|
||||
@@ -30,5 +36,11 @@ extern int nf_unregister_queue_handler(u
|
||||
const struct nf_queue_handler *qh);
|
||||
extern void nf_unregister_queue_handlers(const struct nf_queue_handler *qh);
|
||||
extern void nf_reinject(struct nf_queue_entry *entry, unsigned int verdict);
|
||||
+extern void nf_queue_entry_release_refs(struct nf_queue_entry *entry);
|
||||
+
|
||||
+#if defined(CONFIG_IMQ) || defined(CONFIG_IMQ_MODULE)
|
||||
+extern void nf_register_queue_imq_handler(const struct nf_queue_handler *qh);
|
||||
+extern void nf_unregister_queue_imq_handler(void);
|
||||
+#endif
|
||||
|
||||
#endif /* _NF_QUEUE_H */
|
||||
--- a/net/core/skbuff.c
|
||||
+++ b/net/core/skbuff.c
|
||||
@@ -72,6 +72,9 @@
|
||||
|
@ -926,7 +988,7 @@
|
|||
|
||||
static void sock_pipe_buf_release(struct pipe_inode_info *pipe,
|
||||
struct pipe_buffer *buf)
|
||||
@@ -91,6 +94,80 @@ static int sock_pipe_buf_steal(struct pi
|
||||
@@ -91,6 +94,83 @@ static int sock_pipe_buf_steal(struct pi
|
||||
return 1;
|
||||
}
|
||||
|
||||
|
@ -986,17 +1048,20 @@
|
|||
+}
|
||||
+EXPORT_SYMBOL(skb_restore_cb);
|
||||
+
|
||||
+static void skb_copy_stored_cb(struct sk_buff *new, struct sk_buff *old)
|
||||
+static void skb_copy_stored_cb(struct sk_buff *new, const struct sk_buff *__old)
|
||||
+{
|
||||
+ struct skb_cb_table *next;
|
||||
+ struct sk_buff *old;
|
||||
+
|
||||
+ if (!old->cb_next) {
|
||||
+ new->cb_next = 0;
|
||||
+ if (!__old->cb_next) {
|
||||
+ new->cb_next = NULL;
|
||||
+ return;
|
||||
+ }
|
||||
+
|
||||
+ spin_lock(&skb_cb_store_lock);
|
||||
+
|
||||
+ old = (struct sk_buff *)__old;
|
||||
+
|
||||
+ next = old->cb_next;
|
||||
+ atomic_inc(&next->refcnt);
|
||||
+ new->cb_next = next;
|
||||
|
@ -1007,7 +1072,7 @@
|
|||
|
||||
/* Pipe buffer operations for a socket. */
|
||||
static struct pipe_buf_operations sock_pipe_buf_ops = {
|
||||
@@ -398,6 +475,15 @@ static void skb_release_head_state(struc
|
||||
@@ -398,6 +478,26 @@ static void skb_release_head_state(struc
|
||||
WARN_ON(in_irq());
|
||||
skb->destructor(skb);
|
||||
}
|
||||
|
@ -1015,15 +1080,26 @@
|
|||
+ /* This should not happen. When it does, avoid memleak by restoring
|
||||
+ the chain of cb-backups. */
|
||||
+ while(skb->cb_next != NULL) {
|
||||
+ printk(KERN_WARNING "kfree_skb: skb->cb_next: %08x\n",
|
||||
+ skb->cb_next);
|
||||
+ if (net_ratelimit())
|
||||
+ printk(KERN_WARNING "IMQ: kfree_skb: skb->cb_next: "
|
||||
+ "%08x\n", (unsigned int)skb->cb_next);
|
||||
+
|
||||
+ skb_restore_cb(skb);
|
||||
+ }
|
||||
+ /* This should not happen either, nf_queue_entry is nullified in
|
||||
+ * imq_dev_xmit(). If we have non-NULL nf_queue_entry then we are
|
||||
+ * leaking entry pointers, maybe memory. We don't know if this is
|
||||
+ * pointer to already freed memory, or should this be freed.
|
||||
+ * If this happens we need to add refcounting, etc for nf_queue_entry.
|
||||
+ */
|
||||
+ if (skb->nf_queue_entry && net_ratelimit())
|
||||
+ printk(KERN_WARNING
|
||||
+ "IMQ: kfree_skb: skb->nf_queue_entry != NULL");
|
||||
+#endif
|
||||
#if defined(CONFIG_NF_CONNTRACK) || defined(CONFIG_NF_CONNTRACK_MODULE)
|
||||
nf_conntrack_put(skb->nfct);
|
||||
nf_conntrack_put_reasm(skb->nfct_reasm);
|
||||
@@ -535,6 +621,9 @@ static void __copy_skb_header(struct sk_
|
||||
@@ -535,6 +635,9 @@ static void __copy_skb_header(struct sk_
|
||||
new->sp = secpath_get(old->sp);
|
||||
#endif
|
||||
memcpy(new->cb, old->cb, sizeof(old->cb));
|
||||
|
@ -1033,7 +1109,7 @@
|
|||
new->csum = old->csum;
|
||||
new->local_df = old->local_df;
|
||||
new->pkt_type = old->pkt_type;
|
||||
@@ -2776,6 +2865,13 @@ void __init skb_init(void)
|
||||
@@ -2776,6 +2879,13 @@ void __init skb_init(void)
|
||||
0,
|
||||
SLAB_HWCACHE_ALIGN|SLAB_PANIC,
|
||||
NULL);
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
--- a/include/linux/skbuff.h
|
||||
+++ b/include/linux/skbuff.h
|
||||
@@ -1383,11 +1383,18 @@ static inline int skb_network_offset(con
|
||||
@@ -1384,11 +1384,18 @@ static inline int skb_network_offset(con
|
||||
*
|
||||
* Various parts of the networking layer expect at least 32 bytes of
|
||||
* headroom, you should not reduce this.
|
||||
|
@ -19,7 +19,7 @@
|
|||
extern int ___pskb_trim(struct sk_buff *skb, unsigned int len);
|
||||
|
||||
static inline void __skb_trim(struct sk_buff *skb, unsigned int len)
|
||||
@@ -1477,9 +1484,9 @@ static inline void __skb_queue_purge(str
|
||||
@@ -1478,9 +1485,9 @@ static inline void __skb_queue_purge(str
|
||||
static inline struct sk_buff *__dev_alloc_skb(unsigned int length,
|
||||
gfp_t gfp_mask)
|
||||
{
|
||||
|
@ -31,7 +31,7 @@
|
|||
return skb;
|
||||
}
|
||||
|
||||
@@ -1552,7 +1559,7 @@ static inline int __skb_cow(struct sk_bu
|
||||
@@ -1553,7 +1560,7 @@ static inline int __skb_cow(struct sk_bu
|
||||
delta = headroom - skb_headroom(skb);
|
||||
|
||||
if (delta || cloned)
|
||||
|
@ -42,7 +42,7 @@
|
|||
}
|
||||
--- a/net/core/skbuff.c
|
||||
+++ b/net/core/skbuff.c
|
||||
@@ -336,9 +336,9 @@ struct sk_buff *__netdev_alloc_skb(struc
|
||||
@@ -339,9 +339,9 @@ struct sk_buff *__netdev_alloc_skb(struc
|
||||
int node = dev->dev.parent ? dev_to_node(dev->dev.parent) : -1;
|
||||
struct sk_buff *skb;
|
||||
|
||||
|
|
Loading…
Reference in a new issue