summaryrefslogtreecommitdiff
path: root/datapath/linux/compat/nf_conncount.c
diff options
context:
space:
mode:
authorGreg Rose <gvrose8192@gmail.com>2022-07-14 14:23:29 -0700
committerIlya Maximets <i.maximets@ovn.org>2022-07-15 13:45:55 +0200
commit422e90437854c441a0b274d3bf06ce7d6bbf9de0 (patch)
tree0e65141df83a618e59b00eb96e9c806bfbd51a39 /datapath/linux/compat/nf_conncount.c
parent86642de3ad8d42ab2b5dbace241d8dc398bd1195 (diff)
downloadopenvswitch-422e90437854c441a0b274d3bf06ce7d6bbf9de0.tar.gz
make: Remove the Linux datapath.
Update the necessary make and configure files to remove the Linux datapath and then remove the datapath. Move datapath/linux/compat/include/linux/openvswitch.h to include/linux/openvswitch.h because it is needed to generate header files used by the userspace switch. Also remove references to the Linux datapath from auxiliary files and utilities since it is no longer supported. Signed-off-by: Greg Rose <gvrose8192@gmail.com> Reviewed-by: David Marchand <david.marchand@redhat.com> Signed-off-by: Ilya Maximets <i.maximets@ovn.org>
Diffstat (limited to 'datapath/linux/compat/nf_conncount.c')
-rw-r--r--datapath/linux/compat/nf_conncount.c621
1 files changed, 0 insertions, 621 deletions
diff --git a/datapath/linux/compat/nf_conncount.c b/datapath/linux/compat/nf_conncount.c
deleted file mode 100644
index 97bdfb933..000000000
--- a/datapath/linux/compat/nf_conncount.c
+++ /dev/null
@@ -1,621 +0,0 @@
-/*
- * Backported from upstream commit 5c789e131cbb ("netfilter:
- * nf_conncount: Add list lock and gc worker, and RCU for init tree search")
- *
- * count the number of connections matching an arbitrary key.
- *
- * (C) 2017 Red Hat GmbH
- * Author: Florian Westphal <fw@strlen.de>
- *
- * split from xt_connlimit.c:
- * (c) 2000 Gerd Knorr <kraxel@bytesex.org>
- * Nov 2002: Martin Bene <martin.bene@icomedias.com>:
- * only ignore TIME_WAIT or gone connections
- * (C) CC Computer Consultants GmbH, 2007
- */
-#ifndef HAVE_UPSTREAM_NF_CONNCOUNT
-
-#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
-#include <linux/in.h>
-#include <linux/in6.h>
-#include <linux/ip.h>
-#include <linux/ipv6.h>
-#include <linux/jhash.h>
-#include <linux/slab.h>
-#include <linux/list.h>
-#include <linux/rbtree.h>
-#include <linux/module.h>
-#include <linux/random.h>
-#include <linux/skbuff.h>
-#include <linux/spinlock.h>
-#include <linux/netfilter/nf_conntrack_tcp.h>
-#include <linux/netfilter/x_tables.h>
-#include <net/netfilter/nf_conntrack.h>
-#include <net/netfilter/nf_conntrack_count.h>
-#include <net/netfilter/nf_conntrack_core.h>
-#include <net/netfilter/nf_conntrack_tuple.h>
-#include <net/netfilter/nf_conntrack_zones.h>
-
-#define CONNCOUNT_SLOTS 256U
-
-#define CONNCOUNT_GC_MAX_NODES 8
-#define MAX_KEYLEN 5
-
-/* we will save the tuples of all connections we care about */
-struct nf_conncount_tuple {
- struct list_head node;
- struct nf_conntrack_tuple tuple;
- struct nf_conntrack_zone zone;
- int cpu;
- u32 jiffies32;
-};
-
-struct nf_conncount_rb {
- struct rb_node node;
- struct nf_conncount_list list;
- u32 key[MAX_KEYLEN];
- struct rcu_head rcu_head;
-};
-
-static spinlock_t nf_conncount_locks[CONNCOUNT_SLOTS] __cacheline_aligned_in_smp;
-
-struct nf_conncount_data {
- unsigned int keylen;
- struct rb_root root[CONNCOUNT_SLOTS];
- struct net *net;
- struct work_struct gc_work;
- unsigned long pending_trees[BITS_TO_LONGS(CONNCOUNT_SLOTS)];
- unsigned int gc_tree;
-};
-
-static u_int32_t conncount_rnd __read_mostly;
-static struct kmem_cache *conncount_rb_cachep __read_mostly;
-static struct kmem_cache *conncount_conn_cachep __read_mostly;
-
-static inline bool already_closed(const struct nf_conn *conn)
-{
- if (nf_ct_protonum(conn) == IPPROTO_TCP)
- return conn->proto.tcp.state == TCP_CONNTRACK_TIME_WAIT ||
- conn->proto.tcp.state == TCP_CONNTRACK_CLOSE;
- else
- return false;
-}
-
-static int key_diff(const u32 *a, const u32 *b, unsigned int klen)
-{
- return memcmp(a, b, klen * sizeof(u32));
-}
-
-static void conn_free(struct nf_conncount_list *list,
- struct nf_conncount_tuple *conn)
-{
- lockdep_assert_held(&list->list_lock);
-
- list->count--;
- list_del(&conn->node);
-
- kmem_cache_free(conncount_conn_cachep, conn);
-}
-
-static const struct nf_conntrack_tuple_hash *
-find_or_evict(struct net *net, struct nf_conncount_list *list,
- struct nf_conncount_tuple *conn)
-{
- const struct nf_conntrack_tuple_hash *found;
- unsigned long a, b;
- int cpu = raw_smp_processor_id();
- u32 age;
-
- found = nf_conntrack_find_get(net, &conn->zone, &conn->tuple);
- if (found)
- return found;
- b = conn->jiffies32;
- a = (u32)jiffies;
-
- /* conn might have been added just before by another cpu and
- * might still be unconfirmed. In this case, nf_conntrack_find()
- * returns no result. Thus only evict if this cpu added the
- * stale entry or if the entry is older than two jiffies.
- */
- age = a - b;
- if (conn->cpu == cpu || age >= 2) {
- conn_free(list, conn);
- return ERR_PTR(-ENOENT);
- }
-
- return ERR_PTR(-EAGAIN);
-}
-
-static int __nf_conncount_add(struct net *net,
- struct nf_conncount_list *list,
- const struct nf_conntrack_tuple *tuple,
- const struct nf_conntrack_zone *zone)
-{
- const struct nf_conntrack_tuple_hash *found;
- struct nf_conncount_tuple *conn, *conn_n;
- struct nf_conn *found_ct;
- unsigned int collect = 0;
-
- /* check the saved connections */
- list_for_each_entry_safe(conn, conn_n, &list->head, node) {
- if (collect > CONNCOUNT_GC_MAX_NODES)
- break;
-
- found = find_or_evict(net, list, conn);
- if (IS_ERR(found)) {
- /* Not found, but might be about to be confirmed */
- if (PTR_ERR(found) == -EAGAIN) {
- if (nf_ct_tuple_equal(&conn->tuple, tuple) &&
- nf_ct_zone_id(&conn->zone, conn->zone.dir) ==
- nf_ct_zone_id(zone, zone->dir))
- return 0; /* already exists */
- } else {
- collect++;
- }
- continue;
- }
-
- found_ct = nf_ct_tuplehash_to_ctrack(found);
-
- if (nf_ct_tuple_equal(&conn->tuple, tuple) &&
- nf_ct_zone_equal(found_ct, zone, zone->dir)) {
- /*
- * We should not see tuples twice unless someone hooks
- * this into a table without "-p tcp --syn".
- *
- * Attempt to avoid a re-add in this case.
- */
- nf_ct_put(found_ct);
- return 0;
- } else if (already_closed(found_ct)) {
- /*
- * we do not care about connections which are
- * closed already -> ditch it
- */
- nf_ct_put(found_ct);
- conn_free(list, conn);
- collect++;
- continue;
- }
-
- nf_ct_put(found_ct);
- }
-
- if (WARN_ON_ONCE(list->count > INT_MAX))
- return -EOVERFLOW;
-
- conn = kmem_cache_alloc(conncount_conn_cachep, GFP_ATOMIC);
- if (conn == NULL)
- return -ENOMEM;
-
- conn->tuple = *tuple;
- conn->zone = *zone;
- conn->cpu = raw_smp_processor_id();
- conn->jiffies32 = (u32)jiffies;
- list_add_tail(&conn->node, &list->head);
- list->count++;
- return 0;
-}
-
-int nf_conncount_add(struct net *net,
- struct nf_conncount_list *list,
- const struct nf_conntrack_tuple *tuple,
- const struct nf_conntrack_zone *zone)
-{
- int ret;
-
- /* check the saved connections */
- spin_lock_bh(&list->list_lock);
- ret = __nf_conncount_add(net, list, tuple, zone);
- spin_unlock_bh(&list->list_lock);
-
- return ret;
-}
-
-static void nf_conncount_list_init(struct nf_conncount_list *list)
-{
- spin_lock_init(&list->list_lock);
- INIT_LIST_HEAD(&list->head);
- list->count = 0;
-}
-
-/* Return true if the list is empty. Must be called with BH disabled. */
-static bool nf_conncount_gc_list(struct net *net,
- struct nf_conncount_list *list)
-{
- const struct nf_conntrack_tuple_hash *found;
- struct nf_conncount_tuple *conn, *conn_n;
- struct nf_conn *found_ct;
- unsigned int collected = 0;
- bool ret = false;
-
- /* don't bother if other cpu is already doing GC */
- if (!spin_trylock(&list->list_lock))
- return false;
-
- list_for_each_entry_safe(conn, conn_n, &list->head, node) {
- found = find_or_evict(net, list, conn);
- if (IS_ERR(found)) {
- if (PTR_ERR(found) == -ENOENT)
- collected++;
- continue;
- }
-
- found_ct = nf_ct_tuplehash_to_ctrack(found);
- if (already_closed(found_ct)) {
- /*
- * we do not care about connections which are
- * closed already -> ditch it
- */
- nf_ct_put(found_ct);
- conn_free(list, conn);
- collected++;
- continue;
- }
-
- nf_ct_put(found_ct);
- if (collected > CONNCOUNT_GC_MAX_NODES)
- break;
- }
-
- if (!list->count)
- ret = true;
- spin_unlock(&list->list_lock);
-
- return ret;
-}
-
-static void __tree_nodes_free(struct rcu_head *h)
-{
- struct nf_conncount_rb *rbconn;
-
- rbconn = container_of(h, struct nf_conncount_rb, rcu_head);
- kmem_cache_free(conncount_rb_cachep, rbconn);
-}
-
-/* caller must hold tree nf_conncount_locks[] lock */
-static void tree_nodes_free(struct rb_root *root,
- struct nf_conncount_rb *gc_nodes[],
- unsigned int gc_count)
-{
- struct nf_conncount_rb *rbconn;
-
- while (gc_count) {
- rbconn = gc_nodes[--gc_count];
- spin_lock(&rbconn->list.list_lock);
- if (!rbconn->list.count) {
- rb_erase(&rbconn->node, root);
- call_rcu(&rbconn->rcu_head, __tree_nodes_free);
- }
- spin_unlock(&rbconn->list.list_lock);
- }
-}
-
-static void schedule_gc_worker(struct nf_conncount_data *data, int tree)
-{
- set_bit(tree, data->pending_trees);
- schedule_work(&data->gc_work);
-}
-
-static unsigned int
-insert_tree(struct net *net,
- struct nf_conncount_data *data,
- struct rb_root *root,
- unsigned int hash,
- const u32 *key,
- const struct nf_conntrack_tuple *tuple,
- const struct nf_conntrack_zone *zone)
-{
- struct nf_conncount_rb *gc_nodes[CONNCOUNT_GC_MAX_NODES];
- struct rb_node **rbnode, *parent;
- struct nf_conncount_rb *rbconn;
- struct nf_conncount_tuple *conn;
- unsigned int count = 0, gc_count = 0;
- u8 keylen = data->keylen;
- bool do_gc = true;
-
- spin_lock_bh(&nf_conncount_locks[hash]);
-restart:
- parent = NULL;
- rbnode = &(root->rb_node);
- while (*rbnode) {
- int diff;
- rbconn = rb_entry(*rbnode, struct nf_conncount_rb, node);
-
- parent = *rbnode;
- diff = key_diff(key, rbconn->key, keylen);
- if (diff < 0) {
- rbnode = &((*rbnode)->rb_left);
- } else if (diff > 0) {
- rbnode = &((*rbnode)->rb_right);
- } else {
- int ret;
-
- ret = nf_conncount_add(net, &rbconn->list, tuple, zone);
- if (ret)
- count = 0; /* hotdrop */
- else
- count = rbconn->list.count;
- tree_nodes_free(root, gc_nodes, gc_count);
- goto out_unlock;
- }
-
- if (gc_count >= ARRAY_SIZE(gc_nodes))
- continue;
-
- if (do_gc && nf_conncount_gc_list(net, &rbconn->list))
- gc_nodes[gc_count++] = rbconn;
- }
-
- if (gc_count) {
- tree_nodes_free(root, gc_nodes, gc_count);
- schedule_gc_worker(data, hash);
- gc_count = 0;
- do_gc = false;
- goto restart;
- }
-
- /* expected case: match, insert new node */
- rbconn = kmem_cache_alloc(conncount_rb_cachep, GFP_ATOMIC);
- if (rbconn == NULL)
- goto out_unlock;
-
- conn = kmem_cache_alloc(conncount_conn_cachep, GFP_ATOMIC);
- if (conn == NULL) {
- kmem_cache_free(conncount_rb_cachep, rbconn);
- goto out_unlock;
- }
-
- conn->tuple = *tuple;
- conn->zone = *zone;
- memcpy(rbconn->key, key, sizeof(u32) * keylen);
-
- nf_conncount_list_init(&rbconn->list);
- list_add(&conn->node, &rbconn->list.head);
- count = 1;
- rbconn->list.count = count;
-
- rb_link_node_rcu(&rbconn->node, parent, rbnode);
- rb_insert_color(&rbconn->node, root);
-out_unlock:
- spin_unlock_bh(&nf_conncount_locks[hash]);
- return count;
-}
-
-static unsigned int
-count_tree(struct net *net,
- struct nf_conncount_data *data,
- const u32 *key,
- const struct nf_conntrack_tuple *tuple,
- const struct nf_conntrack_zone *zone)
-{
- struct rb_root *root;
- struct rb_node *parent;
- struct nf_conncount_rb *rbconn;
- unsigned int hash;
- u8 keylen = data->keylen;
-
- hash = jhash2(key, data->keylen, conncount_rnd) % CONNCOUNT_SLOTS;
- root = &data->root[hash];
-
- parent = rcu_dereference_raw(root->rb_node);
- while (parent) {
- int diff;
-
- rbconn = rb_entry(parent, struct nf_conncount_rb, node);
-
- diff = key_diff(key, rbconn->key, keylen);
- if (diff < 0) {
- parent = rcu_dereference_raw(parent->rb_left);
- } else if (diff > 0) {
- parent = rcu_dereference_raw(parent->rb_right);
- } else {
- int ret;
-
- if (!tuple) {
- nf_conncount_gc_list(net, &rbconn->list);
- return rbconn->list.count;
- }
-
- spin_lock_bh(&rbconn->list.list_lock);
- /* Node might be about to be free'd.
- * We need to defer to insert_tree() in this case.
- */
- if (rbconn->list.count == 0) {
- spin_unlock_bh(&rbconn->list.list_lock);
- break;
- }
-
- /* same source network -> be counted! */
- ret = __nf_conncount_add(net, &rbconn->list, tuple, zone);
- spin_unlock_bh(&rbconn->list.list_lock);
- if (ret)
- return 0; /* hotdrop */
- else
- return rbconn->list.count;
- }
- }
-
- if (!tuple)
- return 0;
-
- return insert_tree(net, data, root, hash, key, tuple, zone);
-}
-
-static void tree_gc_worker(struct work_struct *work)
-{
- struct nf_conncount_data *data = container_of(work, struct nf_conncount_data, gc_work);
- struct nf_conncount_rb *gc_nodes[CONNCOUNT_GC_MAX_NODES], *rbconn;
- struct rb_root *root;
- struct rb_node *node;
- unsigned int tree, next_tree, gc_count = 0;
-
- tree = data->gc_tree % CONNCOUNT_SLOTS;
- root = &data->root[tree];
-
- local_bh_disable();
- rcu_read_lock();
- for (node = rb_first(root); node != NULL; node = rb_next(node)) {
- rbconn = rb_entry(node, struct nf_conncount_rb, node);
- if (nf_conncount_gc_list(data->net, &rbconn->list))
- gc_count++;
- }
- rcu_read_unlock();
- local_bh_enable();
-
- cond_resched();
-
- spin_lock_bh(&nf_conncount_locks[tree]);
- if (gc_count < ARRAY_SIZE(gc_nodes))
- goto next; /* do not bother */
-
- gc_count = 0;
- node = rb_first(root);
- while (node != NULL) {
- rbconn = rb_entry(node, struct nf_conncount_rb, node);
- node = rb_next(node);
-
- if (rbconn->list.count > 0)
- continue;
-
- gc_nodes[gc_count++] = rbconn;
- if (gc_count >= ARRAY_SIZE(gc_nodes)) {
- tree_nodes_free(root, gc_nodes, gc_count);
- gc_count = 0;
- }
- }
-
- tree_nodes_free(root, gc_nodes, gc_count);
-next:
-
- clear_bit(tree, data->pending_trees);
-
- next_tree = (tree + 1) % CONNCOUNT_SLOTS;
- next_tree = find_next_bit(data->pending_trees, CONNCOUNT_SLOTS, next_tree);
-
- if (next_tree < CONNCOUNT_SLOTS) {
- data->gc_tree = next_tree;
- schedule_work(work);
- }
-
- spin_unlock_bh(&nf_conncount_locks[tree]);
-}
-
-/* Count and return number of conntrack entries in 'net' with particular 'key'.
- * If 'tuple' is not null, insert it into the accounting data structure.
- * Call with RCU read lock.
- */
-unsigned int rpl_nf_conncount_count(struct net *net,
- struct nf_conncount_data *data,
- const u32 *key,
- const struct nf_conntrack_tuple *tuple,
- const struct nf_conntrack_zone *zone)
-{
- return count_tree(net, data, key, tuple, zone);
-}
-EXPORT_SYMBOL_GPL(rpl_nf_conncount_count);
-
-struct nf_conncount_data *rpl_nf_conncount_init(struct net *net, unsigned int family,
- unsigned int keylen)
-{
- struct nf_conncount_data *data;
- int ret, i;
-
- if (keylen % sizeof(u32) ||
- keylen / sizeof(u32) > MAX_KEYLEN ||
- keylen == 0)
- return ERR_PTR(-EINVAL);
-
- net_get_random_once(&conncount_rnd, sizeof(conncount_rnd));
-
- data = kmalloc(sizeof(*data), GFP_KERNEL);
- if (!data)
- return ERR_PTR(-ENOMEM);
-
- ret = nf_ct_netns_get(net, family);
- if (ret < 0) {
- kfree(data);
- return ERR_PTR(ret);
- }
-
- for (i = 0; i < ARRAY_SIZE(data->root); ++i)
- data->root[i] = RB_ROOT;
-
- data->keylen = keylen / sizeof(u32);
- data->net = net;
- INIT_WORK(&data->gc_work, tree_gc_worker);
-
- return data;
-}
-EXPORT_SYMBOL_GPL(rpl_nf_conncount_init);
-
-static void nf_conncount_cache_free(struct nf_conncount_list *list)
-{
- struct nf_conncount_tuple *conn, *conn_n;
-
- list_for_each_entry_safe(conn, conn_n, &list->head, node)
- kmem_cache_free(conncount_conn_cachep, conn);
-}
-
-static void destroy_tree(struct rb_root *r)
-{
- struct nf_conncount_rb *rbconn;
- struct rb_node *node;
-
- while ((node = rb_first(r)) != NULL) {
- rbconn = rb_entry(node, struct nf_conncount_rb, node);
-
- rb_erase(node, r);
-
- nf_conncount_cache_free(&rbconn->list);
-
- kmem_cache_free(conncount_rb_cachep, rbconn);
- }
-}
-
-void rpl_nf_conncount_destroy(struct net *net, unsigned int family,
- struct nf_conncount_data *data)
-{
- unsigned int i;
-
- cancel_work_sync(&data->gc_work);
- nf_ct_netns_put(net, family);
-
- for (i = 0; i < ARRAY_SIZE(data->root); ++i)
- destroy_tree(&data->root[i]);
-
- kfree(data);
-}
-EXPORT_SYMBOL_GPL(rpl_nf_conncount_destroy);
-
-int rpl_nf_conncount_modinit(void)
-{
- int i;
-
- for (i = 0; i < CONNCOUNT_SLOTS; ++i)
- spin_lock_init(&nf_conncount_locks[i]);
-
- conncount_conn_cachep = kmem_cache_create("nf_conncount_tuple",
- sizeof(struct nf_conncount_tuple),
- 0, 0, NULL);
- if (!conncount_conn_cachep)
- return -ENOMEM;
-
- conncount_rb_cachep = kmem_cache_create("nf_conncount_rb",
- sizeof(struct nf_conncount_rb),
- 0, 0, NULL);
- if (!conncount_rb_cachep) {
- kmem_cache_destroy(conncount_conn_cachep);
- return -ENOMEM;
- }
-
- return 0;
-}
-
-void rpl_nf_conncount_modexit(void)
-{
- kmem_cache_destroy(conncount_conn_cachep);
- kmem_cache_destroy(conncount_rb_cachep);
-}
-
-#endif /* HAVE_UPSTREAM_NF_CONNCOUNT */