2 * Distributed under the terms of the GNU GPL version 2.
3 * Copyright (c) 2007, 2008 The Board of Trustees of The Leland
4 * Stanford Junior University
12 #include <linux/slab.h>
13 #include <linux/vmalloc.h>
15 #include <linux/highmem.h>
16 #include <asm/pgtable.h>
18 static void *kmem_alloc(size_t);
19 static void *kmem_zalloc(size_t);
20 static void kmem_free(void *, size_t);
22 struct sw_table_hash {
26 unsigned int bucket_mask; /* Number of buckets minus 1. */
27 struct sw_flow **buckets;
30 static struct sw_flow **find_bucket(struct sw_table *swt,
31 const struct sw_flow_key *key)
33 struct sw_table_hash *th = (struct sw_table_hash *) swt;
34 unsigned int crc = crc32_calculate(&th->crc32, key, sizeof *key);
35 return &th->buckets[crc & th->bucket_mask];
38 static struct sw_flow *table_hash_lookup(struct sw_table *swt,
39 const struct sw_flow_key *key)
41 struct sw_flow *flow = *find_bucket(swt, key);
42 return flow && !memcmp(&flow->key, key, sizeof *key) ? flow : NULL;
45 static int table_hash_insert(struct sw_table *swt, struct sw_flow *flow)
47 struct sw_table_hash *th = (struct sw_table_hash *) swt;
48 struct sw_flow **bucket;
51 if (flow->key.wildcards != 0)
54 bucket = find_bucket(swt, &flow->key);
55 if (*bucket == NULL) {
56 atomic_inc(&th->n_flows);
57 rcu_assign_pointer(*bucket, flow);
60 struct sw_flow *old_flow = *bucket;
61 if (!memcmp(&old_flow->key, &flow->key, sizeof flow->key)
62 && flow_del(old_flow)) {
63 rcu_assign_pointer(*bucket, flow);
64 flow_deferred_free(old_flow);
73 /* Caller must update n_flows. */
74 static int do_delete(struct sw_flow **bucket, struct sw_flow *flow)
77 rcu_assign_pointer(*bucket, NULL);
78 flow_deferred_free(flow);
84 /* Returns number of deleted flows. We can ignore the priority
85 * argument, since all exact-match entries are the same (highest)
87 static int table_hash_delete(struct sw_table *swt,
88 const struct sw_flow_key *key,
89 uint16_t priority, int strict)
91 struct sw_table_hash *th = (struct sw_table_hash *) swt;
92 unsigned int count = 0;
94 if (key->wildcards == 0) {
95 struct sw_flow **bucket = find_bucket(swt, key);
96 struct sw_flow *flow = *bucket;
97 if (flow && !memcmp(&flow->key, key, sizeof *key))
98 count = do_delete(bucket, flow);
102 for (i = 0; i <= th->bucket_mask; i++) {
103 struct sw_flow **bucket = &th->buckets[i];
104 struct sw_flow *flow = *bucket;
105 if (flow && flow_del_matches(&flow->key, key, strict))
106 count += do_delete(bucket, flow);
110 atomic_sub(count, &th->n_flows);
114 static int table_hash_timeout(struct datapath *dp, struct sw_table *swt)
116 struct sw_table_hash *th = (struct sw_table_hash *) swt;
120 mutex_lock(&dp_mutex);
121 for (i = 0; i <= th->bucket_mask; i++) {
122 struct sw_flow **bucket = &th->buckets[i];
123 struct sw_flow *flow = *bucket;
124 if (flow && flow_timeout(flow)) {
125 count += do_delete(bucket, flow);
126 if (dp->flags & OFPC_SEND_FLOW_EXP)
127 dp_send_flow_expired(dp, flow);
130 mutex_unlock(&dp_mutex);
133 atomic_sub(count, &th->n_flows);
137 static void table_hash_destroy(struct sw_table *swt)
139 struct sw_table_hash *th = (struct sw_table_hash *) swt;
141 for (i = 0; i <= th->bucket_mask; i++)
143 flow_free(th->buckets[i]);
144 kmem_free(th->buckets, (th->bucket_mask + 1) * sizeof *th->buckets);
148 static int table_hash_iterate(struct sw_table *swt,
149 const struct sw_flow_key *key,
150 struct sw_table_position *position,
151 int (*callback)(struct sw_flow *, void *private),
154 struct sw_table_hash *th = (struct sw_table_hash *) swt;
156 if (position->private[0] > th->bucket_mask)
159 if (key->wildcards == 0) {
160 struct sw_flow *flow;
163 flow = table_hash_lookup(swt, key);
167 error = callback(flow, private);
169 position->private[0] = -1;
174 for (i = position->private[0]; i <= th->bucket_mask; i++) {
175 struct sw_flow *flow = th->buckets[i];
176 if (flow && flow_matches(key, &flow->key)) {
177 int error = callback(flow, private);
179 position->private[0] = i;
187 static void table_hash_stats(struct sw_table *swt,
188 struct sw_table_stats *stats)
190 struct sw_table_hash *th = (struct sw_table_hash *) swt;
191 stats->name = "hash";
192 stats->n_flows = atomic_read(&th->n_flows);
193 stats->max_flows = th->bucket_mask + 1;
196 struct sw_table *table_hash_create(unsigned int polynomial,
197 unsigned int n_buckets)
199 struct sw_table_hash *th;
200 struct sw_table *swt;
202 th = kmalloc(sizeof *th, GFP_KERNEL);
206 BUG_ON(n_buckets & (n_buckets - 1));
207 th->buckets = kmem_zalloc(n_buckets * sizeof *th->buckets);
208 if (th->buckets == NULL) {
209 printk("failed to allocate %u buckets\n", n_buckets);
213 th->bucket_mask = n_buckets - 1;
216 swt->lookup = table_hash_lookup;
217 swt->insert = table_hash_insert;
218 swt->delete = table_hash_delete;
219 swt->timeout = table_hash_timeout;
220 swt->destroy = table_hash_destroy;
221 swt->iterate = table_hash_iterate;
222 swt->stats = table_hash_stats;
224 crc32_init(&th->crc32, polynomial);
225 atomic_set(&th->n_flows, 0);
230 /* Double-hashing table. */
232 struct sw_table_hash2 {
234 struct sw_table *subtable[2];
237 static struct sw_flow *table_hash2_lookup(struct sw_table *swt,
238 const struct sw_flow_key *key)
240 struct sw_table_hash2 *t2 = (struct sw_table_hash2 *) swt;
243 for (i = 0; i < 2; i++) {
244 struct sw_flow *flow = *find_bucket(t2->subtable[i], key);
245 if (flow && !memcmp(&flow->key, key, sizeof *key))
251 static int table_hash2_insert(struct sw_table *swt, struct sw_flow *flow)
253 struct sw_table_hash2 *t2 = (struct sw_table_hash2 *) swt;
255 if (table_hash_insert(t2->subtable[0], flow))
257 return table_hash_insert(t2->subtable[1], flow);
260 static int table_hash2_delete(struct sw_table *swt,
261 const struct sw_flow_key *key,
262 uint16_t priority, int strict)
264 struct sw_table_hash2 *t2 = (struct sw_table_hash2 *) swt;
265 return (table_hash_delete(t2->subtable[0], key, priority, strict)
266 + table_hash_delete(t2->subtable[1], key, priority, strict));
269 static int table_hash2_timeout(struct datapath *dp, struct sw_table *swt)
271 struct sw_table_hash2 *t2 = (struct sw_table_hash2 *) swt;
272 return (table_hash_timeout(dp, t2->subtable[0])
273 + table_hash_timeout(dp, t2->subtable[1]));
276 static void table_hash2_destroy(struct sw_table *swt)
278 struct sw_table_hash2 *t2 = (struct sw_table_hash2 *) swt;
279 table_hash_destroy(t2->subtable[0]);
280 table_hash_destroy(t2->subtable[1]);
284 static int table_hash2_iterate(struct sw_table *swt,
285 const struct sw_flow_key *key,
286 struct sw_table_position *position,
287 int (*callback)(struct sw_flow *, void *),
290 struct sw_table_hash2 *t2 = (struct sw_table_hash2 *) swt;
293 for (i = position->private[1]; i < 2; i++) {
294 int error = table_hash_iterate(t2->subtable[i], key, position,
299 position->private[0] = 0;
300 position->private[1]++;
305 static void table_hash2_stats(struct sw_table *swt,
306 struct sw_table_stats *stats)
308 struct sw_table_hash2 *t2 = (struct sw_table_hash2 *) swt;
309 struct sw_table_stats substats[2];
312 for (i = 0; i < 2; i++)
313 table_hash_stats(t2->subtable[i], &substats[i]);
314 stats->name = "hash2";
315 stats->n_flows = substats[0].n_flows + substats[1].n_flows;
316 stats->max_flows = substats[0].max_flows + substats[1].max_flows;
319 struct sw_table *table_hash2_create(unsigned int poly0, unsigned int buckets0,
320 unsigned int poly1, unsigned int buckets1)
323 struct sw_table_hash2 *t2;
324 struct sw_table *swt;
326 t2 = kmalloc(sizeof *t2, GFP_KERNEL);
330 t2->subtable[0] = table_hash_create(poly0, buckets0);
331 if (t2->subtable[0] == NULL)
334 t2->subtable[1] = table_hash_create(poly1, buckets1);
335 if (t2->subtable[1] == NULL)
336 goto out_free_subtable0;
339 swt->lookup = table_hash2_lookup;
340 swt->insert = table_hash2_insert;
341 swt->delete = table_hash2_delete;
342 swt->timeout = table_hash2_timeout;
343 swt->destroy = table_hash2_destroy;
344 swt->iterate = table_hash2_iterate;
345 swt->stats = table_hash2_stats;
350 table_hash_destroy(t2->subtable[0]);
356 /* From fs/xfs/linux-2.4/kmem.c. */
359 kmem_alloc(size_t size)
363 #ifdef KMALLOC_MAX_SIZE
364 if (size > KMALLOC_MAX_SIZE)
367 ptr = kmalloc(size, GFP_KERNEL);
371 printk("openflow: used vmalloc for %lu bytes\n",
372 (unsigned long)size);
378 kmem_zalloc(size_t size)
380 void *ptr = kmem_alloc(size);
382 memset(ptr, 0, size);
387 kmem_free(void *ptr, size_t size)
389 if (((unsigned long)ptr < VMALLOC_START) ||
390 ((unsigned long)ptr >= VMALLOC_END)) {