Staging
v0.5.1
https://github.com/torvalds/linux
Revision 97fd8da281f80e7e69e0114bc906575734d4dfaf authored by Maor Gottlieb on 05 September 2019, 06:56:10 UTC, committed by Saeed Mahameed on 20 November 2019, 20:33:06 UTC
Once all the large flow groups (defined by the user when the flow table
is created - max_num_groups) were created, then all the following new
flow groups will have only one flow table entry, even though the flow table
has place to larger groups.
Fix the condition to prefer large flow group.

Fixes: f0d22d187473 ("net/mlx5_core: Introduce flow steering autogrouped flow table")
Signed-off-by: Maor Gottlieb <maorg@mellanox.com>
Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
1 parent 9c98f7e
Raw File
Tip revision: 97fd8da281f80e7e69e0114bc906575734d4dfaf authored by Maor Gottlieb on 05 September 2019, 06:56:10 UTC
net/mlx5: Fix auto group size calculation
Tip revision: 97fd8da
bucket_locks.c
#include <linux/export.h>
#include <linux/kernel.h>
#include <linux/mm.h>
#include <linux/slab.h>
#include <linux/vmalloc.h>

/* Allocate an array of spinlocks to be accessed by a hash. Two arguments
 * indicate the number of elements to allocate in the array. max_size
 * gives the maximum number of elements to allocate. cpu_mult gives
 * the number of locks per CPU to allocate. The size is rounded up
 * to a power of 2 to be suitable as a hash table.
 */

int __alloc_bucket_spinlocks(spinlock_t **locks, unsigned int *locks_mask,
			     size_t max_size, unsigned int cpu_mult, gfp_t gfp,
			     const char *name, struct lock_class_key *key)
{
	spinlock_t *tlocks = NULL;
	unsigned int i, size;
#if defined(CONFIG_PROVE_LOCKING)
	unsigned int nr_pcpus = 2;
#else
	unsigned int nr_pcpus = num_possible_cpus();
#endif

	if (cpu_mult) {
		nr_pcpus = min_t(unsigned int, nr_pcpus, 64UL);
		size = min_t(unsigned int, nr_pcpus * cpu_mult, max_size);
	} else {
		size = max_size;
	}

	if (sizeof(spinlock_t) != 0) {
		tlocks = kvmalloc_array(size, sizeof(spinlock_t), gfp);
		if (!tlocks)
			return -ENOMEM;
		for (i = 0; i < size; i++) {
			spin_lock_init(&tlocks[i]);
			lockdep_init_map(&tlocks[i].dep_map, name, key, 0);
		}
	}

	*locks = tlocks;
	*locks_mask = size - 1;

	return 0;
}
EXPORT_SYMBOL(__alloc_bucket_spinlocks);

void free_bucket_spinlocks(spinlock_t *locks)
{
	kvfree(locks);
}
EXPORT_SYMBOL(free_bucket_spinlocks);
back to top