git: 81b38bce0794 - main - mlx5e tls: Ensure all allocated tags have a hw context associated
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Sat, 23 Nov 2024 10:05:07 UTC
The branch main has been updated by kib:
URL: https://cgit.FreeBSD.org/src/commit/?id=81b38bce07940b7a4001dfeb8cd63774229ca950
commit 81b38bce07940b7a4001dfeb8cd63774229ca950
Author: Andrew Gallatin <gallatin@FreeBSD.org>
AuthorDate: 2024-11-23 09:43:17 +0000
Commit: Konstantin Belousov <kib@FreeBSD.org>
CommitDate: 2024-11-23 10:01:50 +0000
mlx5e tls: Ensure all allocated tags have a hw context associated
Ensure all allocated tags have a hardware context associated.
The hardware context allocation is moved into the zone import
routine, as suggested by kib. This is safe because these zone
allocations are always done in a sleepable context.
I have removed the now pointless num_resources tracking,
and added sysctls / tunables to control UMA zone limits
for these tls tags, as well as a tunable to let the
driver pre-allocate tags at boot.
MFC after: 2 weeks
---
sys/dev/mlx5/mlx5_en/en_hw_tls.h | 2 +-
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c | 132 ++++++++++++++++++++++++----------
2 files changed, 95 insertions(+), 39 deletions(-)
diff --git a/sys/dev/mlx5/mlx5_en/en_hw_tls.h b/sys/dev/mlx5/mlx5_en/en_hw_tls.h
index 2018198e5e52..d637314e040e 100644
--- a/sys/dev/mlx5/mlx5_en/en_hw_tls.h
+++ b/sys/dev/mlx5/mlx5_en/en_hw_tls.h
@@ -84,7 +84,7 @@ struct mlx5e_tls {
struct workqueue_struct *wq;
uma_zone_t zone;
uint32_t max_resources; /* max number of resources */
- volatile uint32_t num_resources; /* current number of resources */
+ int zone_max;
int init; /* set when ready */
char zname[32];
};
diff --git a/sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c b/sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
index b5caa3ba53dd..8b5121066bdf 100644
--- a/sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
+++ b/sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
@@ -81,17 +81,57 @@ static const char *mlx5e_tls_stats_desc[] = {
static void mlx5e_tls_work(struct work_struct *);
+/*
+ * Expand the tls tag UMA zone in a sleepable context
+ */
+
+static void
+mlx5e_prealloc_tags(struct mlx5e_priv *priv, int nitems)
+{
+ struct mlx5e_tls_tag **tags;
+ int i;
+
+ tags = malloc(sizeof(tags[0]) * nitems,
+ M_MLX5E_TLS, M_WAITOK);
+ for (i = 0; i < nitems; i++)
+ tags[i] = uma_zalloc(priv->tls.zone, M_WAITOK);
+ __compiler_membar();
+ for (i = 0; i < nitems; i++)
+ uma_zfree(priv->tls.zone, tags[i]);
+ free(tags, M_MLX5E_TLS);
+}
+
static int
mlx5e_tls_tag_import(void *arg, void **store, int cnt, int domain, int flags)
{
struct mlx5e_tls_tag *ptag;
- int i;
+ struct mlx5e_priv *priv = arg;
+ int err, i;
+
+ /*
+ * mlx5_tls_open_tis() sleeps on a firmware command, so
+ * zone allocations must be done from a sleepable context.
+ * Note that the uma_zalloc() in mlx5e_tls_snd_tag_alloc()
+ * is done with M_NOWAIT so that hitting the zone limit does
+ * not cause the allocation to pause forever.
+ */
for (i = 0; i != cnt; i++) {
ptag = malloc_domainset(sizeof(*ptag), M_MLX5E_TLS,
mlx5_dev_domainset(arg), flags | M_ZERO);
+ if (ptag == NULL)
+ return (i);
+ ptag->tls = &priv->tls;
mtx_init(&ptag->mtx, "mlx5-tls-tag-mtx", NULL, MTX_DEF);
INIT_WORK(&ptag->work, mlx5e_tls_work);
+ err = mlx5_tls_open_tis(priv->mdev, 0, priv->tdn,
+ priv->pdn, &ptag->tisn);
+ if (err) {
+ MLX5E_TLS_STAT_INC(ptag, tx_error, 1);
+ free(ptag, M_MLX5E_TLS);
+ return (i);
+ }
+
store[i] = ptag;
}
return (i);
@@ -114,7 +154,6 @@ mlx5e_tls_tag_release(void *arg, void **store, int cnt)
if (ptag->tisn != 0) {
mlx5_tls_close_tis(priv->mdev, ptag->tisn);
- atomic_add_32(&ptls->num_resources, -1U);
}
mtx_destroy(&ptag->mtx);
@@ -136,20 +175,38 @@ mlx5e_tls_tag_zfree(struct mlx5e_tls_tag *ptag)
/* avoid leaking keys */
memset(ptag->crypto_params, 0, sizeof(ptag->crypto_params));
- /* update number of TIS contexts */
- if (ptag->tisn == 0)
- atomic_add_32(&ptag->tls->num_resources, -1U);
-
/* return tag to UMA */
uma_zfree(ptag->tls->zone, ptag);
}
+static int
+mlx5e_max_tag_proc(SYSCTL_HANDLER_ARGS)
+{
+ struct mlx5e_priv *priv = (struct mlx5e_priv *)arg1;
+ struct mlx5e_tls *ptls = &priv->tls;
+ int err;
+ unsigned int max_tags;
+
+ max_tags = ptls->zone_max;
+ err = sysctl_handle_int(oidp, &max_tags, arg2, req);
+ if (err != 0 || req->newptr == NULL )
+ return err;
+ if (max_tags == ptls->zone_max)
+ return 0;
+ if (max_tags > priv->tls.max_resources || max_tags == 0)
+ return (EINVAL);
+ ptls->zone_max = max_tags;
+ uma_zone_set_max(ptls->zone, ptls->zone_max);
+ return 0;
+}
+
int
mlx5e_tls_init(struct mlx5e_priv *priv)
{
struct mlx5e_tls *ptls = &priv->tls;
struct sysctl_oid *node;
- uint32_t x;
+ uint32_t max_dek, max_tis, x;
+ int zone_max = 0, prealloc_tags = 0;
if (MLX5_CAP_GEN(priv->mdev, tls_tx) == 0 ||
MLX5_CAP_GEN(priv->mdev, log_max_dek) == 0)
@@ -164,13 +221,31 @@ mlx5e_tls_init(struct mlx5e_priv *priv)
snprintf(ptls->zname, sizeof(ptls->zname),
"mlx5_%u_tls", device_get_unit(priv->mdev->pdev->dev.bsddev));
+
+ TUNABLE_INT_FETCH("hw.mlx5.tls_max_tags", &zone_max);
+ TUNABLE_INT_FETCH("hw.mlx5.tls_prealloc_tags", &prealloc_tags);
+
ptls->zone = uma_zcache_create(ptls->zname,
sizeof(struct mlx5e_tls_tag), NULL, NULL, NULL, NULL,
- mlx5e_tls_tag_import, mlx5e_tls_tag_release, priv->mdev,
- UMA_ZONE_UNMANAGED);
+ mlx5e_tls_tag_import, mlx5e_tls_tag_release, priv,
+ UMA_ZONE_UNMANAGED | (prealloc_tags ? UMA_ZONE_NOFREE : 0));
/* shared between RX and TX TLS */
- ptls->max_resources = 1U << (MLX5_CAP_GEN(priv->mdev, log_max_dek) - 1);
+ max_dek = 1U << (MLX5_CAP_GEN(priv->mdev, log_max_dek) - 1);
+ max_tis = 1U << (MLX5_CAP_GEN(priv->mdev, log_max_tis) - 1);
+ ptls->max_resources = MIN(max_dek, max_tis);
+
+ if (zone_max != 0) {
+ ptls->zone_max = zone_max;
+ if (ptls->zone_max > priv->tls.max_resources)
+ ptls->zone_max = priv->tls.max_resources;
+ } else {
+ ptls->zone_max = priv->tls.max_resources;
+ }
+
+ uma_zone_set_max(ptls->zone, ptls->zone_max);
+ if (prealloc_tags != 0)
+ mlx5e_prealloc_tags(priv, ptls->zone_max);
for (x = 0; x != MLX5E_TLS_STATS_NUM; x++)
ptls->stats.arg[x] = counter_u64_alloc(M_WAITOK);
@@ -183,6 +258,10 @@ mlx5e_tls_init(struct mlx5e_priv *priv)
if (node == NULL)
return (0);
+ SYSCTL_ADD_PROC(&priv->sysctl_ctx, SYSCTL_CHILDREN(node), OID_AUTO, "tls_max_tag",
+ CTLFLAG_RW | CTLTYPE_UINT | CTLFLAG_MPSAFE, priv, 0, mlx5e_max_tag_proc,
+ "IU", "Max number of TLS offload session tags");
+
mlx5e_create_counter_stats(&ptls->ctx,
SYSCTL_CHILDREN(node), "stats",
mlx5e_tls_stats_desc, MLX5E_TLS_STATS_NUM,
@@ -206,9 +285,6 @@ mlx5e_tls_cleanup(struct mlx5e_priv *priv)
uma_zdestroy(ptls->zone);
destroy_workqueue(ptls->wq);
- /* check if all resources are freed */
- MPASS(priv->tls.num_resources == 0);
-
for (x = 0; x != MLX5E_TLS_STATS_NUM; x++)
counter_u64_free(ptls->stats.arg[x]);
}
@@ -334,28 +410,16 @@ mlx5e_tls_snd_tag_alloc(if_t ifp,
if (priv->gone != 0 || priv->tls.init == 0)
return (EOPNOTSUPP);
- /* allocate new tag from zone, if any */
ptag = uma_zalloc(priv->tls.zone, M_WAITOK);
+ if (ptag == NULL)
+ return (ENOMEM);
/* sanity check default values */
MPASS(ptag->dek_index == 0);
MPASS(ptag->dek_index_ok == 0);
- /* setup TLS tag */
- ptag->tls = &priv->tls;
-
/* check if there is no TIS context */
- if (ptag->tisn == 0) {
- uint32_t value;
-
- value = atomic_fetchadd_32(&priv->tls.num_resources, 1U);
-
- /* check resource limits */
- if (value >= priv->tls.max_resources) {
- error = ENOMEM;
- goto failure;
- }
- }
+ KASSERT(ptag->tisn != 0, ("ptag %p w/0 tisn", ptag));
en = ¶ms->tls.tls->params;
@@ -448,17 +512,9 @@ mlx5e_tls_snd_tag_alloc(if_t ifp,
/* reset state */
ptag->state = MLX5E_TLS_ST_INIT;
- /*
- * Try to immediately init the tag. We may fail if the NIC's
- * resources are tied up with send tags that are in the work
- * queue, waiting to be freed. So if we fail, put ourselves
- * on the queue so as to try again after resouces have been freed.
- */
error = mlx5e_tls_st_init(priv, ptag);
- if (error != 0) {
- queue_work(priv->tls.wq, &ptag->work);
- flush_work(&ptag->work);
- }
+ if (error != 0)
+ goto failure;
return (0);