svn commit: r364497 - in head/sys/dev/cxgbe: . common
Navdeep Parhar
np at FreeBSD.org
Sun Aug 23 04:16:21 UTC 2020
Author: np
Date: Sun Aug 23 04:16:20 2020
New Revision: 364497
URL: https://svnweb.freebsd.org/changeset/base/364497
Log:
cxgbe(4): Use large clusters for TOE rx queues when TOE+TLS is enabled.
Rx is more efficient within the chip when the receive buffer size
matches the TLS PDU size.
MFC after: 3 days
Sponsored by: Chelsio Communications
Differential Revision: https://reviews.freebsd.org/D26127
Modified:
head/sys/dev/cxgbe/common/common.h
head/sys/dev/cxgbe/common/t4_hw.c
head/sys/dev/cxgbe/t4_main.c
head/sys/dev/cxgbe/t4_sge.c
Modified: head/sys/dev/cxgbe/common/common.h
==============================================================================
--- head/sys/dev/cxgbe/common/common.h Sat Aug 22 22:56:50 2020 (r364496)
+++ head/sys/dev/cxgbe/common/common.h Sun Aug 23 04:16:20 2020 (r364497)
@@ -246,6 +246,8 @@ struct tp_params {
uint32_t vlan_pri_map;
uint32_t ingress_config;
+ uint32_t max_rx_pdu;
+ uint32_t max_tx_pdu;
uint64_t hash_filter_mask;
__be16 err_vec_mask;
Modified: head/sys/dev/cxgbe/common/t4_hw.c
==============================================================================
--- head/sys/dev/cxgbe/common/t4_hw.c Sat Aug 22 22:56:50 2020 (r364496)
+++ head/sys/dev/cxgbe/common/t4_hw.c Sun Aug 23 04:16:20 2020 (r364497)
@@ -9614,7 +9614,7 @@ static void read_filter_mode_and_ingress_config(struct
int t4_init_tp_params(struct adapter *adap, bool sleep_ok)
{
int chan;
- u32 v;
+ u32 tx_len, rx_len, r, v;
struct tp_params *tpp = &adap->params.tp;
v = t4_read_reg(adap, A_TP_TIMER_RESOLUTION);
@@ -9640,6 +9640,21 @@ int t4_init_tp_params(struct adapter *adap, bool sleep
htobe16(V_T6_COMPR_RXERR_VEC(M_T6_COMPR_RXERR_VEC));
}
}
+
+ rx_len = t4_read_reg(adap, A_TP_PMM_RX_PAGE_SIZE);
+ tx_len = t4_read_reg(adap, A_TP_PMM_TX_PAGE_SIZE);
+
+ r = t4_read_reg(adap, A_TP_PARA_REG2);
+ rx_len = min(rx_len, G_MAXRXDATA(r));
+ tx_len = min(tx_len, G_MAXRXDATA(r));
+
+ r = t4_read_reg(adap, A_TP_PARA_REG7);
+ v = min(G_PMMAXXFERLEN0(r), G_PMMAXXFERLEN1(r));
+ rx_len = min(rx_len, v);
+ tx_len = min(tx_len, v);
+
+ tpp->max_tx_pdu = tx_len;
+ tpp->max_rx_pdu = rx_len;
return 0;
}
Modified: head/sys/dev/cxgbe/t4_main.c
==============================================================================
--- head/sys/dev/cxgbe/t4_main.c Sat Aug 22 22:56:50 2020 (r364496)
+++ head/sys/dev/cxgbe/t4_main.c Sun Aug 23 04:16:20 2020 (r364497)
@@ -736,6 +736,7 @@ static int sysctl_ulprx_la(SYSCTL_HANDLER_ARGS);
static int sysctl_wcwr_stats(SYSCTL_HANDLER_ARGS);
static int sysctl_cpus(SYSCTL_HANDLER_ARGS);
#ifdef TCP_OFFLOAD
+static int sysctl_tls(SYSCTL_HANDLER_ARGS);
static int sysctl_tls_rx_ports(SYSCTL_HANDLER_ARGS);
static int sysctl_tp_tick(SYSCTL_HANDLER_ARGS);
static int sysctl_tp_dack_timer(SYSCTL_HANDLER_ARGS);
@@ -6607,8 +6608,9 @@ t4_sysctls(struct adapter *sc)
CTLFLAG_RW, &sc->tt.rx_coalesce, 0, "receive coalescing");
sc->tt.tls = 0;
- SYSCTL_ADD_INT(ctx, children, OID_AUTO, "tls", CTLFLAG_RW,
- &sc->tt.tls, 0, "Inline TLS allowed");
+ SYSCTL_ADD_PROC(ctx, children, OID_AUTO, "tls", CTLTYPE_INT |
+ CTLFLAG_RW | CTLFLAG_NEEDGIANT, sc, 0, sysctl_tls, "I",
+ "Inline TLS allowed");
SYSCTL_ADD_PROC(ctx, children, OID_AUTO, "tls_rx_ports",
CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_NEEDGIANT, sc, 0,
@@ -9699,6 +9701,37 @@ sysctl_cpus(SYSCTL_HANDLER_ARGS)
}
#ifdef TCP_OFFLOAD
+static int
+sysctl_tls(SYSCTL_HANDLER_ARGS)
+{
+ struct adapter *sc = arg1;
+ int i, j, v, rc;
+ struct vi_info *vi;
+
+ v = sc->tt.tls;
+ rc = sysctl_handle_int(oidp, &v, 0, req);
+ if (rc != 0 || req->newptr == NULL)
+ return (rc);
+
+ if (v != 0 && !(sc->cryptocaps & FW_CAPS_CONFIG_TLSKEYS))
+ return (ENOTSUP);
+
+ rc = begin_synchronized_op(sc, NULL, SLEEP_OK | INTR_OK, "t4stls");
+ if (rc)
+ return (rc);
+ sc->tt.tls = !!v;
+ for_each_port(sc, i) {
+ for_each_vi(sc->port[i], j, vi) {
+ if (vi->flags & VI_INIT_DONE)
+ t4_update_fl_bufsize(vi->ifp);
+ }
+ }
+ end_synchronized_op(sc, 0);
+
+ return (0);
+
+}
+
static int
sysctl_tls_rx_ports(SYSCTL_HANDLER_ARGS)
{
Modified: head/sys/dev/cxgbe/t4_sge.c
==============================================================================
--- head/sys/dev/cxgbe/t4_sge.c Sat Aug 22 22:56:50 2020 (r364496)
+++ head/sys/dev/cxgbe/t4_sge.c Sun Aug 23 04:16:20 2020 (r364497)
@@ -1032,14 +1032,19 @@ t4_teardown_adapter_queues(struct adapter *sc)
return (0);
}
-/* Maximum payload that can be delivered with a single iq descriptor */
+/* Maximum payload that could arrive with a single iq descriptor. */
static inline int
-mtu_to_max_payload(struct adapter *sc, int mtu)
+max_rx_payload(struct adapter *sc, struct ifnet *ifp, const bool ofld)
{
+ int maxp;
/* large enough even when hw VLAN extraction is disabled */
- return (sc->params.sge.fl_pktshift + ETHER_HDR_LEN +
- ETHER_VLAN_ENCAP_LEN + mtu);
+ maxp = sc->params.sge.fl_pktshift + ETHER_HDR_LEN +
+ ETHER_VLAN_ENCAP_LEN + ifp->if_mtu;
+ if (ofld && sc->tt.tls && sc->cryptocaps & FW_CAPS_CONFIG_TLSKEYS &&
+ maxp < sc->params.tp.max_rx_pdu)
+ maxp = sc->params.tp.max_rx_pdu;
+ return (maxp);
}
int
@@ -1065,7 +1070,7 @@ t4_setup_vi_queues(struct vi_info *vi)
struct ifnet *ifp = vi->ifp;
struct sysctl_oid *oid = device_get_sysctl_tree(vi->dev);
struct sysctl_oid_list *children = SYSCTL_CHILDREN(oid);
- int maxp, mtu = ifp->if_mtu;
+ int maxp;
/* Interrupt vector to start from (when using multiple vectors) */
intr_idx = vi->first_intr;
@@ -1109,7 +1114,7 @@ t4_setup_vi_queues(struct vi_info *vi)
* Allocate rx queues first because a default iqid is required when
* creating a tx queue.
*/
- maxp = mtu_to_max_payload(sc, mtu);
+ maxp = max_rx_payload(sc, ifp, false);
oid = SYSCTL_ADD_NODE(&vi->ctx, children, OID_AUTO, "rxq",
CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, "rx queues");
for_each_rxq(vi, i, rxq) {
@@ -1131,6 +1136,7 @@ t4_setup_vi_queues(struct vi_info *vi)
intr_idx = saved_idx + max(vi->nrxq, vi->nnmrxq);
#endif
#ifdef TCP_OFFLOAD
+ maxp = max_rx_payload(sc, ifp, true);
oid = SYSCTL_ADD_NODE(&vi->ctx, children, OID_AUTO, "ofld_rxq",
CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, "rx queues for offloaded TCP connections");
for_each_ofld_rxq(vi, i, ofld_rxq) {
@@ -2144,9 +2150,9 @@ t4_update_fl_bufsize(struct ifnet *ifp)
struct sge_ofld_rxq *ofld_rxq;
#endif
struct sge_fl *fl;
- int i, maxp, mtu = ifp->if_mtu;
+ int i, maxp;
- maxp = mtu_to_max_payload(sc, mtu);
+ maxp = max_rx_payload(sc, ifp, false);
for_each_rxq(vi, i, rxq) {
fl = &rxq->fl;
@@ -2156,6 +2162,7 @@ t4_update_fl_bufsize(struct ifnet *ifp)
FL_UNLOCK(fl);
}
#ifdef TCP_OFFLOAD
+ maxp = max_rx_payload(sc, ifp, true);
for_each_ofld_rxq(vi, i, ofld_rxq) {
fl = &ofld_rxq->fl;
More information about the svn-src-head
mailing list