svn commit: r363844 - in stable/12/sys: dev/vmware/vmxnet3 net
Joe Clarke
jclarke at marcuscom.com
Tue Aug 4 15:54:16 UTC 2020
Thanks, Mark.
Joe
> On Aug 4, 2020, at 11:48, Mark Johnston <markj at freebsd.org> wrote:
>
> Author: markj
> Date: Tue Aug 4 15:48:12 2020
> New Revision: 363844
> URL: https://svnweb.freebsd.org/changeset/base/363844
>
> Log:
> MFC r358996 (by pkelsey):
> Fix iflib zero-length fragment handling
>
> MFC r358997 (by pkelsey):
> Remove freelist contiguous-indexes assertion from rxd_frag_to_sd()
>
> MFC r358998 (by pkelsey):
> Allow iflib drivers to specify the buffer size used for each receive queue
>
> MFC r358999 (by pkelsey):
> Remove refill budget from iflib
>
> MFC r359000 (by pkelsey):
> Fix if_vmx receive checksum offload bug and harden against the device
> skipping receive descriptors
>
> MFC r359001 (by pkelsey):
> Adjust if_vmx default receive parameters for better out-of-box performance
>
> MFC r359002 (by pkelsey):
> Remove extraneous code from iflib
>
> MFC r359029 (by pkelsey):
> Restore power-of-2 queue count constraint from r290948
>
> Tested by: marcus, mav
>
> Modified:
> stable/12/sys/dev/vmware/vmxnet3/if_vmx.c
> stable/12/sys/dev/vmware/vmxnet3/if_vmxvar.h
> stable/12/sys/net/iflib.c
> stable/12/sys/net/iflib.h
> Directory Properties:
> stable/12/ (props changed)
>
> Modified: stable/12/sys/dev/vmware/vmxnet3/if_vmx.c
> ==============================================================================
> --- stable/12/sys/dev/vmware/vmxnet3/if_vmx.c Tue Aug 4 15:27:32 2020 (r363843)
> +++ stable/12/sys/dev/vmware/vmxnet3/if_vmx.c Tue Aug 4 15:48:12 2020 (r363844)
> @@ -321,6 +321,13 @@ vmxnet3_register(device_t dev)
> }
>
> static int
> +trunc_powerof2(int val)
> +{
> +
> + return (1U << (fls(val) - 1));
> +}
> +
> +static int
> vmxnet3_attach_pre(if_ctx_t ctx)
> {
> device_t dev;
> @@ -349,12 +356,16 @@ vmxnet3_attach_pre(if_ctx_t ctx)
> /* If 0, the iflib tunable was not set, so set to the default */
> if (scctx->isc_nrxqsets == 0)
> scctx->isc_nrxqsets = VMXNET3_DEF_RX_QUEUES;
> + scctx->isc_nrxqsets = trunc_powerof2(scctx->isc_nrxqsets);
> scctx->isc_nrxqsets_max = min(VMXNET3_MAX_RX_QUEUES, mp_ncpus);
> + scctx->isc_nrxqsets_max = trunc_powerof2(scctx->isc_nrxqsets_max);
>
> /* If 0, the iflib tunable was not set, so set to the default */
> if (scctx->isc_ntxqsets == 0)
> scctx->isc_ntxqsets = VMXNET3_DEF_TX_QUEUES;
> + scctx->isc_ntxqsets = trunc_powerof2(scctx->isc_ntxqsets);
> scctx->isc_ntxqsets_max = min(VMXNET3_MAX_TX_QUEUES, mp_ncpus);
> + scctx->isc_ntxqsets_max = trunc_powerof2(scctx->isc_ntxqsets_max);
>
> /*
> * Enforce that the transmit completion queue descriptor count is
> @@ -381,6 +392,12 @@ vmxnet3_attach_pre(if_ctx_t ctx)
> scctx->isc_rxqsizes[2] =
> sizeof(struct vmxnet3_rxdesc) * scctx->isc_nrxd[2];
>
> + /*
> + * Initialize the max frame size and descriptor queue buffer
> + * sizes.
> + */
> + vmxnet3_mtu_set(ctx, if_getmtu(sc->vmx_ifp));
> +
> scctx->isc_rss_table_size = UPT1_RSS_MAX_IND_TABLE_SIZE;
>
> /* Map PCI BARs */
> @@ -1494,6 +1511,7 @@ vmxnet3_isc_rxd_pkt_get(void *vsc, if_rxd_info_t ri)
> int cqidx;
> uint16_t total_len;
> uint8_t nfrags;
> + uint8_t i;
> uint8_t flid;
>
> sc = vsc;
> @@ -1517,6 +1535,7 @@ vmxnet3_isc_rxd_pkt_get(void *vsc, if_rxd_info_t ri)
> KASSERT(rxcd->sop && rxcd->eop,
> ("%s: zero-length packet without both sop and eop set",
> __func__));
> + rxc->vxcr_zero_length++;
> if (++cqidx == rxc->vxcr_ndesc) {
> cqidx = 0;
> rxc->vxcr_gen ^= 1;
> @@ -1572,31 +1591,6 @@ vmxnet3_isc_rxd_pkt_get(void *vsc, if_rxd_info_t ri)
> }
> }
>
> - /* VLAN */
> - if (rxcd->vlan) {
> - ri->iri_flags |= M_VLANTAG;
> - ri->iri_vtag = rxcd->vtag;
> - }
> -
> - /* Checksum offload */
> - if (!rxcd->no_csum) {
> - uint32_t csum_flags = 0;
> -
> - if (rxcd->ipv4) {
> - csum_flags |= CSUM_IP_CHECKED;
> - if (rxcd->ipcsum_ok)
> - csum_flags |= CSUM_IP_VALID;
> - }
> - if (!rxcd->fragment && (rxcd->tcp || rxcd->udp)) {
> - csum_flags |= CSUM_L4_CALC;
> - if (rxcd->csum_ok) {
> - csum_flags |= CSUM_L4_VALID;
> - ri->iri_csum_data = 0xffff;
> - }
> - }
> - ri->iri_csum_flags = csum_flags;
> - }
> -
> /*
> * The queue numbering scheme used for rxcd->qid is as follows:
> * - All of the command ring 0s are numbered [0, nrxqsets - 1]
> @@ -1632,6 +1626,46 @@ vmxnet3_isc_rxd_pkt_get(void *vsc, if_rxd_info_t ri)
> ri->iri_nfrags = nfrags;
> ri->iri_len = total_len;
>
> + /*
> + * If there's an error, the last descriptor in the packet will
> + * have the error indicator set. In this case, set all
> + * fragment lengths to zero. This will cause iflib to discard
> + * the packet, but process all associated descriptors through
> + * the refill mechanism.
> + */
> + if (__predict_false(rxcd->error)) {
> + rxc->vxcr_pkt_errors++;
> + for (i = 0; i < nfrags; i++) {
> + frag = &ri->iri_frags[i];
> + frag->irf_len = 0;
> + }
> + } else {
> + /* Checksum offload information is in the last descriptor. */
> + if (!rxcd->no_csum) {
> + uint32_t csum_flags = 0;
> +
> + if (rxcd->ipv4) {
> + csum_flags |= CSUM_IP_CHECKED;
> + if (rxcd->ipcsum_ok)
> + csum_flags |= CSUM_IP_VALID;
> + }
> + if (!rxcd->fragment && (rxcd->tcp || rxcd->udp)) {
> + csum_flags |= CSUM_L4_CALC;
> + if (rxcd->csum_ok) {
> + csum_flags |= CSUM_L4_VALID;
> + ri->iri_csum_data = 0xffff;
> + }
> + }
> + ri->iri_csum_flags = csum_flags;
> + }
> +
> + /* VLAN information is in the last descriptor. */
> + if (rxcd->vlan) {
> + ri->iri_flags |= M_VLANTAG;
> + ri->iri_vtag = rxcd->vtag;
> + }
> + }
> +
> return (0);
> }
>
> @@ -1645,14 +1679,13 @@ vmxnet3_isc_rxd_refill(void *vsc, if_rxd_update_t iru)
> uint64_t *paddrs;
> int count;
> int len;
> - int pidx;
> + int idx;
> int i;
> uint8_t flid;
> uint8_t btype;
>
> count = iru->iru_count;
> len = iru->iru_buf_size;
> - pidx = iru->iru_pidx;
> flid = iru->iru_flidx;
> paddrs = iru->iru_paddrs;
>
> @@ -1666,17 +1699,32 @@ vmxnet3_isc_rxd_refill(void *vsc, if_rxd_update_t iru)
> * command ring 1 is filled with BTYPE_BODY descriptors.
> */
> btype = (flid == 0) ? VMXNET3_BTYPE_HEAD : VMXNET3_BTYPE_BODY;
> - for (i = 0; i < count; i++) {
> - rxd[pidx].addr = paddrs[i];
> - rxd[pidx].len = len;
> - rxd[pidx].btype = btype;
> - rxd[pidx].gen = rxr->vxrxr_gen;
> + /*
> + * The refill entries from iflib will advance monotonically,
> + * but the refilled descriptors may not be contiguous due to
> + * earlier skipping of descriptors by the device. The refill
> + * entries from iflib need an entire state update, while the
> + * descriptors previously skipped by the device only need to
> + * have their generation numbers updated.
> + */
> + idx = rxr->vxrxr_refill_start;
> + i = 0;
> + do {
> + if (idx == iru->iru_idxs[i]) {
> + rxd[idx].addr = paddrs[i];
> + rxd[idx].len = len;
> + rxd[idx].btype = btype;
> + i++;
> + } else
> + rxr->vxrxr_desc_skips++;
> + rxd[idx].gen = rxr->vxrxr_gen;
>
> - if (++pidx == rxr->vxrxr_ndesc) {
> - pidx = 0;
> + if (++idx == rxr->vxrxr_ndesc) {
> + idx = 0;
> rxr->vxrxr_gen ^= 1;
> }
> - }
> + } while (i != count);
> + rxr->vxrxr_refill_start = idx;
> }
>
> static void
> @@ -1825,6 +1873,8 @@ vmxnet3_rxinit(struct vmxnet3_softc *sc, struct vmxnet
> for (i = 0; i < sc->vmx_sctx->isc_nrxqs - 1; i++) {
> rxr = &rxq->vxrxq_cmd_ring[i];
> rxr->vxrxr_gen = VMXNET3_INIT_GEN;
> + rxr->vxrxr_desc_skips = 0;
> + rxr->vxrxr_refill_start = 0;
> /*
> * iflib has zeroed out the descriptor array during the
> * prior attach or stop
> @@ -1834,6 +1884,8 @@ vmxnet3_rxinit(struct vmxnet3_softc *sc, struct vmxnet
> for (/**/; i < VMXNET3_RXRINGS_PERQ; i++) {
> rxr = &rxq->vxrxq_cmd_ring[i];
> rxr->vxrxr_gen = 0;
> + rxr->vxrxr_desc_skips = 0;
> + rxr->vxrxr_refill_start = 0;
> bzero(rxr->vxrxr_rxd,
> rxr->vxrxr_ndesc * sizeof(struct vmxnet3_rxdesc));
> }
> @@ -1841,6 +1893,8 @@ vmxnet3_rxinit(struct vmxnet3_softc *sc, struct vmxnet
> rxc = &rxq->vxrxq_comp_ring;
> rxc->vxcr_next = 0;
> rxc->vxcr_gen = VMXNET3_INIT_GEN;
> + rxc->vxcr_zero_length = 0;
> + rxc->vxcr_pkt_errors = 0;
> /*
> * iflib has zeroed out the descriptor array during the prior attach
> * or stop
> @@ -1906,14 +1960,9 @@ static void
> vmxnet3_init(if_ctx_t ctx)
> {
> struct vmxnet3_softc *sc;
> - if_softc_ctx_t scctx;
>
> sc = iflib_get_softc(ctx);
> - scctx = sc->vmx_scctx;
>
> - scctx->isc_max_frame_size = if_getmtu(iflib_get_ifp(ctx)) +
> - ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN + ETHER_CRC_LEN;
> -
> /* Use the current MAC address. */
> bcopy(IF_LLADDR(sc->vmx_ifp), sc->vmx_lladdr, ETHER_ADDR_LEN);
> vmxnet3_set_lladdr(sc);
> @@ -1938,11 +1987,37 @@ vmxnet3_multi_set(if_ctx_t ctx)
> static int
> vmxnet3_mtu_set(if_ctx_t ctx, uint32_t mtu)
> {
> + struct vmxnet3_softc *sc;
> + if_softc_ctx_t scctx;
>
> + sc = iflib_get_softc(ctx);
> + scctx = sc->vmx_scctx;
> +
> if (mtu > VMXNET3_TX_MAXSIZE - (ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN +
> ETHER_CRC_LEN))
> return (EINVAL);
>
> + /*
> + * Update the max frame size so that the rx mbuf size is
> + * chosen based on the new mtu during the interface init that
> + * will occur after this routine returns.
> + */
> + scctx->isc_max_frame_size = mtu +
> + ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN + ETHER_CRC_LEN;
> + /* RX completion queue - n/a */
> + scctx->isc_rxd_buf_size[0] = 0;
> + /*
> + * For header-type descriptors (used for first segment of
> + * packet), let iflib determine the buffer size based on the
> + * max frame size.
> + */
> + scctx->isc_rxd_buf_size[1] = 0;
> + /*
> + * For body-type descriptors (used for jumbo frames and LRO),
> + * always use page-sized buffers.
> + */
> + scctx->isc_rxd_buf_size[2] = MJUMPAGESIZE;
> +
> return (0);
> }
>
> @@ -2288,14 +2363,22 @@ vmxnet3_setup_debug_sysctl(struct vmxnet3_softc *sc,
> &rxq->vxrxq_cmd_ring[0].vxrxr_ndesc, 0, "");
> SYSCTL_ADD_INT(ctx, list, OID_AUTO, "cmd0_gen", CTLFLAG_RD,
> &rxq->vxrxq_cmd_ring[0].vxrxr_gen, 0, "");
> + SYSCTL_ADD_U64(ctx, list, OID_AUTO, "cmd0_desc_skips", CTLFLAG_RD,
> + &rxq->vxrxq_cmd_ring[0].vxrxr_desc_skips, 0, "");
> SYSCTL_ADD_UINT(ctx, list, OID_AUTO, "cmd1_ndesc", CTLFLAG_RD,
> &rxq->vxrxq_cmd_ring[1].vxrxr_ndesc, 0, "");
> SYSCTL_ADD_INT(ctx, list, OID_AUTO, "cmd1_gen", CTLFLAG_RD,
> &rxq->vxrxq_cmd_ring[1].vxrxr_gen, 0, "");
> + SYSCTL_ADD_U64(ctx, list, OID_AUTO, "cmd1_desc_skips", CTLFLAG_RD,
> + &rxq->vxrxq_cmd_ring[1].vxrxr_desc_skips, 0, "");
> SYSCTL_ADD_UINT(ctx, list, OID_AUTO, "comp_ndesc", CTLFLAG_RD,
> &rxq->vxrxq_comp_ring.vxcr_ndesc, 0,"");
> SYSCTL_ADD_INT(ctx, list, OID_AUTO, "comp_gen", CTLFLAG_RD,
> &rxq->vxrxq_comp_ring.vxcr_gen, 0, "");
> + SYSCTL_ADD_U64(ctx, list, OID_AUTO, "comp_zero_length", CTLFLAG_RD,
> + &rxq->vxrxq_comp_ring.vxcr_zero_length, 0, "");
> + SYSCTL_ADD_U64(ctx, list, OID_AUTO, "comp_pkt_errors", CTLFLAG_RD,
> + &rxq->vxrxq_comp_ring.vxcr_pkt_errors, 0, "");
> }
> }
>
>
> Modified: stable/12/sys/dev/vmware/vmxnet3/if_vmxvar.h
> ==============================================================================
> --- stable/12/sys/dev/vmware/vmxnet3/if_vmxvar.h Tue Aug 4 15:27:32 2020 (r363843)
> +++ stable/12/sys/dev/vmware/vmxnet3/if_vmxvar.h Tue Aug 4 15:48:12 2020 (r363844)
> @@ -41,7 +41,7 @@ struct vmxnet3_softc;
> #define VMXNET3_MAX_TX_NDESC 4096
> #define VMXNET3_MIN_TX_NDESC 32
> #define VMXNET3_MASK_TX_NDESC 0x1F
> -#define VMXNET3_DEF_RX_NDESC 256
> +#define VMXNET3_DEF_RX_NDESC 512
> #define VMXNET3_MAX_RX_NDESC 2048
> #define VMXNET3_MIN_RX_NDESC 32
> #define VMXNET3_MASK_RX_NDESC 0x1F
> @@ -63,6 +63,8 @@ struct vmxnet3_rxring {
> u_int vxrxr_ndesc;
> int vxrxr_gen;
> bus_addr_t vxrxr_paddr;
> + uint64_t vxrxr_desc_skips;
> + uint16_t vxrxr_refill_start;
> };
>
> struct vmxnet3_comp_ring {
> @@ -78,6 +80,8 @@ struct vmxnet3_comp_ring {
> u_int vxcr_ndesc;
> int vxcr_gen;
> bus_addr_t vxcr_paddr;
> + uint64_t vxcr_zero_length;
> + uint64_t vxcr_pkt_errors;
> };
>
> struct vmxnet3_txqueue {
>
> Modified: stable/12/sys/net/iflib.c
> ==============================================================================
> --- stable/12/sys/net/iflib.c Tue Aug 4 15:27:32 2020 (r363843)
> +++ stable/12/sys/net/iflib.c Tue Aug 4 15:48:12 2020 (r363844)
> @@ -453,7 +453,6 @@ typedef struct if_rxsd {
> caddr_t *ifsd_cl;
> struct mbuf **ifsd_m;
> iflib_fl_t ifsd_fl;
> - qidx_t ifsd_cidx;
> } *if_rxsd_t;
>
> /* multiple of word size */
> @@ -713,6 +712,7 @@ static int iflib_altq_if_transmit(if_t ifp, struct mbu
> static int iflib_register(if_ctx_t);
> static void iflib_deregister(if_ctx_t);
> static void iflib_unregister_vlan_handlers(if_ctx_t ctx);
> +static uint16_t iflib_get_mbuf_size_for(unsigned int size);
> static void iflib_init_locked(if_ctx_t ctx);
> static void iflib_add_device_sysctl_pre(if_ctx_t ctx);
> static void iflib_add_device_sysctl_post(if_ctx_t ctx);
> @@ -2130,7 +2130,7 @@ _iflib_fl_refill(if_ctx_t ctx, iflib_fl_t fl, int coun
> }
>
> static __inline uint8_t
> -__iflib_fl_refill_lt(if_ctx_t ctx, iflib_fl_t fl, int max)
> +__iflib_fl_refill_all(if_ctx_t ctx, iflib_fl_t fl)
> {
> /* we avoid allowing pidx to catch up with cidx as it confuses ixl */
> int32_t reclaimable = fl->ifl_size - fl->ifl_credits - 1;
> @@ -2142,7 +2142,7 @@ __iflib_fl_refill_lt(if_ctx_t ctx, iflib_fl_t fl, int
> MPASS(reclaimable == delta);
>
> if (reclaimable > 0)
> - return (_iflib_fl_refill(ctx, fl, min(max, reclaimable)));
> + return (_iflib_fl_refill(ctx, fl, reclaimable));
> return (0);
> }
>
> @@ -2213,6 +2213,8 @@ iflib_fl_setup(iflib_fl_t fl)
> {
> iflib_rxq_t rxq = fl->ifl_rxq;
> if_ctx_t ctx = rxq->ifr_ctx;
> + if_softc_ctx_t scctx = &ctx->ifc_softc_ctx;
> + int qidx;
>
> bit_nclear(fl->ifl_rx_bitmap, 0, fl->ifl_size - 1);
> /*
> @@ -2221,7 +2223,16 @@ iflib_fl_setup(iflib_fl_t fl)
> iflib_fl_bufs_free(fl);
> /* Now replenish the mbufs */
> MPASS(fl->ifl_credits == 0);
> - fl->ifl_buf_size = ctx->ifc_rx_mbuf_sz;
> + qidx = rxq->ifr_fl_offset + fl->ifl_id;
> + if (scctx->isc_rxd_buf_size[qidx] != 0)
> + fl->ifl_buf_size = scctx->isc_rxd_buf_size[qidx];
> + else
> + fl->ifl_buf_size = ctx->ifc_rx_mbuf_sz;
> + /*
> + * ifl_buf_size may be a driver-supplied value, so pull it up
> + * to the selected mbuf size.
> + */
> + fl->ifl_buf_size = iflib_get_mbuf_size_for(fl->ifl_buf_size);
> if (fl->ifl_buf_size > ctx->ifc_max_fl_buf_size)
> ctx->ifc_max_fl_buf_size = fl->ifl_buf_size;
> fl->ifl_cltype = m_gettype(fl->ifl_buf_size);
> @@ -2353,6 +2364,16 @@ iflib_timer(void *arg)
> STATE_UNLOCK(ctx);
> }
>
> +static uint16_t
> +iflib_get_mbuf_size_for(unsigned int size)
> +{
> +
> + if (size <= MCLBYTES)
> + return (MCLBYTES);
> + else
> + return (MJUMPAGESIZE);
> +}
> +
> static void
> iflib_calc_rx_mbuf_sz(if_ctx_t ctx)
> {
> @@ -2362,10 +2383,8 @@ iflib_calc_rx_mbuf_sz(if_ctx_t ctx)
> * XXX don't set the max_frame_size to larger
> * than the hardware can handle
> */
> - if (sctx->isc_max_frame_size <= MCLBYTES)
> - ctx->ifc_rx_mbuf_sz = MCLBYTES;
> - else
> - ctx->ifc_rx_mbuf_sz = MJUMPAGESIZE;
> + ctx->ifc_rx_mbuf_sz =
> + iflib_get_mbuf_size_for(sctx->isc_max_frame_size);
> }
>
> uint32_t
> @@ -2578,7 +2597,6 @@ rxd_frag_to_sd(iflib_rxq_t rxq, if_rxd_frag_t irf, int
> cidx = irf->irf_idx;
> fl = &rxq->ifr_fl[flid];
> sd->ifsd_fl = fl;
> - sd->ifsd_cidx = cidx;
> sd->ifsd_m = &fl->ifl_sds.ifsd_m[cidx];
> sd->ifsd_cl = &fl->ifl_sds.ifsd_cl[cidx];
> fl->ifl_credits--;
> @@ -2590,12 +2608,10 @@ rxd_frag_to_sd(iflib_rxq_t rxq, if_rxd_frag_t irf, int
> next = (cidx + CACHE_PTR_INCREMENT) & (fl->ifl_size-1);
> prefetch(&fl->ifl_sds.ifsd_map[next]);
> map = fl->ifl_sds.ifsd_map[cidx];
> - next = (cidx + CACHE_LINE_SIZE) & (fl->ifl_size-1);
>
> - /* not valid assert if bxe really does SGE from non-contiguous elements */
> - MPASS(fl->ifl_cidx == cidx);
> bus_dmamap_sync(fl->ifl_buf_tag, map, BUS_DMASYNC_POSTREAD);
> - if (unload)
> +
> + if (unload && irf->irf_len != 0)
> bus_dmamap_unload(fl->ifl_buf_tag, map);
> fl->ifl_cidx = (fl->ifl_cidx + 1) & (fl->ifl_size-1);
> if (__predict_false(fl->ifl_cidx == 0))
> @@ -2667,6 +2683,7 @@ iflib_rxd_pkt_get(iflib_rxq_t rxq, if_rxd_info_t ri)
>
> /* should I merge this back in now that the two paths are basically duplicated? */
> if (ri->iri_nfrags == 1 &&
> + ri->iri_frags[0].irf_len != 0 &&
> ri->iri_frags[0].irf_len <= MIN(IFLIB_RX_COPY_THRESH, MHLEN)) {
> rxd_frag_to_sd(rxq, &ri->iri_frags[0], FALSE, &sd);
> m = *sd.ifsd_m;
> @@ -2680,6 +2697,8 @@ iflib_rxd_pkt_get(iflib_rxq_t rxq, if_rxd_info_t ri)
> m->m_len = ri->iri_frags[0].irf_len;
> } else {
> m = assemble_segments(rxq, ri, &sd);
> + if (m == NULL)
> + return (NULL);
> }
> m->m_pkthdr.len = ri->iri_len;
> m->m_pkthdr.rcvif = ri->iri_ifp;
> @@ -2780,7 +2799,7 @@ iflib_rxeof(iflib_rxq_t rxq, qidx_t budget)
> cidxp = &rxq->ifr_fl[0].ifl_cidx;
> if ((avail = iflib_rxd_avail(ctx, rxq, *cidxp, budget)) == 0) {
> for (i = 0, fl = &rxq->ifr_fl[0]; i < sctx->isc_nfl; i++, fl++)
> - retval |= __iflib_fl_refill_lt(ctx, fl, budget + 8);
> + retval |= __iflib_fl_refill_all(ctx, fl);
> DBG_COUNTER_INC(rx_unavail);
> return (retval);
> }
> @@ -2836,7 +2855,7 @@ iflib_rxeof(iflib_rxq_t rxq, qidx_t budget)
> }
> /* make sure that we can refill faster than drain */
> for (i = 0, fl = &rxq->ifr_fl[0]; i < sctx->isc_nfl; i++, fl++)
> - retval |= __iflib_fl_refill_lt(ctx, fl, budget + 8);
> + retval |= __iflib_fl_refill_all(ctx, fl);
>
> lro_enabled = (if_getcapenable(ifp) & IFCAP_LRO);
> if (lro_enabled)
> @@ -6702,6 +6721,9 @@ iflib_add_device_sysctl_post(if_ctx_t ctx)
> SYSCTL_ADD_U16(ctx_list, fl_list, OID_AUTO, "credits",
> CTLFLAG_RD,
> &fl->ifl_credits, 1, "credits available");
> + SYSCTL_ADD_U16(ctx_list, fl_list, OID_AUTO, "buf_size",
> + CTLFLAG_RD,
> + &fl->ifl_buf_size, 1, "buffer size");
> #if MEMORY_LOGGING
> SYSCTL_ADD_QUAD(ctx_list, fl_list, OID_AUTO, "fl_m_enqueued",
> CTLFLAG_RD,
>
> Modified: stable/12/sys/net/iflib.h
> ==============================================================================
> --- stable/12/sys/net/iflib.h Tue Aug 4 15:27:32 2020 (r363843)
> +++ stable/12/sys/net/iflib.h Tue Aug 4 15:48:12 2020 (r363844)
> @@ -221,6 +221,9 @@ typedef struct if_softc_ctx {
> uint32_t isc_tx_qdepth;
>
> iflib_intr_mode_t isc_intr;
> + uint16_t isc_rxd_buf_size[8]; /* set at init time by driver, 0
> + means use iflib-calculated size
> + based on isc_max_frame_size */
> uint16_t isc_max_frame_size; /* set at init time by driver */
> uint16_t isc_min_frame_size; /* set at init time by driver, only used if
> IFLIB_NEED_ETHER_PAD is set. */
> _______________________________________________
> svn-src-all at freebsd.org mailing list
> https://lists.freebsd.org/mailman/listinfo/svn-src-all
> To unsubscribe, send any mail to "svn-src-all-unsubscribe at freebsd.org"
---
PGP Key : http://www.marcuscom.com/pgp.asc
More information about the svn-src-all
mailing list