svn commit: r307574 - in head/sys: dev/netmap net
Luigi Rizzo
luigi at FreeBSD.org
Tue Oct 18 16:18:27 UTC 2016
Author: luigi
Date: Tue Oct 18 16:18:25 2016
New Revision: 307574
URL: https://svnweb.freebsd.org/changeset/base/307574
Log:
remove stale and unused code from various files
fix build on 32 bit platforms
simplify logic in netmap_virt.h
The commands (in net/netmap.h) to configure communication with the
hypervisor may be revised soon.
At the moment they are unused so this will not be a change of API.
Modified:
head/sys/dev/netmap/if_lem_netmap.h
head/sys/dev/netmap/if_ptnet.c
head/sys/dev/netmap/netmap.c
head/sys/dev/netmap/netmap_freebsd.c
head/sys/dev/netmap/netmap_generic.c
head/sys/dev/netmap/netmap_kern.h
head/sys/dev/netmap/netmap_pt.c
head/sys/net/netmap.h
head/sys/net/netmap_virt.h
Modified: head/sys/dev/netmap/if_lem_netmap.h
==============================================================================
--- head/sys/dev/netmap/if_lem_netmap.h Tue Oct 18 15:50:20 2016 (r307573)
+++ head/sys/dev/netmap/if_lem_netmap.h Tue Oct 18 16:18:25 2016 (r307574)
@@ -35,12 +35,8 @@
#include <net/netmap.h>
#include <sys/selinfo.h>
-#include <vm/vm.h>
-#include <vm/pmap.h> /* vtophys ? */
#include <dev/netmap/netmap_kern.h>
-extern int netmap_adaptive_io;
-
/*
* Register/unregister. We are already under netmap lock.
*/
Modified: head/sys/dev/netmap/if_ptnet.c
==============================================================================
--- head/sys/dev/netmap/if_ptnet.c Tue Oct 18 15:50:20 2016 (r307573)
+++ head/sys/dev/netmap/if_ptnet.c Tue Oct 18 16:18:25 2016 (r307574)
@@ -341,7 +341,11 @@ ptnet_attach(device_t dev)
}
{
- vm_paddr_t paddr = vtophys(sc->csb);
+ /*
+ * We use uint64_t rather than vm_paddr_t since we
+ * need 64 bit addresses even on 32 bit platforms.
+ */
+ uint64_t paddr = vtophys(sc->csb);
bus_write_4(sc->iomem, PTNET_IO_CSBBAH,
(paddr >> 32) & 0xffffffff);
@@ -1139,9 +1143,11 @@ ptnet_sync_from_csb(struct ptnet_softc *
static void
ptnet_update_vnet_hdr(struct ptnet_softc *sc)
{
- sc->vnet_hdr_len = ptnet_vnet_hdr ? PTNET_HDR_SIZE : 0;
+ unsigned int wanted_hdr_len = ptnet_vnet_hdr ? PTNET_HDR_SIZE : 0;
+
+ bus_write_4(sc->iomem, PTNET_IO_VNET_HDR_LEN, wanted_hdr_len);
+ sc->vnet_hdr_len = bus_read_4(sc->iomem, PTNET_IO_VNET_HDR_LEN);
sc->ptna->hwup.up.virt_hdr_len = sc->vnet_hdr_len;
- bus_write_4(sc->iomem, PTNET_IO_VNET_HDR_LEN, sc->vnet_hdr_len);
}
static int
Modified: head/sys/dev/netmap/netmap.c
==============================================================================
--- head/sys/dev/netmap/netmap.c Tue Oct 18 15:50:20 2016 (r307573)
+++ head/sys/dev/netmap/netmap.c Tue Oct 18 16:18:25 2016 (r307574)
@@ -483,7 +483,6 @@ static int netmap_no_timestamp; /* don't
int netmap_mitigate = 1;
int netmap_no_pendintr = 1;
int netmap_txsync_retry = 2;
-int netmap_adaptive_io = 0;
int netmap_flags = 0; /* debug flags */
static int netmap_fwd = 0; /* force transparent mode */
@@ -540,8 +539,6 @@ SYSCTL_INT(_dev_netmap, OID_AUTO, no_pen
CTLFLAG_RW, &netmap_no_pendintr, 0, "Always look for new received packets.");
SYSCTL_INT(_dev_netmap, OID_AUTO, txsync_retry, CTLFLAG_RW,
&netmap_txsync_retry, 0 , "Number of txsync loops in bridge's flush.");
-SYSCTL_INT(_dev_netmap, OID_AUTO, adaptive_io, CTLFLAG_RW,
- &netmap_adaptive_io, 0 , "Adaptive I/O on paravirt");
SYSCTL_INT(_dev_netmap, OID_AUTO, flags, CTLFLAG_RW, &netmap_flags, 0 , "");
SYSCTL_INT(_dev_netmap, OID_AUTO, fwd, CTLFLAG_RW, &netmap_fwd, 0 , "");
@@ -1559,7 +1556,7 @@ nm_txsync_prologue(struct netmap_kring *
}
}
if (ring->tail != kring->rtail) {
- RD(5, "tail overwritten was %d need %d",
+ RD(5, "%s tail overwritten was %d need %d", kring->name,
ring->tail, kring->rtail);
ring->tail = kring->rtail;
}
Modified: head/sys/dev/netmap/netmap_freebsd.c
==============================================================================
--- head/sys/dev/netmap/netmap_freebsd.c Tue Oct 18 15:50:20 2016 (r307573)
+++ head/sys/dev/netmap/netmap_freebsd.c Tue Oct 18 16:18:25 2016 (r307574)
@@ -353,7 +353,7 @@ nm_os_generic_xmit_frame(struct nm_os_ge
bcopy(a->addr, m->m_data, len);
#else /* __FreeBSD_version >= 1100000 */
/* New FreeBSD versions. Link the external storage to
- * the netmap buffer, so that no copy is necessary. */
+ * the netmap buffer, so that no copy is necessary. */
m->m_ext.ext_buf = m->m_data = a->addr;
m->m_ext.ext_size = len;
#endif /* __FreeBSD_version >= 1100000 */
@@ -644,7 +644,8 @@ DRIVER_MODULE_ORDERED(ptn_memdev, pci, p
* of the netmap memory mapped in the guest.
*/
int
-nm_os_pt_memdev_iomap(struct ptnetmap_memdev *ptn_dev, vm_paddr_t *nm_paddr, void **nm_addr)
+nm_os_pt_memdev_iomap(struct ptnetmap_memdev *ptn_dev, vm_paddr_t *nm_paddr,
+ void **nm_addr)
{
uint32_t mem_size;
int rid;
@@ -668,8 +669,8 @@ nm_os_pt_memdev_iomap(struct ptnetmap_me
D("=== BAR %d start %lx len %lx mem_size %x ===",
PTNETMAP_MEM_PCI_BAR,
- *nm_paddr,
- rman_get_size(ptn_dev->pci_mem),
+ (unsigned long)(*nm_paddr),
+ (unsigned long)rman_get_size(ptn_dev->pci_mem),
mem_size);
return (0);
}
Modified: head/sys/dev/netmap/netmap_generic.c
==============================================================================
--- head/sys/dev/netmap/netmap_generic.c Tue Oct 18 15:50:20 2016 (r307573)
+++ head/sys/dev/netmap/netmap_generic.c Tue Oct 18 16:18:25 2016 (r307574)
@@ -95,7 +95,7 @@ __FBSDID("$FreeBSD$");
/*
* For older versions of FreeBSD:
- *
+ *
* We allocate EXT_PACKET mbuf+clusters, but need to set M_NOFREE
* so that the destructor, if invoked, will not free the packet.
* In principle we should set the destructor only on demand,
@@ -628,8 +628,6 @@ generic_mbuf_destructor(struct mbuf *m)
#endif
}
-extern int netmap_adaptive_io;
-
/* Record completed transmissions and update hwtail.
*
* The oldest tx buffer not yet completed is at nr_hwtail + 1,
@@ -690,23 +688,6 @@ generic_netmap_tx_clean(struct netmap_kr
n++;
nm_i = nm_next(nm_i, lim);
-#if 0 /* rate adaptation */
- if (netmap_adaptive_io > 1) {
- if (n >= netmap_adaptive_io)
- break;
- } else if (netmap_adaptive_io) {
- /* if hwcur - nm_i < lim/8 do an early break
- * so we prevent the sender from stalling. See CVT.
- */
- if (hwcur >= nm_i) {
- if (hwcur - nm_i < lim/2)
- break;
- } else {
- if (hwcur + lim + 1 - nm_i < lim/2)
- break;
- }
- }
-#endif
}
kring->nr_hwtail = nm_prev(nm_i, lim);
ND("tx completed [%d] -> hwtail %d", n, kring->nr_hwtail);
Modified: head/sys/dev/netmap/netmap_kern.h
==============================================================================
--- head/sys/dev/netmap/netmap_kern.h Tue Oct 18 15:50:20 2016 (r307573)
+++ head/sys/dev/netmap/netmap_kern.h Tue Oct 18 16:18:25 2016 (r307574)
@@ -1512,9 +1512,9 @@ int netmap_adapter_put(struct netmap_ada
*/
#define NETMAP_BUF_BASE(_na) ((_na)->na_lut.lut[0].vaddr)
#define NETMAP_BUF_SIZE(_na) ((_na)->na_lut.objsize)
-extern int netmap_mitigate; // XXX not really used
extern int netmap_no_pendintr;
-extern int netmap_verbose; // XXX debugging
+extern int netmap_mitigate;
+extern int netmap_verbose; /* for debugging */
enum { /* verbose flags */
NM_VERB_ON = 1, /* generic verbose */
NM_VERB_HOST = 0x2, /* verbose host stack */
@@ -1527,7 +1527,6 @@ enum {
};
extern int netmap_txsync_retry;
-extern int netmap_adaptive_io;
extern int netmap_flags;
extern int netmap_generic_mit;
extern int netmap_generic_ringsize;
Modified: head/sys/dev/netmap/netmap_pt.c
==============================================================================
--- head/sys/dev/netmap/netmap_pt.c Tue Oct 18 15:50:20 2016 (r307573)
+++ head/sys/dev/netmap/netmap_pt.c Tue Oct 18 16:18:25 2016 (r307574)
@@ -64,13 +64,10 @@
* results in random drops in the VALE txsync. */
//#define PTN_TX_BATCH_LIM(_n) ((_n >> 1))
-/* XXX: avoid nm_*sync_prologue(). XXX-vin: this should go away,
- * we should never trust the guest. */
-#define PTN_AVOID_NM_PROLOGUE
//#define BUSY_WAIT
-#define DEBUG /* Enables communication debugging. */
-#ifdef DEBUG
+#define NETMAP_PT_DEBUG /* Enables communication debugging. */
+#ifdef NETMAP_PT_DEBUG
#define DBG(x) x
#else
#define DBG(x)
@@ -196,22 +193,6 @@ ptnetmap_kring_dump(const char *title, c
kring->ring->head, kring->ring->cur, kring->ring->tail);
}
-#if 0
-static inline void
-ptnetmap_ring_reinit(struct netmap_kring *kring, uint32_t g_head, uint32_t g_cur)
-{
- struct netmap_ring *ring = kring->ring;
-
- //XXX: trust guest?
- ring->head = g_head;
- ring->cur = g_cur;
- ring->tail = NM_ACCESS_ONCE(kring->nr_hwtail);
-
- netmap_ring_reinit(kring);
- ptnetmap_kring_dump("kring reinit", kring);
-}
-#endif
-
/*
* TX functions to set/get and to handle host/guest kick.
*/
@@ -251,7 +232,7 @@ ptnetmap_tx_handler(void *data)
(struct netmap_pt_host_adapter *)kring->na->na_private;
struct ptnetmap_state *ptns = pth_na->ptns;
struct ptnet_ring __user *ptring;
- struct netmap_ring g_ring; /* guest ring pointer, copied from CSB */
+ struct netmap_ring shadow_ring; /* shadow copy of the netmap_ring */
bool more_txspace = false;
struct nm_kthread *kth;
uint32_t num_slots;
@@ -281,18 +262,18 @@ ptnetmap_tx_handler(void *data)
kth = ptns->kthreads[kring->ring_id];
num_slots = kring->nkr_num_slots;
- g_ring.head = kring->rhead;
- g_ring.cur = kring->rcur;
+ shadow_ring.head = kring->rhead;
+ shadow_ring.cur = kring->rcur;
/* Disable guest --> host notifications. */
ptring_kick_enable(ptring, 0);
/* Copy the guest kring pointers from the CSB */
- ptnetmap_host_read_kring_csb(ptring, &g_ring, num_slots);
+ ptnetmap_host_read_kring_csb(ptring, &shadow_ring, num_slots);
for (;;) {
/* If guest moves ahead too fast, let's cut the move so
* that we don't exceed our batch limit. */
- batch = g_ring.head - kring->nr_hwcur;
+ batch = shadow_ring.head - kring->nr_hwcur;
if (batch < 0)
batch += num_slots;
@@ -302,37 +283,35 @@ ptnetmap_tx_handler(void *data)
if (head_lim >= num_slots)
head_lim -= num_slots;
- ND(1, "batch: %d head: %d head_lim: %d", batch, g_ring.head,
+ ND(1, "batch: %d head: %d head_lim: %d", batch, shadow_ring.head,
head_lim);
- g_ring.head = head_lim;
+ shadow_ring.head = head_lim;
batch = PTN_TX_BATCH_LIM(num_slots);
}
#endif /* PTN_TX_BATCH_LIM */
if (nm_kr_txspace(kring) <= (num_slots >> 1)) {
- g_ring.flags |= NAF_FORCE_RECLAIM;
+ shadow_ring.flags |= NAF_FORCE_RECLAIM;
}
-#ifndef PTN_AVOID_NM_PROLOGUE
+
/* Netmap prologue */
- if (unlikely(nm_txsync_prologue(kring, &g_ring) >= num_slots)) {
- ptnetmap_ring_reinit(kring, g_ring.head, g_ring.cur);
- /* Reenable notifications. */
+ shadow_ring.tail = kring->rtail;
+ if (unlikely(nm_txsync_prologue(kring, &shadow_ring) >= num_slots)) {
+ /* Reinit ring and enable notifications. */
+ netmap_ring_reinit(kring);
ptring_kick_enable(ptring, 1);
break;
}
-#else /* PTN_AVOID_NM_PROLOGUE */
- kring->rhead = g_ring.head;
- kring->rcur = g_ring.cur;
-#endif /* !PTN_AVOID_NM_PROLOGUE */
+
if (unlikely(netmap_verbose & NM_VERB_TXSYNC)) {
ptnetmap_kring_dump("pre txsync", kring);
}
IFRATE(pre_tail = kring->rtail);
- if (unlikely(kring->nm_sync(kring, g_ring.flags))) {
+ if (unlikely(kring->nm_sync(kring, shadow_ring.flags))) {
/* Reenable notifications. */
ptring_kick_enable(ptring, 1);
- D("ERROR txsync");
+ D("ERROR txsync()");
break;
}
@@ -350,7 +329,7 @@ ptnetmap_tx_handler(void *data)
}
IFRATE(rate_batch_stats_update(&ptns->rate_ctx.new.txbs, pre_tail,
- kring->rtail, num_slots));
+ kring->rtail, num_slots));
if (unlikely(netmap_verbose & NM_VERB_TXSYNC)) {
ptnetmap_kring_dump("post txsync", kring);
@@ -367,9 +346,9 @@ ptnetmap_tx_handler(void *data)
}
#endif
/* Read CSB to see if there is more work to do. */
- ptnetmap_host_read_kring_csb(ptring, &g_ring, num_slots);
+ ptnetmap_host_read_kring_csb(ptring, &shadow_ring, num_slots);
#ifndef BUSY_WAIT
- if (g_ring.head == kring->rhead) {
+ if (shadow_ring.head == kring->rhead) {
/*
* No more packets to transmit. We enable notifications and
* go to sleep, waiting for a kick from the guest when new
@@ -379,8 +358,8 @@ ptnetmap_tx_handler(void *data)
/* Reenable notifications. */
ptring_kick_enable(ptring, 1);
/* Doublecheck. */
- ptnetmap_host_read_kring_csb(ptring, &g_ring, num_slots);
- if (g_ring.head != kring->rhead) {
+ ptnetmap_host_read_kring_csb(ptring, &shadow_ring, num_slots);
+ if (shadow_ring.head != kring->rhead) {
/* We won the race condition, there are more packets to
* transmit. Disable notifications and do another cycle */
ptring_kick_enable(ptring, 0);
@@ -433,7 +412,7 @@ ptnetmap_rx_handler(void *data)
(struct netmap_pt_host_adapter *)kring->na->na_private;
struct ptnetmap_state *ptns = pth_na->ptns;
struct ptnet_ring __user *ptring;
- struct netmap_ring g_ring; /* guest ring pointer, copied from CSB */
+ struct netmap_ring shadow_ring; /* shadow copy of the netmap_ring */
struct nm_kthread *kth;
uint32_t num_slots;
int dry_cycles = 0;
@@ -464,36 +443,32 @@ ptnetmap_rx_handler(void *data)
kth = ptns->kthreads[pth_na->up.num_tx_rings + kring->ring_id];
num_slots = kring->nkr_num_slots;
- g_ring.head = kring->rhead;
- g_ring.cur = kring->rcur;
+ shadow_ring.head = kring->rhead;
+ shadow_ring.cur = kring->rcur;
/* Disable notifications. */
ptring_kick_enable(ptring, 0);
/* Copy the guest kring pointers from the CSB */
- ptnetmap_host_read_kring_csb(ptring, &g_ring, num_slots);
+ ptnetmap_host_read_kring_csb(ptring, &shadow_ring, num_slots);
for (;;) {
uint32_t hwtail;
-#ifndef PTN_AVOID_NM_PROLOGUE
/* Netmap prologue */
- if (unlikely(nm_rxsync_prologue(kring, &g_ring) >= num_slots)) {
- ptnetmap_ring_reinit(kring, g_ring.head, g_ring.cur);
- /* Reenable notifications. */
+ shadow_ring.tail = kring->rtail;
+ if (unlikely(nm_rxsync_prologue(kring, &shadow_ring) >= num_slots)) {
+ /* Reinit ring and enable notifications. */
+ netmap_ring_reinit(kring);
ptring_kick_enable(ptring, 1);
break;
}
-#else /* PTN_AVOID_NM_PROLOGUE */
- kring->rhead = g_ring.head;
- kring->rcur = g_ring.cur;
-#endif /* !PTN_AVOID_NM_PROLOGUE */
- if (unlikely(netmap_verbose & NM_VERB_RXSYNC))
+ if (unlikely(netmap_verbose & NM_VERB_RXSYNC)) {
ptnetmap_kring_dump("pre rxsync", kring);
+ }
IFRATE(pre_tail = kring->rtail);
-
- if (unlikely(kring->nm_sync(kring, g_ring.flags))) {
+ if (unlikely(kring->nm_sync(kring, shadow_ring.flags))) {
/* Reenable notifications. */
ptring_kick_enable(ptring, 1);
D("ERROR rxsync()");
@@ -516,8 +491,9 @@ ptnetmap_rx_handler(void *data)
IFRATE(rate_batch_stats_update(&ptns->rate_ctx.new.rxbs, pre_tail,
kring->rtail, num_slots));
- if (unlikely(netmap_verbose & NM_VERB_RXSYNC))
+ if (unlikely(netmap_verbose & NM_VERB_RXSYNC)) {
ptnetmap_kring_dump("post rxsync", kring);
+ }
#ifndef BUSY_WAIT
/* Interrupt the guest if needed. */
@@ -530,9 +506,9 @@ ptnetmap_rx_handler(void *data)
}
#endif
/* Read CSB to see if there is more work to do. */
- ptnetmap_host_read_kring_csb(ptring, &g_ring, num_slots);
+ ptnetmap_host_read_kring_csb(ptring, &shadow_ring, num_slots);
#ifndef BUSY_WAIT
- if (ptnetmap_norxslots(kring, g_ring.head)) {
+ if (ptnetmap_norxslots(kring, shadow_ring.head)) {
/*
* No more slots available for reception. We enable notification and
* go to sleep, waiting for a kick from the guest when new receive
@@ -542,8 +518,8 @@ ptnetmap_rx_handler(void *data)
/* Reenable notifications. */
ptring_kick_enable(ptring, 1);
/* Doublecheck. */
- ptnetmap_host_read_kring_csb(ptring, &g_ring, num_slots);
- if (!ptnetmap_norxslots(kring, g_ring.head)) {
+ ptnetmap_host_read_kring_csb(ptring, &shadow_ring, num_slots);
+ if (!ptnetmap_norxslots(kring, shadow_ring.head)) {
/* We won the race condition, more slots are available. Disable
* notifications and do another cycle. */
ptring_kick_enable(ptring, 0);
@@ -578,7 +554,7 @@ ptnetmap_rx_handler(void *data)
}
}
-#ifdef DEBUG
+#ifdef NETMAP_PT_DEBUG
static void
ptnetmap_print_configuration(struct ptnetmap_cfg *cfg)
{
@@ -594,7 +570,7 @@ ptnetmap_print_configuration(struct ptne
}
}
-#endif
+#endif /* NETMAP_PT_DEBUG */
/* Copy actual state of the host ring into the CSB for the guest init */
static int
Modified: head/sys/net/netmap.h
==============================================================================
--- head/sys/net/netmap.h Tue Oct 18 15:50:20 2016 (r307573)
+++ head/sys/net/netmap.h Tue Oct 18 16:18:25 2016 (r307574)
@@ -655,8 +655,8 @@ struct ptn_vmm_ioctl_msix {
/* IOCTL parameters */
struct nm_kth_ioctl {
- u_long com;
- /* TODO: use union */
+ uint64_t com;
+ /* We use union to support more ioctl commands. */
union {
struct ptn_vmm_ioctl_msix msix;
} data;
@@ -667,5 +667,6 @@ struct ptnet_ring_cfg {
uint64_t ioeventfd; /* eventfd in linux, tsleep() parameter in FreeBSD */
uint64_t irqfd; /* eventfd in linux, ioctl fd in FreeBSD */
struct nm_kth_ioctl ioctl; /* ioctl parameter to send irq (only used in bhyve/FreeBSD) */
+ uint64_t reserved[4]; /* reserved to support of more hypervisors */
};
#endif /* _NET_NETMAP_H_ */
Modified: head/sys/net/netmap_virt.h
==============================================================================
--- head/sys/net/netmap_virt.h Tue Oct 18 15:50:20 2016 (r307573)
+++ head/sys/net/netmap_virt.h Tue Oct 18 16:18:25 2016 (r307574)
@@ -58,13 +58,11 @@
/* Registers for the ptnetmap memdev */
/* 32 bit r/o */
-#define PTNETMAP_IO_PCI_FEATURES 0 /* XXX should be removed */
-/* 32 bit r/o */
-#define PTNETMAP_IO_PCI_MEMSIZE 4 /* size of the netmap memory shared
+#define PTNETMAP_IO_PCI_MEMSIZE 0 /* size of the netmap memory shared
* between guest and host */
/* 16 bit r/o */
-#define PTNETMAP_IO_PCI_HOSTID 8 /* memory allocator ID in netmap host */
-#define PTNETMAP_IO_SIZE 10
+#define PTNETMAP_IO_PCI_HOSTID 4 /* memory allocator ID in netmap host */
+#define PTNETMAP_IO_SIZE 6
/*
* ptnetmap configuration
@@ -115,18 +113,17 @@ ptnetmap_write_cfg(struct nmreq *nmr, st
#define PTNET_IO_PTFEAT 0
#define PTNET_IO_PTCTL 4
#define PTNET_IO_PTSTS 8
-/* hole */
-#define PTNET_IO_MAC_LO 16
-#define PTNET_IO_MAC_HI 20
-#define PTNET_IO_CSBBAH 24
-#define PTNET_IO_CSBBAL 28
-#define PTNET_IO_NIFP_OFS 32
-#define PTNET_IO_NUM_TX_RINGS 36
-#define PTNET_IO_NUM_RX_RINGS 40
-#define PTNET_IO_NUM_TX_SLOTS 44
-#define PTNET_IO_NUM_RX_SLOTS 48
-#define PTNET_IO_VNET_HDR_LEN 52
-#define PTNET_IO_END 56
+#define PTNET_IO_MAC_LO 12
+#define PTNET_IO_MAC_HI 16
+#define PTNET_IO_CSBBAH 20
+#define PTNET_IO_CSBBAL 24
+#define PTNET_IO_NIFP_OFS 28
+#define PTNET_IO_NUM_TX_RINGS 32
+#define PTNET_IO_NUM_RX_RINGS 36
+#define PTNET_IO_NUM_TX_SLOTS 40
+#define PTNET_IO_NUM_RX_SLOTS 44
+#define PTNET_IO_VNET_HDR_LEN 48
+#define PTNET_IO_END 52
#define PTNET_IO_KICK_BASE 128
#define PTNET_IO_MASK 0xff
@@ -139,11 +136,11 @@ struct ptnet_ring {
uint32_t head; /* GW+ HR+ the head of the guest netmap_ring */
uint32_t cur; /* GW+ HR+ the cur of the guest netmap_ring */
uint32_t guest_need_kick; /* GW+ HR+ host-->guest notification enable */
- char pad[4];
+ uint32_t sync_flags; /* GW+ HR+ the flags of the guest [tx|rx]sync() */
uint32_t hwcur; /* GR+ HW+ the hwcur of the host netmap_kring */
uint32_t hwtail; /* GR+ HW+ the hwtail of the host netmap_kring */
uint32_t host_need_kick; /* GR+ HW+ guest-->host notification enable */
- uint32_t sync_flags; /* GW+ HR+ the flags of the guest [tx|rx]sync() */
+ char pad[4];
};
/* CSB for the ptnet device. */
@@ -165,161 +162,119 @@ ptn_sub(uint32_t l_elem, uint32_t r_elem
}
#endif /* WITH_PTNETMAP_HOST || WITH_PTNETMAP_GUEST */
-#ifdef WITH_PTNETMAP_HOST
-/*
- * ptnetmap kernel thread routines
- * */
-
-/* Functions to read and write CSB fields in the host */
-#if defined (linux)
-#define CSB_READ(csb, field, r) (get_user(r, &csb->field))
-#define CSB_WRITE(csb, field, v) (put_user(v, &csb->field))
-#else /* ! linux */
-#define CSB_READ(csb, field, r) (r = fuword32(&csb->field))
-#define CSB_WRITE(csb, field, v) (suword32(&csb->field, v))
-#endif /* ! linux */
+#ifdef WITH_PTNETMAP_GUEST
-/*
- * HOST read/write kring pointers from/in CSB
- */
+/* ptnetmap_memdev routines used to talk with ptnetmap_memdev device driver */
+struct ptnetmap_memdev;
+int nm_os_pt_memdev_iomap(struct ptnetmap_memdev *, vm_paddr_t *, void **);
+void nm_os_pt_memdev_iounmap(struct ptnetmap_memdev *);
-/* Host: Read kring pointers (head, cur, sync_flags) from CSB */
+/* Guest driver: Write kring pointers (cur, head) to the CSB.
+ * This routine is coupled with ptnetmap_host_read_kring_csb(). */
static inline void
-ptnetmap_host_read_kring_csb(struct ptnet_ring __user *ptr,
- struct netmap_ring *g_ring,
- uint32_t num_slots)
+ptnetmap_guest_write_kring_csb(struct ptnet_ring *ptr, uint32_t cur,
+ uint32_t head)
{
- uint32_t old_head = g_ring->head, old_cur = g_ring->cur;
- uint32_t d, inc_h, inc_c;
-
- //mb(); /* Force memory complete before read CSB */
-
/*
- * We must first read head and then cur with a barrier in the
- * middle, because cur can exceed head, but not vice versa.
- * The guest must first write cur and then head with a barrier.
- *
- * head <= cur
- *
- * guest host
- *
- * STORE(cur) LOAD(head)
- * mb() ----------- mb()
- * STORE(head) LOAD(cur)
- *
- * This approach ensures that every head that we read is
- * associated with the correct cur. In this way head can not exceed cur.
+ * We need to write cur and head to the CSB but we cannot do it atomically.
+ * There is no way we can prevent the host from reading the updated value
+ * of one of the two and the old value of the other. However, if we make
+ * sure that the host never reads a value of head more recent than the
+ * value of cur we are safe. We can allow the host to read a value of cur
+ * more recent than the value of head, since in the netmap ring cur can be
+ * ahead of head and cur cannot wrap around head because it must be behind
+ * tail. Inverting the order of writes below could instead result into the
+ * host to think head went ahead of cur, which would cause the sync
+ * prologue to fail.
+ *
+ * The following memory barrier scheme is used to make this happen:
+ *
+ * Guest Host
+ *
+ * STORE(cur) LOAD(head)
+ * mb() <-----------> mb()
+ * STORE(head) LOAD(cur)
*/
- CSB_READ(ptr, head, g_ring->head);
+ ptr->cur = cur;
mb();
- CSB_READ(ptr, cur, g_ring->cur);
- CSB_READ(ptr, sync_flags, g_ring->flags);
-
- /*
- * Even with the previous barrier, it is still possible that we read an
- * updated cur and an old head.
- * To detect this situation, we can check if the new cur overtakes
- * the (apparently) new head.
- */
- d = ptn_sub(old_cur, old_head, num_slots); /* previous distance */
- inc_c = ptn_sub(g_ring->cur, old_cur, num_slots); /* increase of cur */
- inc_h = ptn_sub(g_ring->head, old_head, num_slots); /* increase of head */
-
- if (unlikely(inc_c > num_slots - d + inc_h)) { /* cur overtakes head */
- ND(1,"ERROR cur overtakes head - old_cur: %u cur: %u old_head: %u head: %u",
- old_cur, g_ring->cur, old_head, g_ring->head);
- g_ring->cur = nm_prev(g_ring->head, num_slots - 1);
- //*g_cur = *g_head;
- }
+ ptr->head = head;
}
-/* Host: Write kring pointers (hwcur, hwtail) into the CSB */
+/* Guest driver: Read kring pointers (hwcur, hwtail) from the CSB.
+ * This routine is coupled with ptnetmap_host_write_kring_csb(). */
static inline void
-ptnetmap_host_write_kring_csb(struct ptnet_ring __user *ptr, uint32_t hwcur,
- uint32_t hwtail)
+ptnetmap_guest_read_kring_csb(struct ptnet_ring *ptr, struct netmap_kring *kring)
{
- /* We must write hwtail before hwcur (see below). */
- CSB_WRITE(ptr, hwtail, hwtail);
+ /*
+ * We place a memory barrier to make sure that the update of hwtail never
+ * overtakes the update of hwcur.
+ * (see explanation in ptnetmap_host_write_kring_csb).
+ */
+ kring->nr_hwtail = ptr->hwtail;
mb();
- CSB_WRITE(ptr, hwcur, hwcur);
-
- //mb(); /* Force memory complete before send notification */
+ kring->nr_hwcur = ptr->hwcur;
}
-#endif /* WITH_PTNETMAP_HOST */
+#endif /* WITH_PTNETMAP_GUEST */
-#ifdef WITH_PTNETMAP_GUEST
+#ifdef WITH_PTNETMAP_HOST
/*
- * GUEST read/write kring pointers from/in CSB.
- * To use into device driver.
- */
+ * ptnetmap kernel thread routines
+ * */
+
+/* Functions to read and write CSB fields in the host */
+#if defined (linux)
+#define CSB_READ(csb, field, r) (get_user(r, &csb->field))
+#define CSB_WRITE(csb, field, v) (put_user(v, &csb->field))
+#else /* ! linux */
+#define CSB_READ(csb, field, r) (r = fuword32(&csb->field))
+#define CSB_WRITE(csb, field, v) (suword32(&csb->field, v))
+#endif /* ! linux */
-/* Guest: Write kring pointers (cur, head) into the CSB */
+/* Host netmap: Write kring pointers (hwcur, hwtail) to the CSB.
+ * This routine is coupled with ptnetmap_guest_read_kring_csb(). */
static inline void
-ptnetmap_guest_write_kring_csb(struct ptnet_ring *ptr, uint32_t cur,
- uint32_t head)
+ptnetmap_host_write_kring_csb(struct ptnet_ring __user *ptr, uint32_t hwcur,
+ uint32_t hwtail)
{
- /* We must write cur before head for sync reason (see above) */
- ptr->cur = cur;
+ /*
+ * The same scheme used in ptnetmap_guest_write_kring_csb() applies here.
+ * We allow the guest to read a value of hwcur more recent than the value
+ * of hwtail, since this would anyway result in a consistent view of the
+ * ring state (and hwcur can never wraparound hwtail, since hwcur must be
+ * behind head).
+ *
+ * The following memory barrier scheme is used to make this happen:
+ *
+ * Guest Host
+ *
+ * STORE(hwcur) LOAD(hwtail)
+ * mb() <-------------> mb()
+ * STORE(hwtail) LOAD(hwcur)
+ */
+ CSB_WRITE(ptr, hwcur, hwcur);
mb();
- ptr->head = head;
-
- //mb(); /* Force memory complete before send notification */
+ CSB_WRITE(ptr, hwtail, hwtail);
}
-/* Guest: Read kring pointers (hwcur, hwtail) from CSB */
+/* Host netmap: Read kring pointers (head, cur, sync_flags) from the CSB.
+ * This routine is coupled with ptnetmap_guest_write_kring_csb(). */
static inline void
-ptnetmap_guest_read_kring_csb(struct ptnet_ring *ptr, struct netmap_kring *kring)
+ptnetmap_host_read_kring_csb(struct ptnet_ring __user *ptr,
+ struct netmap_ring *shadow_ring,
+ uint32_t num_slots)
{
- uint32_t old_hwcur = kring->nr_hwcur, old_hwtail = kring->nr_hwtail;
- uint32_t num_slots = kring->nkr_num_slots;
- uint32_t d, inc_hc, inc_ht;
-
- //mb(); /* Force memory complete before read CSB */
-
/*
- * We must first read hwcur and then hwtail with a barrier in the
- * middle, because hwtail can exceed hwcur, but not vice versa.
- * The host must first write hwtail and then hwcur with a barrier.
- *
- * hwcur <= hwtail
- *
- * host guest
- *
- * STORE(hwtail) LOAD(hwcur)
- * mb() --------- mb()
- * STORE(hwcur) LOAD(hwtail)
- *
- * This approach ensures that every hwcur that the guest reads is
- * associated with the correct hwtail. In this way hwcur can not exceed
- * hwtail.
+ * We place a memory barrier to make sure that the update of head never
+ * overtakes the update of cur.
+ * (see explanation in ptnetmap_guest_write_kring_csb).
*/
- kring->nr_hwcur = ptr->hwcur;
+ CSB_READ(ptr, head, shadow_ring->head);
mb();
- kring->nr_hwtail = ptr->hwtail;
-
- /*
- * Even with the previous barrier, it is still possible that we read an
- * updated hwtail and an old hwcur.
- * To detect this situation, we can check if the new hwtail overtakes
- * the (apparently) new hwcur.
- */
- d = ptn_sub(old_hwtail, old_hwcur, num_slots); /* previous distance */
- inc_ht = ptn_sub(kring->nr_hwtail, old_hwtail, num_slots); /* increase of hwtail */
- inc_hc = ptn_sub(kring->nr_hwcur, old_hwcur, num_slots); /* increase of hwcur */
-
- if (unlikely(inc_ht > num_slots - d + inc_hc)) {
- ND(1, "ERROR hwtail overtakes hwcur - old_hwtail: %u hwtail: %u old_hwcur: %u hwcur: %u",
- old_hwtail, kring->nr_hwtail, old_hwcur, kring->nr_hwcur);
- kring->nr_hwtail = nm_prev(kring->nr_hwcur, num_slots - 1);
- //kring->nr_hwtail = kring->nr_hwcur;
- }
+ CSB_READ(ptr, cur, shadow_ring->cur);
+ CSB_READ(ptr, sync_flags, shadow_ring->flags);
}
-/* ptnetmap_memdev routines used to talk with ptnetmap_memdev device driver */
-struct ptnetmap_memdev;
-int nm_os_pt_memdev_iomap(struct ptnetmap_memdev *, vm_paddr_t *, void **);
-void nm_os_pt_memdev_iounmap(struct ptnetmap_memdev *);
-#endif /* WITH_PTNETMAP_GUEST */
+#endif /* WITH_PTNETMAP_HOST */
#endif /* NETMAP_VIRT_H */
More information about the svn-src-all
mailing list