svn commit: r269212 - head/sys/arm/arm

Ian Lepore ian at FreeBSD.org
Tue Jul 29 02:36:42 UTC 2014


Author: ian
Date: Tue Jul 29 02:36:41 2014
New Revision: 269212
URL: http://svnweb.freebsd.org/changeset/base/269212

Log:
  Memory belonging to an mbuf, or allocated by bus_dmamem_alloc(), never
  triggers a need to bounce due to cacheline alignment.  These buffers
  are always aligned to cacheline boundaries, and even when the DMA operation
  starts at an offset within the buffer or doesn't extend to the end of the
  buffer, it's safe to flush the complete cachelines that were only partially
  involved in the DMA.  This is because there's a very strict rule on these
  types of buffers that there will not be concurrent access by the CPU and
  one or more DMA transfers within the buffer.
  
  Reviewed by:	cognet

Modified:
  head/sys/arm/arm/busdma_machdep-v6.c

Modified: head/sys/arm/arm/busdma_machdep-v6.c
==============================================================================
--- head/sys/arm/arm/busdma_machdep-v6.c	Tue Jul 29 02:36:27 2014	(r269211)
+++ head/sys/arm/arm/busdma_machdep-v6.c	Tue Jul 29 02:36:41 2014	(r269212)
@@ -162,6 +162,8 @@ struct bus_dmamap {
 	void		      *callback_arg;
 	int		      flags;
 #define DMAMAP_COHERENT		(1 << 0)
+#define DMAMAP_DMAMEM_ALLOC	(1 << 1)
+#define DMAMAP_MBUF		(1 << 2)
 	STAILQ_ENTRY(bus_dmamap) links;
 	int		       sync_count;
 	struct sync_list       slist[];
@@ -279,12 +281,22 @@ alignment_bounce(bus_dma_tag_t dmat, bus
 }
 
 /*
- * Return true if the buffer start or end does not fall on a cacheline boundary.
+ * Return true if the DMA should bounce because the start or end does not fall
+ * on a cacheline boundary (which would require a partial cacheline flush).
+ * COHERENT memory doesn't trigger cacheline flushes.  Memory allocated by
+ * bus_dmamem_alloc() is always aligned to cacheline boundaries, and there's a
+ * strict rule that such memory cannot be accessed by the CPU while DMA is in
+ * progress (or by multiple DMA engines at once), so that it's always safe to do
+ * full cacheline flushes even if that affects memory outside the range of a
+ * given DMA operation that doesn't involve the full allocated buffer.  If we're
+ * mapping an mbuf, that follows the same rules as a buffer we allocated.
  */
 static __inline int
-cacheline_bounce(bus_addr_t addr, bus_size_t size)
+cacheline_bounce(bus_dmamap_t map, bus_addr_t addr, bus_size_t size)
 {
 
+	if (map->flags & (DMAMAP_DMAMEM_ALLOC | DMAMAP_COHERENT | DMAMAP_MBUF))
+		return (0);
 	return ((addr | size) & arm_dcache_align_mask);
 }
 
@@ -302,8 +314,9 @@ static __inline int
 might_bounce(bus_dma_tag_t dmat, bus_dmamap_t map, bus_addr_t addr, 
     bus_size_t size)
 {
-	return ((dmat->flags & BUS_DMA_COULD_BOUNCE) ||
-	    !((map->flags & DMAMAP_COHERENT) && cacheline_bounce(addr, size)));
+	return ((dmat->flags & BUS_DMA_EXCL_BOUNCE) ||
+	    alignment_bounce(dmat, addr) ||
+	    cacheline_bounce(map, addr, size));
 }
 
 /*
@@ -322,8 +335,7 @@ must_bounce(bus_dma_tag_t dmat, bus_dmam
     bus_size_t size)
 {
 
-	/* Coherent memory doesn't need to bounce due to cache alignment. */
-	if (!(map->flags & DMAMAP_COHERENT) && cacheline_bounce(paddr, size))
+	if (cacheline_bounce(map, paddr, size))
 		return (1);
 
 	/*
@@ -727,7 +739,9 @@ bus_dmamem_alloc(bus_dma_tag_t dmat, voi
 		return (ENOMEM);
 	}
 
+	(*mapp)->flags = DMAMAP_DMAMEM_ALLOC;
 	(*mapp)->sync_count = 0;
+
 	/* We may need bounce pages, even for allocated memory */
 	error = allocate_bz_and_pages(dmat, *mapp);
 	if (error != 0) {
@@ -1080,6 +1094,9 @@ _bus_dmamap_load_buffer(bus_dma_tag_t dm
 	if (segs == NULL)
 		segs = dmat->segments;
 
+	if (flags & BUS_DMA_LOAD_MBUF)
+		map->flags |= DMAMAP_MBUF;
+
 	map->pmap = pmap;
 
 	if (might_bounce(dmat, map, (bus_addr_t)buf, buflen)) {
@@ -1196,6 +1213,7 @@ _bus_dmamap_unload(bus_dma_tag_t dmat, b
 		map->pagesneeded = 0;
 	}
 	map->sync_count = 0;
+	map->flags &= ~DMAMAP_MBUF;
 }
 
 #ifdef notyetbounceuser


More information about the svn-src-all mailing list