svn commit: r348484 - head/sys/vm

Kubilay Kocak koobs at FreeBSD.org
Sat Jun 1 03:05:22 UTC 2019


On 1/06/2019 7:02 am, Doug Moore wrote:
> Author: dougm
> Date: Fri May 31 21:02:42 2019
> New Revision: 348484
> URL: https://svnweb.freebsd.org/changeset/base/348484
> 
> Log:
>    The function vm_phys_free_contig invokes vm_phys_free_pages for every
>    power-of-two page block it frees, launching an unsuccessful search for
>    a buddy to pair up with each time.  The only possible buddy-up mergers
>    are across the boundaries of the freed region, so change
>    vm_phys_free_contig simply to enqueue the freed interior blocks, via a
>    new function vm_phys_enqueue_contig, and then call vm_phys_free_pages
>    on the bounding blocks to create as big a cross-boundary block as
>    possible after buddy-merging.
>    
>    The only callers of vm_phys_free_contig at the moment call it in
>    situations where merging blocks across the boundary is clearly
>    impossible, so just call vm_phys_enqueue_contig in those places and
>    avoid trying to buddy-up at all.
>    
>    One beneficiary of this change is in breaking reservations.  For the
>    case where memory is freed in breaking a reservation with only the
>    first and last pages allocated, the number of cycles consumed by the
>    operation drops about 11% with this change.
>    
>    Suggested by: alc
>    Reviewed by: alc
>    Approved by: kib, markj (mentors)
>    Differential Revision: https://reviews.freebsd.org/D16901

Can this be an MFC candidate?


> Modified:
>    head/sys/vm/vm_page.c
>    head/sys/vm/vm_phys.c
>    head/sys/vm/vm_phys.h
>    head/sys/vm/vm_reserv.c
> 
> Modified: head/sys/vm/vm_page.c
> ==============================================================================
> --- head/sys/vm/vm_page.c	Fri May 31 20:36:32 2019	(r348483)
> +++ head/sys/vm/vm_page.c	Fri May 31 21:02:42 2019	(r348484)
> @@ -833,7 +833,7 @@ vm_page_startup(vm_offset_t vaddr)
>   
>   			vmd = VM_DOMAIN(seg->domain);
>   			vm_domain_free_lock(vmd);
> -			vm_phys_free_contig(m, pagecount);
> +			vm_phys_enqueue_contig(m, pagecount);
>   			vm_domain_free_unlock(vmd);
>   			vm_domain_freecnt_inc(vmd, pagecount);
>   			vm_cnt.v_page_count += (u_int)pagecount;
> 
> Modified: head/sys/vm/vm_phys.c
> ==============================================================================
> --- head/sys/vm/vm_phys.c	Fri May 31 20:36:32 2019	(r348483)
> +++ head/sys/vm/vm_phys.c	Fri May 31 21:02:42 2019	(r348484)
> @@ -1095,14 +1095,35 @@ vm_phys_free_pages(vm_page_t m, int order)
>   }
>   
>   /*
> - * Free a contiguous, arbitrarily sized set of physical pages.
> + * Return the largest possible order of a set of pages starting at m.
> + */
> +static int
> +max_order(vm_page_t m)
> +{
> +
> +	/*
> +	 * Unsigned "min" is used here so that "order" is assigned
> +	 * "VM_NFREEORDER - 1" when "m"'s physical address is zero
> +	 * or the low-order bits of its physical address are zero
> +	 * because the size of a physical address exceeds the size of
> +	 * a long.
> +	 */
> +	return (min(ffsl(VM_PAGE_TO_PHYS(m) >> PAGE_SHIFT) - 1,
> +	    VM_NFREEORDER - 1));
> +}
> +
> +/*
> + * Free a contiguous, arbitrarily sized set of physical pages, without
> + * merging across set boundaries.
>    *
>    * The free page queues must be locked.
>    */
>   void
> -vm_phys_free_contig(vm_page_t m, u_long npages)
> +vm_phys_enqueue_contig(vm_page_t m, u_long npages)
>   {
> -	u_int n;
> +	struct vm_freelist *fl;
> +	struct vm_phys_seg *seg;
> +	vm_page_t m_end;
>   	int order;
>   
>   	/*
> @@ -1110,29 +1131,68 @@ vm_phys_free_contig(vm_page_t m, u_long npages)
>   	 * possible power-of-two-sized subsets.
>   	 */
>   	vm_domain_free_assert_locked(vm_pagequeue_domain(m));
> -	for (;; npages -= n) {
> -		/*
> -		 * Unsigned "min" is used here so that "order" is assigned
> -		 * "VM_NFREEORDER - 1" when "m"'s physical address is zero
> -		 * or the low-order bits of its physical address are zero
> -		 * because the size of a physical address exceeds the size of
> -		 * a long.
> -		 */
> -		order = min(ffsl(VM_PAGE_TO_PHYS(m) >> PAGE_SHIFT) - 1,
> -		    VM_NFREEORDER - 1);
> -		n = 1 << order;
> -		if (npages < n)
> -			break;
> -		vm_phys_free_pages(m, order);
> -		m += n;
> +	seg = &vm_phys_segs[m->segind];
> +	fl = (*seg->free_queues)[m->pool];
> +	m_end = m + npages;
> +	/* Free blocks of increasing size. */
> +	while ((order = max_order(m)) < VM_NFREEORDER - 1 &&
> +	    m + (1 << order) <= m_end) {
> +		KASSERT(seg == &vm_phys_segs[m->segind],
> +		    ("%s: page range [%p,%p) spans multiple segments",
> +		    __func__, m_end - npages, m));
> +		vm_freelist_add(fl, m, order, 1);
> +		m += 1 << order;
>   	}
> -	/* The residual "npages" is less than "1 << (VM_NFREEORDER - 1)". */
> -	for (; npages > 0; npages -= n) {
> -		order = flsl(npages) - 1;
> -		n = 1 << order;
> -		vm_phys_free_pages(m, order);
> -		m += n;
> +	/* Free blocks of maximum size. */
> +	while (m + (1 << order) <= m_end) {
> +		KASSERT(seg == &vm_phys_segs[m->segind],
> +		    ("%s: page range [%p,%p) spans multiple segments",
> +		    __func__, m_end - npages, m));
> +		vm_freelist_add(fl, m, order, 1);
> +		m += 1 << order;
>   	}
> +	/* Free blocks of diminishing size. */
> +	while (m < m_end) {
> +		KASSERT(seg == &vm_phys_segs[m->segind],
> +		    ("%s: page range [%p,%p) spans multiple segments",
> +		    __func__, m_end - npages, m));
> +		order = flsl(m_end - m) - 1;
> +		vm_freelist_add(fl, m, order, 1);
> +		m += 1 << order;
> +	}
> +}
> +
> +/*
> + * Free a contiguous, arbitrarily sized set of physical pages.
> + *
> + * The free page queues must be locked.
> + */
> +void
> +vm_phys_free_contig(vm_page_t m, u_long npages)
> +{
> +	int order_start, order_end;
> +	vm_page_t m_start, m_end;
> +
> +	vm_domain_free_assert_locked(vm_pagequeue_domain(m));
> +
> +	m_start = m;
> +	order_start = max_order(m_start);
> +	if (order_start < VM_NFREEORDER - 1)
> +		m_start += 1 << order_start;
> +	m_end = m + npages;
> +	order_end = max_order(m_end);
> +	if (order_end < VM_NFREEORDER - 1)
> +		m_end -= 1 << order_end;
> +	/*
> +	 * Avoid unnecessary coalescing by freeing the pages at the start and
> +	 * end of the range last.
> +	 */
> +	if (m_start < m_end)
> +		vm_phys_enqueue_contig(m_start, m_end - m_start);
> +	if (order_start < VM_NFREEORDER - 1)
> +		vm_phys_free_pages(m, order_start);
> +	if (order_end < VM_NFREEORDER - 1)
> +		vm_phys_free_pages(m_end, order_end);
>   }
>   
>   /*
> 
> Modified: head/sys/vm/vm_phys.h
> ==============================================================================
> --- head/sys/vm/vm_phys.h	Fri May 31 20:36:32 2019	(r348483)
> +++ head/sys/vm/vm_phys.h	Fri May 31 21:02:42 2019	(r348484)
> @@ -84,6 +84,7 @@ vm_page_t vm_phys_alloc_freelist_pages(int domain, int
>   int vm_phys_alloc_npages(int domain, int pool, int npages, vm_page_t ma[]);
>   vm_page_t vm_phys_alloc_pages(int domain, int pool, int order);
>   int vm_phys_domain_match(int prefer, vm_paddr_t low, vm_paddr_t high);
> +void vm_phys_enqueue_contig(vm_page_t m, u_long npages);
>   int vm_phys_fictitious_reg_range(vm_paddr_t start, vm_paddr_t end,
>       vm_memattr_t memattr);
>   void vm_phys_fictitious_unreg_range(vm_paddr_t start, vm_paddr_t end);
> 
> Modified: head/sys/vm/vm_reserv.c
> ==============================================================================
> --- head/sys/vm/vm_reserv.c	Fri May 31 20:36:32 2019	(r348483)
> +++ head/sys/vm/vm_reserv.c	Fri May 31 21:02:42 2019	(r348484)
> @@ -1066,7 +1066,7 @@ vm_reserv_break(vm_reserv_t rv)
>   			else {
>   				hi = NBPOPMAP * i + bitpos;
>   				vm_domain_free_lock(VM_DOMAIN(rv->domain));
> -				vm_phys_free_contig(&rv->pages[lo], hi - lo);
> +				vm_phys_enqueue_contig(&rv->pages[lo], hi - lo);
>   				vm_domain_free_unlock(VM_DOMAIN(rv->domain));
>   				lo = hi;
>   			}



More information about the svn-src-head mailing list