Skip to content

Commit

Permalink
mm: reduce lock contention of pcp buffer refill
Browse files Browse the repository at this point in the history
rmqueue_bulk() batches the allocation of multiple elements to refill the
per-CPU buffers into a single hold of the zone lock.  Each element is
allocated and checked using check_pcp_refill().  The check touches every
related struct page which is especially expensive for higher order
allocations (huge pages).

This patch reduces the time holding the lock by moving the check out of
the critical section similar to rmqueue_buddy() which allocates a single
element.

Measurements of parallel allocation-heavy workloads show a reduction of
the average huge page allocation latency of 50 percent for two cores and
nearly 90 percent for 24 cores.

Link: https://lkml.kernel.org/r/20230201162549.68384-1-halbuer@sra.uni-hannover.de
Signed-off-by: Alexander Halbuer <halbuer@sra.uni-hannover.de>
Cc: Mel Gorman <mgorman@techsingularity.net>
Cc: Vlastimil Babka <vbabka@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
  • Loading branch information
halbuer authored and akpm00 committed Mar 28, 2023
1 parent a4a4659 commit 2ede3c1
Showing 1 changed file with 18 additions and 4 deletions.
22 changes: 18 additions & 4 deletions mm/page_alloc.c
Original file line number Diff line number Diff line change
Expand Up @@ -3138,6 +3138,8 @@ static int rmqueue_bulk(struct zone *zone, unsigned int order,
{
unsigned long flags;
int i, allocated = 0;
struct list_head *prev_tail = list->prev;
struct page *pos, *n;

spin_lock_irqsave(&zone->lock, flags);
for (i = 0; i < count; ++i) {
Expand All @@ -3146,9 +3148,6 @@ static int rmqueue_bulk(struct zone *zone, unsigned int order,
if (unlikely(page == NULL))
break;

if (unlikely(check_pcp_refill(page, order)))
continue;

/*
* Split buddy pages returned by expand() are received here in
* physical page order. The page is added to the tail of
Expand All @@ -3160,7 +3159,6 @@ static int rmqueue_bulk(struct zone *zone, unsigned int order,
* pages are ordered properly.
*/
list_add_tail(&page->pcp_list, list);
allocated++;
if (is_migrate_cma(get_pcppage_migratetype(page)))
__mod_zone_page_state(zone, NR_FREE_CMA_PAGES,
-(1 << order));
Expand All @@ -3174,6 +3172,22 @@ static int rmqueue_bulk(struct zone *zone, unsigned int order,
*/
__mod_zone_page_state(zone, NR_FREE_PAGES, -(i << order));
spin_unlock_irqrestore(&zone->lock, flags);

/*
* Pages are appended to the pcp list without checking to reduce the
* time holding the zone lock. Checking the appended pages happens right
* after the critical section while still holding the pcp lock.
*/
pos = list_first_entry(prev_tail, struct page, pcp_list);
list_for_each_entry_safe_from(pos, n, list, pcp_list) {
if (unlikely(check_pcp_refill(pos, order))) {
list_del(&pos->pcp_list);
continue;
}

allocated++;
}

return allocated;
}

Expand Down

0 comments on commit 2ede3c1

Please sign in to comment.