// SPDX-License-Identifier: GPL-2.0-only /* * Common interface for implementing a memory balloon, including support * for migration of pages inflated in a memory balloon. * * Copyright (C) 2012, Red Hat, Inc. Rafael Aquini */ #include #include #include #include /* * Lock protecting the balloon_dev_info of all devices. We don't really * expect more than one device. */ static DEFINE_SPINLOCK(balloon_pages_lock); /** * balloon_page_insert - insert a page into the balloon's page list and make * the page->private assignment accordingly. * @balloon : pointer to balloon device * @page : page to be assigned as a 'balloon page' * * Caller must ensure the balloon_pages_lock is held. */ static void balloon_page_insert(struct balloon_dev_info *balloon, struct page *page) { lockdep_assert_held(&balloon_pages_lock); __SetPageOffline(page); if (IS_ENABLED(CONFIG_BALLOON_MIGRATION)) { SetPageMovableOps(page); set_page_private(page, (unsigned long)balloon); } list_add(&page->lru, &balloon->pages); } /** * balloon_page_finalize - prepare a balloon page that was removed from the * balloon list for release to the page allocator * @page: page to be released to the page allocator * * Caller must ensure the balloon_pages_lock is held. */ static void balloon_page_finalize(struct page *page) { lockdep_assert_held(&balloon_pages_lock); if (IS_ENABLED(CONFIG_BALLOON_MIGRATION)) set_page_private(page, 0); /* PageOffline is sticky until the page is freed to the buddy. */ } static void balloon_page_enqueue_one(struct balloon_dev_info *b_dev_info, struct page *page) { balloon_page_insert(b_dev_info, page); if (b_dev_info->adjust_managed_page_count) adjust_managed_page_count(page, -1); __count_vm_event(BALLOON_INFLATE); inc_node_page_state(page, NR_BALLOON_PAGES); } /** * balloon_page_list_enqueue() - inserts a list of pages into the balloon page * list. * @b_dev_info: balloon device descriptor where we will insert a new page to * @pages: pages to enqueue - allocated using balloon_page_alloc. * * Driver must call this function to properly enqueue balloon pages before * definitively removing them from the guest system. * * Return: number of pages that were enqueued. */ size_t balloon_page_list_enqueue(struct balloon_dev_info *b_dev_info, struct list_head *pages) { struct page *page, *tmp; unsigned long flags; size_t n_pages = 0; spin_lock_irqsave(&balloon_pages_lock, flags); list_for_each_entry_safe(page, tmp, pages, lru) { list_del(&page->lru); balloon_page_enqueue_one(b_dev_info, page); n_pages++; } spin_unlock_irqrestore(&balloon_pages_lock, flags); return n_pages; } EXPORT_SYMBOL_GPL(balloon_page_list_enqueue); /** * balloon_page_list_dequeue() - removes pages from balloon's page list and * returns a list of the pages. * @b_dev_info: balloon device descriptor where we will grab a page from. * @pages: pointer to the list of pages that would be returned to the caller. * @n_req_pages: number of requested pages. * * Driver must call this function to properly de-allocate a previous enlisted * balloon pages before definitively releasing it back to the guest system. * This function tries to remove @n_req_pages from the ballooned pages and * return them to the caller in the @pages list. * * Note that this function may fail to dequeue some pages even if the balloon * isn't empty - since the page list can be temporarily empty due to compaction * of isolated pages. * * Return: number of pages that were added to the @pages list. */ size_t balloon_page_list_dequeue(struct balloon_dev_info *b_dev_info, struct list_head *pages, size_t n_req_pages) { struct page *page, *tmp; unsigned long flags; size_t n_pages = 0; spin_lock_irqsave(&balloon_pages_lock, flags); list_for_each_entry_safe(page, tmp, &b_dev_info->pages, lru) { if (n_pages == n_req_pages) break; list_del(&page->lru); if (b_dev_info->adjust_managed_page_count) adjust_managed_page_count(page, 1); balloon_page_finalize(page); __count_vm_event(BALLOON_DEFLATE); list_add(&page->lru, pages); dec_node_page_state(page, NR_BALLOON_PAGES); n_pages++; } spin_unlock_irqrestore(&balloon_pages_lock, flags); return n_pages; } EXPORT_SYMBOL_GPL(balloon_page_list_dequeue); /** * balloon_page_alloc - allocates a new page for insertion into the balloon * page list. * * Driver must call this function to properly allocate a new balloon page. * Driver must call balloon_page_enqueue before definitively removing the page * from the guest system. * * Return: struct page for the allocated page or NULL on allocation failure. */ struct page *balloon_page_alloc(void) { gfp_t gfp_flags = __GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN; if (IS_ENABLED(CONFIG_BALLOON_MIGRATION)) gfp_flags |= GFP_HIGHUSER_MOVABLE; else gfp_flags |= GFP_HIGHUSER; return alloc_page(gfp_flags); } EXPORT_SYMBOL_GPL(balloon_page_alloc); /** * balloon_page_enqueue - inserts a new page into the balloon page list. * * @b_dev_info: balloon device descriptor where we will insert a new page * @page: new page to enqueue - allocated using balloon_page_alloc. * * Drivers must call this function to properly enqueue a new allocated balloon * page before definitively removing the page from the guest system. * * Drivers must not enqueue pages while page->lru is still in * use, and must not use page->lru until a page was unqueued again. */ void balloon_page_enqueue(struct balloon_dev_info *b_dev_info, struct page *page) { unsigned long flags; spin_lock_irqsave(&balloon_pages_lock, flags); balloon_page_enqueue_one(b_dev_info, page); spin_unlock_irqrestore(&balloon_pages_lock, flags); } EXPORT_SYMBOL_GPL(balloon_page_enqueue); /** * balloon_page_dequeue - removes a page from balloon's page list and returns * its address to allow the driver to release the page. * @b_dev_info: balloon device descriptor where we will grab a page from. * * Driver must call this function to properly dequeue a previously enqueued page * before definitively releasing it back to the guest system. * * Caller must perform its own accounting to ensure that this * function is called only if some pages are actually enqueued. * * Note that this function may fail to dequeue some pages even if there are * some enqueued pages - since the page list can be temporarily empty due to * the compaction of isolated pages. * * TODO: remove the caller accounting requirements, and allow caller to wait * until all pages can be dequeued. * * Return: struct page for the dequeued page, or NULL if no page was dequeued. */ struct page *balloon_page_dequeue(struct balloon_dev_info *b_dev_info) { unsigned long flags; LIST_HEAD(pages); int n_pages; n_pages = balloon_page_list_dequeue(b_dev_info, &pages, 1); if (n_pages != 1) { /* * If we are unable to dequeue a balloon page because the page * list is empty and there are no isolated pages, then something * went out of track and some balloon pages are lost. * BUG() here, otherwise the balloon driver may get stuck in * an infinite loop while attempting to release all its pages. */ spin_lock_irqsave(&balloon_pages_lock, flags); if (unlikely(list_empty(&b_dev_info->pages) && !b_dev_info->isolated_pages)) BUG(); spin_unlock_irqrestore(&balloon_pages_lock, flags); return NULL; } return list_first_entry(&pages, struct page, lru); } EXPORT_SYMBOL_GPL(balloon_page_dequeue); #ifdef CONFIG_BALLOON_MIGRATION static struct balloon_dev_info *balloon_page_device(struct page *page) { return (struct balloon_dev_info *)page_private(page); } static bool balloon_page_isolate(struct page *page, isolate_mode_t mode) { struct balloon_dev_info *b_dev_info; unsigned long flags; spin_lock_irqsave(&balloon_pages_lock, flags); b_dev_info = balloon_page_device(page); if (!b_dev_info) { /* * The page already got deflated and removed from the * balloon list. */ spin_unlock_irqrestore(&balloon_pages_lock, flags); return false; } list_del(&page->lru); b_dev_info->isolated_pages++; spin_unlock_irqrestore(&balloon_pages_lock, flags); return true; } static void balloon_page_putback(struct page *page) { struct balloon_dev_info *b_dev_info = balloon_page_device(page); unsigned long flags; /* * When we isolated the page, the page was still inflated in a balloon * device. As isolated balloon pages cannot get deflated, we still have * a balloon device here. */ if (WARN_ON_ONCE(!b_dev_info)) return; spin_lock_irqsave(&balloon_pages_lock, flags); list_add(&page->lru, &b_dev_info->pages); b_dev_info->isolated_pages--; spin_unlock_irqrestore(&balloon_pages_lock, flags); } static int balloon_page_migrate(struct page *newpage, struct page *page, enum migrate_mode mode) { struct balloon_dev_info *b_dev_info = balloon_page_device(page); unsigned long flags; int rc; /* * When we isolated the page, the page was still inflated in a balloon * device. As isolated balloon pages cannot get deflated, we still have * a balloon device here. */ if (WARN_ON_ONCE(!b_dev_info)) return -EAGAIN; rc = b_dev_info->migratepage(b_dev_info, newpage, page, mode); if (rc < 0 && rc != -ENOENT) return rc; spin_lock_irqsave(&balloon_pages_lock, flags); if (!rc) { /* Insert the new page into the balloon list. */ get_page(newpage); balloon_page_insert(b_dev_info, newpage); __count_vm_event(BALLOON_MIGRATE); if (b_dev_info->adjust_managed_page_count && page_zone(page) != page_zone(newpage)) { /* * When we migrate a page to a different zone we * have to fixup the count of both involved zones. */ adjust_managed_page_count(page, 1); adjust_managed_page_count(newpage, -1); } } else { /* Old page was deflated but new page not inflated. */ __count_vm_event(BALLOON_DEFLATE); if (b_dev_info->adjust_managed_page_count) adjust_managed_page_count(page, 1); } b_dev_info->isolated_pages--; /* Free the now-deflated page we isolated in balloon_page_isolate(). */ balloon_page_finalize(page); spin_unlock_irqrestore(&balloon_pages_lock, flags); put_page(page); return 0; } static const struct movable_operations balloon_mops = { .migrate_page = balloon_page_migrate, .isolate_page = balloon_page_isolate, .putback_page = balloon_page_putback, }; static int __init balloon_init(void) { return set_movable_ops(&balloon_mops, PGTY_offline); } core_initcall(balloon_init); #endif /* CONFIG_BALLOON_MIGRATION */