diff options
| author | Andrew Morton <akpm@zip.com.au> | 2002-07-04 08:31:25 -0700 |
|---|---|---|
| committer | Linus Torvalds <torvalds@home.transmeta.com> | 2002-07-04 08:31:25 -0700 |
| commit | a2b41d233e7dd5a78f7eac7491f06c5fdce8ae6a (patch) | |
| tree | 12b83bca9768c24f06edc3e48823660dcd3bebf9 | |
| parent | a263b64790ca9e9ddf1337d772351f105d095234 (diff) | |
[PATCH] always update page->flags atomically
move_from_swap_cache() and move_to_swap_cache() are playing with
page->flags nonatomically. The page is on the LRU at the time and
another CPU could be altering page->flags concurrently.
The patch converts those functions to use atomic operations.
It also rationalises the number of bits which are cleared. It's not
really clear to me what page flags we really want to set to a known
state in there.
It had no right to go clearing PG_arch_1. I'm now clearing PG_arch_1
inside rmqueue() which is still a bit presumptious.
btw: shmem uses PAGE_CACHE_SIZE and swapper_space uses PAGE_SIZE. I've
been carefully maintaining the distinction, but it looks like shmem
will break if we ever do make these values different.
Also, __add_to_page_cache() was performing a non-atomic RMW against
page->flags, under the assumption that it was a newly allocated page
which no other CPU would look at. Not true - this function is used for
moving anon pages into swapcache. Those anon pages are on the LRU -
other CPUs can be performing operations against page->flags while
__add_to_swap_cache is stomping on them. This had me running around in
circles for two days.
So let's move the initialisation of the page state into rmqueue(),
where the page really is new (could do it in page_cache_alloc,
perhaps).
The SetPageLocked() in __add_to_page_cache() is also rather curious.
Seems OK for both pagecache and swapcache so I covered that with a
comment.
2.4 has the same problem. Basically, add_to_swap_cache() can stomp on
another CPU's manipulation of page->flags. After a quick review of the
code there, it is barely conceivable that a concurrent refill_inactve()
could get its PG_referenced and PG_active bits scribbled on. Rather
unlikely because swap_out() will probably see PageActive() and bale
out.
Also, mark_dirty_kiobuf() could have its PG_dirty bit accidentally
cleared (but try_to_swap_out() sets it again later).
But there may be other code paths. Really, I think this needs fixing
in 2.4 - it's horrid.
| -rw-r--r-- | mm/filemap.c | 36 | ||||
| -rw-r--r-- | mm/page_alloc.c | 45 | ||||
| -rw-r--r-- | mm/swap_state.c | 21 | ||||
| -rw-r--r-- | mm/vmscan.c | 2 |
4 files changed, 60 insertions, 44 deletions
diff --git a/mm/filemap.c b/mm/filemap.c index 17b9167e6a15..bcbe3e71e0ef 100644 --- a/mm/filemap.c +++ b/mm/filemap.c @@ -515,24 +515,32 @@ int filemap_fdatawait(struct address_space * mapping) } /* - * This adds a page to the page cache, starting out as locked, - * owned by us, but unreferenced, not uptodate and with no errors. - * The caller must hold a write_lock on the mapping->page_lock. + * This adds a page to the page cache, starting out as locked, unreferenced, + * not uptodate and with no errors. + * + * The caller must hold a write_lock on mapping->page_lock. + * + * This function is used for two things: adding newly allocated pagecache + * pages and for moving existing anon pages into swapcache. + * + * In the case of pagecache pages, the page is new, so we can just run + * SetPageLocked() against it. The other page state flags were set by + * rmqueue() + * + * In the case of swapcache, try_to_swap_out() has already locked the page, so + * SetPageLocked() is ugly-but-OK there too. The required page state has been + * set up by swap_out_add_to_swap_cache(). */ static int __add_to_page_cache(struct page *page, struct address_space *mapping, unsigned long offset) { - page_cache_get(page); - if (radix_tree_insert(&mapping->page_tree, offset, page) < 0) - goto nomem; - page->flags &= ~(1 << PG_uptodate | 1 << PG_error | - 1 << PG_referenced | 1 << PG_arch_1 | 1 << PG_checked); - SetPageLocked(page); - ClearPageDirty(page); - ___add_to_page_cache(page, mapping, offset); - return 0; - nomem: - page_cache_release(page); + if (radix_tree_insert(&mapping->page_tree, offset, page) == 0) { + SetPageLocked(page); + ClearPageDirty(page); + ___add_to_page_cache(page, mapping, offset); + page_cache_get(page); + return 0; + } return -ENOMEM; } diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 8266c5a21751..1957b56a1480 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -99,7 +99,6 @@ static void __free_pages_ok (struct page *page, unsigned int order) if (PageWriteback(page)) BUG(); ClearPageDirty(page); - page->flags &= ~(1<<PG_referenced); if (current->flags & PF_FREE_PAGES) goto local_freelist; @@ -188,6 +187,24 @@ static inline struct page * expand (zone_t *zone, struct page *page, return page; } +/* + * This page is about to be returned from the page allocator + */ +static inline void prep_new_page(struct page *page) +{ + BUG_ON(page->mapping); + BUG_ON(PagePrivate(page)); + BUG_ON(PageLocked(page)); + BUG_ON(PageLRU(page)); + BUG_ON(PageActive(page)); + BUG_ON(PageDirty(page)); + BUG_ON(PageWriteback(page)); + page->flags &= ~(1 << PG_uptodate | 1 << PG_error | + 1 << PG_referenced | 1 << PG_arch_1 | + 1 << PG_checked); + set_page_count(page, 1); +} + static FASTCALL(struct page * rmqueue(zone_t *zone, unsigned int order)); static struct page * rmqueue(zone_t *zone, unsigned int order) { @@ -217,13 +234,9 @@ static struct page * rmqueue(zone_t *zone, unsigned int order) page = expand(zone, page, index, order, curr_order, area); spin_unlock_irqrestore(&zone->lock, flags); - set_page_count(page, 1); if (bad_range(zone, page)) BUG(); - if (PageLRU(page)) - BUG(); - if (PageActive(page)) - BUG(); + prep_new_page(page); return page; } curr_order++; @@ -296,25 +309,9 @@ static struct page * balance_classzone(zone_t * classzone, unsigned int gfp_mask tmp = list_entry(entry, struct page, list); if (tmp->index == order && memclass(page_zone(tmp), classzone)) { list_del(entry); - current->nr_local_pages--; - set_page_count(tmp, 1); page = tmp; - - if (PagePrivate(page)) - BUG(); - if (page->mapping) - BUG(); - if (PageLocked(page)) - BUG(); - if (PageLRU(page)) - BUG(); - if (PageActive(page)) - BUG(); - if (PageDirty(page)) - BUG(); - if (PageWriteback(page)) - BUG(); - + current->nr_local_pages--; + prep_new_page(page); break; } } while ((entry = entry->next) != local_pages); diff --git a/mm/swap_state.c b/mm/swap_state.c index c831b5193865..7386e49ccf38 100644 --- a/mm/swap_state.c +++ b/mm/swap_state.c @@ -153,9 +153,13 @@ int move_to_swap_cache(struct page *page, swp_entry_t entry) /* Add it to the swap cache */ *pslot = page; - page->flags &= ~(1 << PG_uptodate | 1 << PG_error - | 1 << PG_referenced | 1 << PG_arch_1 - | 1 << PG_checked); + /* + * This code used to clear PG_uptodate, PG_error, PG_arch1, + * PG_referenced and PG_checked. What _should_ it clear? + */ + ClearPageUptodate(page); + ClearPageReferenced(page); + SetPageLocked(page); ClearPageDirty(page); ___add_to_page_cache(page, &swapper_space, entry.val); @@ -198,9 +202,14 @@ int move_from_swap_cache(struct page *page, unsigned long index, __delete_from_swap_cache(page); *pslot = page; - page->flags &= ~(1 << PG_uptodate | 1 << PG_error | - 1 << PG_referenced | 1 << PG_arch_1 | - 1 << PG_checked); + + /* + * This code used to clear PG_uptodate, PG_error, PG_referenced, + * PG_arch_1 and PG_checked. It's not really clear why. + */ + ClearPageUptodate(page); + ClearPageReferenced(page); + /* * ___add_to_page_cache puts the page on ->clean_pages, * but it's dirty. If it's on ->clean_pages, it will basically diff --git a/mm/vmscan.c b/mm/vmscan.c index 56757cc67ee5..da823781fc06 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -64,6 +64,8 @@ swap_out_add_to_swap_cache(struct page *page, swp_entry_t entry) current->flags &= ~PF_MEMALLOC; current->flags |= PF_RADIX_TREE; + ClearPageUptodate(page); /* why? */ + ClearPageReferenced(page); /* why? */ ret = add_to_swap_cache(page, entry); current->flags = flags; return ret; |
