09cbfeaf1a
PAGE_CACHE_{SIZE,SHIFT,MASK,ALIGN} macros were introduced *long* time ago with promise that one day it will be possible to implement page cache with bigger chunks than PAGE_SIZE. This promise never materialized. And unlikely will. We have many places where PAGE_CACHE_SIZE assumed to be equal to PAGE_SIZE. And it's constant source of confusion on whether PAGE_CACHE_* or PAGE_* constant should be used in a particular case, especially on the border between fs and mm. Global switching to PAGE_CACHE_SIZE != PAGE_SIZE would cause to much breakage to be doable. Let's stop pretending that pages in page cache are special. They are not. The changes are pretty straight-forward: - <foo> << (PAGE_CACHE_SHIFT - PAGE_SHIFT) -> <foo>; - <foo> >> (PAGE_CACHE_SHIFT - PAGE_SHIFT) -> <foo>; - PAGE_CACHE_{SIZE,SHIFT,MASK,ALIGN} -> PAGE_{SIZE,SHIFT,MASK,ALIGN}; - page_cache_get() -> get_page(); - page_cache_release() -> put_page(); This patch contains automated changes generated with coccinelle using script below. For some reason, coccinelle doesn't patch header files. I've called spatch for them manually. The only adjustment after coccinelle is revert of changes to PAGE_CAHCE_ALIGN definition: we are going to drop it later. There are few places in the code where coccinelle didn't reach. I'll fix them manually in a separate patch. Comments and documentation also will be addressed with the separate patch. virtual patch @@ expression E; @@ - E << (PAGE_CACHE_SHIFT - PAGE_SHIFT) + E @@ expression E; @@ - E >> (PAGE_CACHE_SHIFT - PAGE_SHIFT) + E @@ @@ - PAGE_CACHE_SHIFT + PAGE_SHIFT @@ @@ - PAGE_CACHE_SIZE + PAGE_SIZE @@ @@ - PAGE_CACHE_MASK + PAGE_MASK @@ expression E; @@ - PAGE_CACHE_ALIGN(E) + PAGE_ALIGN(E) @@ expression E; @@ - page_cache_get(E) + get_page(E) @@ expression E; @@ - page_cache_release(E) + put_page(E) Signed-off-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com> Acked-by: Michal Hocko <mhocko@suse.com> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
150 lines
3.6 KiB
C
150 lines
3.6 KiB
C
#include <linux/module.h>
|
|
#include <linux/gfp.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/pagemap.h>
|
|
#include <linux/highmem.h>
|
|
#include <linux/ceph/pagelist.h>
|
|
|
|
static void ceph_pagelist_unmap_tail(struct ceph_pagelist *pl)
|
|
{
|
|
if (pl->mapped_tail) {
|
|
struct page *page = list_entry(pl->head.prev, struct page, lru);
|
|
kunmap(page);
|
|
pl->mapped_tail = NULL;
|
|
}
|
|
}
|
|
|
|
void ceph_pagelist_release(struct ceph_pagelist *pl)
|
|
{
|
|
if (!atomic_dec_and_test(&pl->refcnt))
|
|
return;
|
|
ceph_pagelist_unmap_tail(pl);
|
|
while (!list_empty(&pl->head)) {
|
|
struct page *page = list_first_entry(&pl->head, struct page,
|
|
lru);
|
|
list_del(&page->lru);
|
|
__free_page(page);
|
|
}
|
|
ceph_pagelist_free_reserve(pl);
|
|
kfree(pl);
|
|
}
|
|
EXPORT_SYMBOL(ceph_pagelist_release);
|
|
|
|
static int ceph_pagelist_addpage(struct ceph_pagelist *pl)
|
|
{
|
|
struct page *page;
|
|
|
|
if (!pl->num_pages_free) {
|
|
page = __page_cache_alloc(GFP_NOFS);
|
|
} else {
|
|
page = list_first_entry(&pl->free_list, struct page, lru);
|
|
list_del(&page->lru);
|
|
--pl->num_pages_free;
|
|
}
|
|
if (!page)
|
|
return -ENOMEM;
|
|
pl->room += PAGE_SIZE;
|
|
ceph_pagelist_unmap_tail(pl);
|
|
list_add_tail(&page->lru, &pl->head);
|
|
pl->mapped_tail = kmap(page);
|
|
return 0;
|
|
}
|
|
|
|
int ceph_pagelist_append(struct ceph_pagelist *pl, const void *buf, size_t len)
|
|
{
|
|
while (pl->room < len) {
|
|
size_t bit = pl->room;
|
|
int ret;
|
|
|
|
memcpy(pl->mapped_tail + (pl->length & ~PAGE_MASK),
|
|
buf, bit);
|
|
pl->length += bit;
|
|
pl->room -= bit;
|
|
buf += bit;
|
|
len -= bit;
|
|
ret = ceph_pagelist_addpage(pl);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
memcpy(pl->mapped_tail + (pl->length & ~PAGE_MASK), buf, len);
|
|
pl->length += len;
|
|
pl->room -= len;
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(ceph_pagelist_append);
|
|
|
|
/* Allocate enough pages for a pagelist to append the given amount
|
|
* of data without without allocating.
|
|
* Returns: 0 on success, -ENOMEM on error.
|
|
*/
|
|
int ceph_pagelist_reserve(struct ceph_pagelist *pl, size_t space)
|
|
{
|
|
if (space <= pl->room)
|
|
return 0;
|
|
space -= pl->room;
|
|
space = (space + PAGE_SIZE - 1) >> PAGE_SHIFT; /* conv to num pages */
|
|
|
|
while (space > pl->num_pages_free) {
|
|
struct page *page = __page_cache_alloc(GFP_NOFS);
|
|
if (!page)
|
|
return -ENOMEM;
|
|
list_add_tail(&page->lru, &pl->free_list);
|
|
++pl->num_pages_free;
|
|
}
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(ceph_pagelist_reserve);
|
|
|
|
/* Free any pages that have been preallocated. */
|
|
int ceph_pagelist_free_reserve(struct ceph_pagelist *pl)
|
|
{
|
|
while (!list_empty(&pl->free_list)) {
|
|
struct page *page = list_first_entry(&pl->free_list,
|
|
struct page, lru);
|
|
list_del(&page->lru);
|
|
__free_page(page);
|
|
--pl->num_pages_free;
|
|
}
|
|
BUG_ON(pl->num_pages_free);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(ceph_pagelist_free_reserve);
|
|
|
|
/* Create a truncation point. */
|
|
void ceph_pagelist_set_cursor(struct ceph_pagelist *pl,
|
|
struct ceph_pagelist_cursor *c)
|
|
{
|
|
c->pl = pl;
|
|
c->page_lru = pl->head.prev;
|
|
c->room = pl->room;
|
|
}
|
|
EXPORT_SYMBOL(ceph_pagelist_set_cursor);
|
|
|
|
/* Truncate a pagelist to the given point. Move extra pages to reserve.
|
|
* This won't sleep.
|
|
* Returns: 0 on success,
|
|
* -EINVAL if the pagelist doesn't match the trunc point pagelist
|
|
*/
|
|
int ceph_pagelist_truncate(struct ceph_pagelist *pl,
|
|
struct ceph_pagelist_cursor *c)
|
|
{
|
|
struct page *page;
|
|
|
|
if (pl != c->pl)
|
|
return -EINVAL;
|
|
ceph_pagelist_unmap_tail(pl);
|
|
while (pl->head.prev != c->page_lru) {
|
|
page = list_entry(pl->head.prev, struct page, lru);
|
|
/* move from pagelist to reserve */
|
|
list_move_tail(&page->lru, &pl->free_list);
|
|
++pl->num_pages_free;
|
|
}
|
|
pl->room = c->room;
|
|
if (!list_empty(&pl->head)) {
|
|
page = list_entry(pl->head.prev, struct page, lru);
|
|
pl->mapped_tail = kmap(page);
|
|
}
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(ceph_pagelist_truncate);
|