diff options
author | Mingming Cao <cmm@us.ibm.com> | 2007-10-17 02:38:25 +0400 |
---|---|---|
committer | Theodore Ts'o <tytso@mit.edu> | 2007-10-18 02:49:56 +0400 |
commit | c089d490dfbf53bc0893dc9ef57cf3ee6448314d (patch) | |
tree | 8faffea3bdcfdd48ce175ac92d5088ced4f1c969 /include/linux/jbd.h | |
parent | d85714d81cc0408daddb68c10f7fd69eafe7c213 (diff) | |
download | linux-c089d490dfbf53bc0893dc9ef57cf3ee6448314d.tar.xz |
JBD: JBD slab allocation cleanups
JBD: Replace slab allocations with page allocations
JBD allocate memory for committed_data and frozen_data from slab. However
JBD should not pass slab pages down to the block layer. Use page allocator pages instead. This will also prepare JBD for the large blocksize patchset.
Signed-off-by: Christoph Lameter <clameter@sgi.com>
Signed-off-by: Mingming Cao <cmm@us.ibm.com>
Diffstat (limited to 'include/linux/jbd.h')
-rw-r--r-- | include/linux/jbd.h | 13 |
1 files changed, 10 insertions, 3 deletions
diff --git a/include/linux/jbd.h b/include/linux/jbd.h index 72f522372924..1db3b684f557 100644 --- a/include/linux/jbd.h +++ b/include/linux/jbd.h @@ -73,14 +73,21 @@ extern int journal_enable_debug; #endif extern void * __jbd_kmalloc (const char *where, size_t size, gfp_t flags, int retry); -extern void * jbd_slab_alloc(size_t size, gfp_t flags); -extern void jbd_slab_free(void *ptr, size_t size); - #define jbd_kmalloc(size, flags) \ __jbd_kmalloc(__FUNCTION__, (size), (flags), journal_oom_retry) #define jbd_rep_kmalloc(size, flags) \ __jbd_kmalloc(__FUNCTION__, (size), (flags), 1) +static inline void *jbd_alloc(size_t size, gfp_t flags) +{ + return (void *)__get_free_pages(flags, get_order(size)); +} + +static inline void jbd_free(void *ptr, size_t size) +{ + free_pages((unsigned long)ptr, get_order(size)); +}; + #define JFS_MIN_JOURNAL_BLOCKS 1024 #ifdef __KERNEL__ |