1 /* 2 * Copyright (c) 2000-2005 Silicon Graphics, Inc. 3 * All Rights Reserved. 4 * 5 * This program is free software; you can redistribute it and/or 6 * modify it under the terms of the GNU General Public License as 7 * published by the Free Software Foundation. 8 * 9 * This program is distributed in the hope that it would be useful, 10 * but WITHOUT ANY WARRANTY; without even the implied warranty of 11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 12 * GNU General Public License for more details. 13 * 14 * You should have received a copy of the GNU General Public License 15 * along with this program; if not, write the Free Software Foundation, 16 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA 17 */ 18 #ifndef __XFS_SUPPORT_KMEM_H__ 19 #define __XFS_SUPPORT_KMEM_H__ 20 21 #include <linux/slab.h> 22 #include <linux/sched.h> 23 #include <linux/mm.h> 24 #include <linux/vmalloc.h> 25 26 /* 27 * General memory allocation interfaces 28 */ 29 30 typedef unsigned __bitwise xfs_km_flags_t; 31 #define KM_SLEEP ((__force xfs_km_flags_t)0x0001u) 32 #define KM_NOSLEEP ((__force xfs_km_flags_t)0x0002u) 33 #define KM_NOFS ((__force xfs_km_flags_t)0x0004u) 34 #define KM_MAYFAIL ((__force xfs_km_flags_t)0x0008u) 35 #define KM_ZERO ((__force xfs_km_flags_t)0x0010u) 36 37 /* 38 * We use a special process flag to avoid recursive callbacks into 39 * the filesystem during transactions. We will also issue our own 40 * warnings, so we explicitly skip any generic ones (silly of us). 41 */ 42 static inline gfp_t 43 kmem_flags_convert(xfs_km_flags_t flags) 44 { 45 gfp_t lflags; 46 47 BUG_ON(flags & ~(KM_SLEEP|KM_NOSLEEP|KM_NOFS|KM_MAYFAIL|KM_ZERO)); 48 49 if (flags & KM_NOSLEEP) { 50 lflags = GFP_ATOMIC | __GFP_NOWARN; 51 } else { 52 lflags = GFP_KERNEL | __GFP_NOWARN; 53 if (flags & KM_NOFS) 54 lflags &= ~__GFP_FS; 55 } 56 57 /* 58 * Default page/slab allocator behavior is to retry for ever 59 * for small allocations. We can override this behavior by using 60 * __GFP_RETRY_MAYFAIL which will tell the allocator to retry as long 61 * as it is feasible but rather fail than retry forever for all 62 * request sizes. 63 */ 64 if (flags & KM_MAYFAIL) 65 lflags |= __GFP_RETRY_MAYFAIL; 66 67 if (flags & KM_ZERO) 68 lflags |= __GFP_ZERO; 69 70 return lflags; 71 } 72 73 extern void *kmem_alloc(size_t, xfs_km_flags_t); 74 extern void *kmem_alloc_large(size_t size, xfs_km_flags_t); 75 extern void *kmem_realloc(const void *, size_t, xfs_km_flags_t); 76 static inline void kmem_free(const void *ptr) 77 { 78 kvfree(ptr); 79 } 80 81 82 static inline void * 83 kmem_zalloc(size_t size, xfs_km_flags_t flags) 84 { 85 return kmem_alloc(size, flags | KM_ZERO); 86 } 87 88 static inline void * 89 kmem_zalloc_large(size_t size, xfs_km_flags_t flags) 90 { 91 return kmem_alloc_large(size, flags | KM_ZERO); 92 } 93 94 /* 95 * Zone interfaces 96 */ 97 98 #define KM_ZONE_HWALIGN SLAB_HWCACHE_ALIGN 99 #define KM_ZONE_RECLAIM SLAB_RECLAIM_ACCOUNT 100 #define KM_ZONE_SPREAD SLAB_MEM_SPREAD 101 #define KM_ZONE_ACCOUNT SLAB_ACCOUNT 102 103 #define kmem_zone kmem_cache 104 #define kmem_zone_t struct kmem_cache 105 106 static inline kmem_zone_t * 107 kmem_zone_init(int size, char *zone_name) 108 { 109 return kmem_cache_create(zone_name, size, 0, 0, NULL); 110 } 111 112 static inline kmem_zone_t * 113 kmem_zone_init_flags(int size, char *zone_name, slab_flags_t flags, 114 void (*construct)(void *)) 115 { 116 return kmem_cache_create(zone_name, size, 0, flags, construct); 117 } 118 119 static inline void 120 kmem_zone_free(kmem_zone_t *zone, void *ptr) 121 { 122 kmem_cache_free(zone, ptr); 123 } 124 125 static inline void 126 kmem_zone_destroy(kmem_zone_t *zone) 127 { 128 kmem_cache_destroy(zone); 129 } 130 131 extern void *kmem_zone_alloc(kmem_zone_t *, xfs_km_flags_t); 132 133 static inline void * 134 kmem_zone_zalloc(kmem_zone_t *zone, xfs_km_flags_t flags) 135 { 136 return kmem_zone_alloc(zone, flags | KM_ZERO); 137 } 138 139 #endif /* __XFS_SUPPORT_KMEM_H__ */ 140