/openbmc/linux/include/linux/ |
H A D | gfp.h | 235 warn_if_node_offline(nid, gfp_mask); in __alloc_pages_node() 260 return __alloc_pages_node(nid, gfp_mask, order); in alloc_pages_node() 280 #define alloc_page(gfp_mask) alloc_pages(gfp_mask, 0) argument 296 #define __get_free_page(gfp_mask) \ argument 297 __get_free_pages((gfp_mask), 0) 299 #define __get_dma_pages(gfp_mask, order) \ argument 300 __get_free_pages((gfp_mask) | GFP_DMA, (order)) 308 unsigned int fragsz, gfp_t gfp_mask, 312 unsigned int fragsz, gfp_t gfp_mask) in page_frag_alloc() argument 339 bool gfp_pfmemalloc_allowed(gfp_t gfp_mask); [all …]
|
H A D | mempool.h | 13 typedef void * (mempool_alloc_t)(gfp_t gfp_mask, void *pool_data); 41 gfp_t gfp_mask, int node_id); 49 gfp_t gfp_mask, int nid); 53 extern void *mempool_alloc(mempool_t *pool, gfp_t gfp_mask) __malloc; 61 void *mempool_alloc_slab(gfp_t gfp_mask, void *pool_data); 82 void *mempool_kmalloc(gfp_t gfp_mask, void *pool_data); 101 void *mempool_alloc_pages(gfp_t gfp_mask, void *pool_data);
|
H A D | cpuset.h | 85 extern bool cpuset_node_allowed(int node, gfp_t gfp_mask); 87 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed() argument 89 return cpuset_node_allowed(zone_to_nid(z), gfp_mask); in __cpuset_zone_allowed() 92 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed() argument 95 return __cpuset_zone_allowed(z, gfp_mask); in cpuset_zone_allowed() 223 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed() argument 228 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed() argument
|
H A D | blk-crypto.h | 84 gfp_t gfp_mask); 115 int __bio_crypt_clone(struct bio *dst, struct bio *src, gfp_t gfp_mask); 128 gfp_t gfp_mask) in bio_crypt_clone() argument 131 return __bio_crypt_clone(dst, src, gfp_mask); in bio_crypt_clone()
|
H A D | page_owner.h | 13 unsigned short order, gfp_t gfp_mask); 28 unsigned short order, gfp_t gfp_mask) in set_page_owner() argument 31 __set_page_owner(page, order, gfp_mask); in set_page_owner() 59 unsigned int order, gfp_t gfp_mask) in set_page_owner() argument
|
/openbmc/linux/mm/ |
H A D | mempool.c | 187 gfp_t gfp_mask, int node_id) in mempool_init_node() argument 197 gfp_mask, node_id); in mempool_init_node() 268 gfp_t gfp_mask, int node_id) in mempool_create_node() argument 277 gfp_mask, node_id)) { in mempool_create_node() 387 VM_WARN_ON_ONCE(gfp_mask & __GFP_ZERO); in mempool_alloc() 388 might_alloc(gfp_mask); in mempool_alloc() 420 if (gfp_temp != gfp_mask) { in mempool_alloc() 422 gfp_temp = gfp_mask; in mempool_alloc() 518 return kmem_cache_alloc(mem, gfp_mask); in mempool_alloc_slab() 536 return kmalloc(size, gfp_mask); in mempool_kmalloc() [all …]
|
H A D | page_alloc.c | 3130 gfp_mask)) { in get_page_from_freelist() 3247 current->comm, &vaf, gfp_mask, &gfp_mask, in warn_alloc() 3285 .gfp_mask = gfp_mask, in __alloc_pages_may_oom() 3348 WARN_ON_ONCE_GFP(gfp_mask & __GFP_NOFAIL, gfp_mask)) { in __alloc_pages_may_oom() 3566 gfp_mask = current_gfp_context(gfp_mask); in fs_reclaim_acquire() 3583 gfp_mask = current_gfp_context(gfp_mask); in fs_reclaim_release() 4204 might_alloc(gfp_mask); in prepare_alloc_pages() 4551 gfp_t gfp = gfp_mask; in __page_frag_cache_refill() 6126 .gfp_mask = current_gfp_context(gfp_mask), in alloc_contig_range() 6244 gfp_mask); in __alloc_contig_pages() [all …]
|
H A D | fail_page_alloc.c | 24 bool __should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) in __should_fail_alloc_page() argument 30 if (gfp_mask & __GFP_NOFAIL) in __should_fail_alloc_page() 32 if (fail_page_alloc.ignore_gfp_highmem && (gfp_mask & __GFP_HIGHMEM)) in __should_fail_alloc_page() 35 (gfp_mask & __GFP_DIRECT_RECLAIM)) in __should_fail_alloc_page() 39 if (gfp_mask & __GFP_NOWARN) in __should_fail_alloc_page()
|
H A D | page_owner.c | 27 gfp_t gfp_mask; member 162 unsigned short order, gfp_t gfp_mask) in __set_page_owner_handle() argument 172 page_owner->gfp_mask = gfp_mask; in __set_page_owner_handle() 187 gfp_t gfp_mask) in __set_page_owner() argument 192 handle = save_stack(gfp_mask); in __set_page_owner() 250 new_page_owner->gfp_mask = old_page_owner->gfp_mask; in __folio_copy_owner() 413 &page_owner->gfp_mask, page_owner->pid, in print_page_owner() 460 gfp_t gfp_mask; in __dump_page_owner() local 469 gfp_mask = page_owner->gfp_mask; in __dump_page_owner() 470 mt = gfp_migratetype(gfp_mask); in __dump_page_owner() [all …]
|
H A D | swap.h | 47 struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, 51 struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, 84 gfp_t gfp_mask, struct vm_fault *vmf) in swap_cluster_readahead() argument 89 static inline struct page *swapin_readahead(swp_entry_t swp, gfp_t gfp_mask, in swapin_readahead() argument 128 gfp_t gfp_mask, void **shadowp) in add_to_swap_cache() argument
|
H A D | swap_state.c | 456 folio = vma_alloc_folio(gfp_mask, 0, vma, addr, false); in __read_swap_cache_async() 488 if (mem_cgroup_swapin_charge_folio(folio, NULL, gfp_mask, entry)) in __read_swap_cache_async() 527 struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, in read_swap_cache_async() argument 532 struct page *retpage = __read_swap_cache_async(entry, gfp_mask, in read_swap_cache_async() 652 gfp_mask, vma, addr, &page_allocated); in swap_cluster_readahead() 670 return read_swap_cache_async(entry, gfp_mask, vma, addr, NULL); in swap_cluster_readahead() 817 page = __read_swap_cache_async(entry, gfp_mask, vma, in swap_vma_readahead() 837 return read_swap_cache_async(fentry, gfp_mask, vma, vmf->address, in swap_vma_readahead() 853 struct page *swapin_readahead(swp_entry_t entry, gfp_t gfp_mask, in swapin_readahead() argument 857 swap_vma_readahead(entry, gfp_mask, vmf) : in swapin_readahead() [all …]
|
/openbmc/linux/block/ |
H A D | blk-lib.c | 39 sector_t nr_sects, gfp_t gfp_mask, struct bio **biop) in __blkdev_issue_discard() argument 67 bio = blk_next_bio(bio, bdev, 0, REQ_OP_DISCARD, gfp_mask); in __blkdev_issue_discard() 98 sector_t nr_sects, gfp_t gfp_mask) in blkdev_issue_discard() argument 119 sector_t sector, sector_t nr_sects, gfp_t gfp_mask, in __blkdev_issue_write_zeroes() argument 135 bio = blk_next_bio(bio, bdev, 0, REQ_OP_WRITE_ZEROES, gfp_mask); in __blkdev_issue_write_zeroes() 169 sector_t sector, sector_t nr_sects, gfp_t gfp_mask, in __blkdev_issue_zero_pages() argument 181 REQ_OP_WRITE, gfp_mask); in __blkdev_issue_zero_pages() 219 sector_t nr_sects, gfp_t gfp_mask, struct bio **biop, in __blkdev_issue_zeroout() argument 253 sector_t nr_sects, gfp_t gfp_mask, unsigned flags) in blkdev_issue_zeroout() argument 270 gfp_mask, &bio, flags); in blkdev_issue_zeroout() [all …]
|
H A D | blk-map.c | 22 gfp_t gfp_mask) in bio_alloc_map_data() argument 132 struct iov_iter *iter, gfp_t gfp_mask) in bio_copy_user_iov() argument 157 bio = bio_kmalloc(nr_pages, gfp_mask); in bio_copy_user_iov() 254 unsigned int nr_vecs, gfp_t gfp_mask) in blk_rq_map_bio_alloc() argument 264 bio = bio_kmalloc(nr_vecs, gfp_mask); in blk_rq_map_bio_alloc() 273 gfp_t gfp_mask) in bio_map_user_iov() argument 390 unsigned int len, gfp_t gfp_mask) in bio_map_kern() argument 401 bio = bio_kmalloc(nr_pages, gfp_mask); in bio_map_kern() 491 bio = bio_kmalloc(nr_pages, gfp_mask); in bio_copy_kern() 725 gfp_mask); in blk_rq_map_user_io() [all …]
|
H A D | blk-crypto.c | 92 const u64 dun[BLK_CRYPTO_DUN_ARRAY_SIZE], gfp_t gfp_mask) in bio_crypt_set_ctx() argument 100 WARN_ON_ONCE(!(gfp_mask & __GFP_DIRECT_RECLAIM)); in bio_crypt_set_ctx() 102 bc = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); in bio_crypt_set_ctx() 116 int __bio_crypt_clone(struct bio *dst, struct bio *src, gfp_t gfp_mask) in __bio_crypt_clone() argument 118 dst->bi_crypt_context = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); in __bio_crypt_clone() 304 gfp_t gfp_mask) in __blk_crypto_rq_bio_prep() argument 307 rq->crypt_ctx = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); in __blk_crypto_rq_bio_prep()
|
/openbmc/linux/fs/btrfs/ |
H A D | ulist.h | 48 struct ulist *ulist_alloc(gfp_t gfp_mask); 50 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask); 52 u64 *old_aux, gfp_t gfp_mask); 57 void **old_aux, gfp_t gfp_mask) in ulist_add_merge_ptr() argument 61 int ret = ulist_add_merge(ulist, val, (uintptr_t)aux, &old64, gfp_mask); in ulist_add_merge_ptr() 65 return ulist_add_merge(ulist, val, (u64)aux, (u64 *)old_aux, gfp_mask); in ulist_add_merge_ptr()
|
H A D | ulist.c | 97 struct ulist *ulist_alloc(gfp_t gfp_mask) in ulist_alloc() argument 99 struct ulist *ulist = kmalloc(sizeof(*ulist), gfp_mask); in ulist_alloc() 193 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask) in ulist_add() argument 195 return ulist_add_merge(ulist, val, aux, NULL, gfp_mask); in ulist_add() 199 u64 *old_aux, gfp_t gfp_mask) in ulist_add_merge() argument 210 node = kmalloc(sizeof(*node), gfp_mask); in ulist_add_merge()
|
/openbmc/linux/fs/nfs/blocklayout/ |
H A D | dev.c | 242 dev = bl_resolve_deviceid(server, v, gfp_mask); in bl_parse_simple() 407 sizeof(struct pnfs_block_dev), gfp_mask); in bl_parse_concat() 413 volumes, v->concat.volumes[i], gfp_mask); in bl_parse_concat() 436 sizeof(struct pnfs_block_dev), gfp_mask); in bl_parse_stripe() 442 volumes, v->stripe.volumes[i], gfp_mask); in bl_parse_stripe() 464 return bl_parse_slice(server, d, volumes, idx, gfp_mask); in bl_parse_deviceid() 470 return bl_parse_scsi(server, d, volumes, idx, gfp_mask); in bl_parse_deviceid() 479 gfp_t gfp_mask) in bl_alloc_deviceid_node() argument 490 scratch = alloc_page(gfp_mask); in bl_alloc_deviceid_node() 503 gfp_mask); in bl_alloc_deviceid_node() [all …]
|
/openbmc/linux/lib/ |
H A D | generic-radix-tree.c | 79 static inline struct genradix_node *genradix_alloc_node(gfp_t gfp_mask) in genradix_alloc_node() argument 83 node = (struct genradix_node *)__get_free_page(gfp_mask|__GFP_ZERO); in genradix_alloc_node() 90 kmemleak_alloc(node, PAGE_SIZE, 1, gfp_mask); in genradix_alloc_node() 105 gfp_t gfp_mask) in __genradix_ptr_alloc() argument 122 new_node = genradix_alloc_node(gfp_mask); in __genradix_ptr_alloc() 147 new_node = genradix_alloc_node(gfp_mask); in __genradix_ptr_alloc() 231 gfp_t gfp_mask) in __genradix_prealloc() argument 236 if (!__genradix_ptr_alloc(radix, offset, gfp_mask)) in __genradix_prealloc()
|
/openbmc/linux/include/linux/sched/ |
H A D | mm.h | 258 extern void fs_reclaim_acquire(gfp_t gfp_mask); 259 extern void fs_reclaim_release(gfp_t gfp_mask); 263 static inline void fs_reclaim_acquire(gfp_t gfp_mask) { } in fs_reclaim_acquire() argument 264 static inline void fs_reclaim_release(gfp_t gfp_mask) { } in fs_reclaim_release() argument 301 static inline void might_alloc(gfp_t gfp_mask) in might_alloc() argument 303 fs_reclaim_acquire(gfp_mask); in might_alloc() 304 fs_reclaim_release(gfp_mask); in might_alloc() 306 might_sleep_if(gfpflags_allow_blocking(gfp_mask)); in might_alloc()
|
/openbmc/linux/net/sunrpc/auth_gss/ |
H A D | gss_krb5_keys.c | 152 const struct xdr_netobj *in_constant, gfp_t gfp_mask) in krb5_DK() argument 174 inblockdata = kmalloc(blocksize, gfp_mask); in krb5_DK() 178 outblockdata = kmalloc(blocksize, gfp_mask); in krb5_DK() 271 gfp_t gfp_mask) in krb5_derive_key_v2() argument 277 inblock.data = kmalloc(inblock.len, gfp_mask); in krb5_derive_key_v2() 281 ret = krb5_DK(gk5e, inkey, inblock.data, label, gfp_mask); in krb5_derive_key_v2() 372 gfp_t gfp_mask) in krb5_kdf_feedback_cmac() argument 401 step.data = kzalloc(step.len, gfp_mask); in krb5_kdf_feedback_cmac() 406 DR.data = kmalloc(DR.len, gfp_mask); in krb5_kdf_feedback_cmac() 504 gfp_t gfp_mask) in krb5_kdf_hmac_sha2() argument [all …]
|
H A D | gss_krb5_mech.c | 313 KEY_USAGE_SEED_ENCRYPTION, gfp_mask)) in gss_krb5_import_ctx_v2() 329 KEY_USAGE_SEED_ENCRYPTION, gfp_mask)) in gss_krb5_import_ctx_v2() 346 KEY_USAGE_SEED_CHECKSUM, gfp_mask)) in gss_krb5_import_ctx_v2() 354 KEY_USAGE_SEED_CHECKSUM, gfp_mask)) in gss_krb5_import_ctx_v2() 363 KEY_USAGE_SEED_INTEGRITY, gfp_mask)) in gss_krb5_import_ctx_v2() 371 KEY_USAGE_SEED_INTEGRITY, gfp_mask)) in gss_krb5_import_ctx_v2() 396 gfp_t gfp_mask) in gss_import_v2_context() argument 447 gss_kerberos_mech.gm_oid.len, gfp_mask); in gss_import_v2_context() 454 ret = gss_krb5_import_ctx_v2(ctx, gfp_mask); in gss_import_v2_context() 470 time64_t *endtime, gfp_t gfp_mask) in gss_krb5_import_sec_context() argument [all …]
|
H A D | gss_krb5_internal.h | 40 gfp_t gfp_mask); 110 gfp_t gfp_mask); 116 gfp_t gfp_mask); 122 gfp_t gfp_mask); 141 u32 usage, u8 seed, gfp_t gfp_mask) in krb5_derive_key() argument 153 return gk5e->derive_key(gk5e, inkey, outkey, &label, gfp_mask); in krb5_derive_key()
|
/openbmc/linux/fs/ntfs/ |
H A D | malloc.h | 28 static inline void *__ntfs_malloc(unsigned long size, gfp_t gfp_mask) in __ntfs_malloc() argument 33 return kmalloc(PAGE_SIZE, gfp_mask & ~__GFP_HIGHMEM); in __ntfs_malloc() 37 return __vmalloc(size, gfp_mask); in __ntfs_malloc()
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx4/ |
H A D | icm.c | 99 gfp_t gfp_mask, int node) in mlx4_alloc_icm_pages() argument 103 page = alloc_pages_node(node, gfp_mask, order); in mlx4_alloc_icm_pages() 105 page = alloc_pages(gfp_mask, order); in mlx4_alloc_icm_pages() 115 int order, gfp_t gfp_mask) in mlx4_alloc_icm_coherent() argument 118 &buf->dma_addr, gfp_mask); in mlx4_alloc_icm_coherent() 133 gfp_t gfp_mask, int coherent) in mlx4_alloc_icm() argument 142 BUG_ON(coherent && (gfp_mask & __GFP_HIGHMEM)); in mlx4_alloc_icm() 145 gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN), in mlx4_alloc_icm() 162 gfp_mask & ~(__GFP_HIGHMEM | in mlx4_alloc_icm() 167 gfp_mask & ~(__GFP_HIGHMEM | in mlx4_alloc_icm() [all …]
|
/openbmc/linux/drivers/connector/ |
H A D | connector.c | 62 gfp_t gfp_mask, in cn_netlink_send_mult() argument 98 skb = nlmsg_new(size, gfp_mask); in cn_netlink_send_mult() 116 gfp_mask, filter, in cn_netlink_send_mult() 119 !gfpflags_allow_blocking(gfp_mask)); in cn_netlink_send_mult() 125 gfp_t gfp_mask) in cn_netlink_send() argument 127 return cn_netlink_send_mult(msg, msg->len, portid, __group, gfp_mask, in cn_netlink_send()
|