Home
last modified time | relevance | path

Searched refs:order (Results 1 – 25 of 2465) sorted by relevance

12345678910>>...99

/openbmc/linux/scripts/atomic/
H A Dgen-atomic-fallback.sh16 local order="$1"; shift
35 local order="$1"; shift
37 local tmpl_order=${order#_}
39 gen_template_fallback "${tmpl}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "$@"
49 local order="$1"; shift
51 local tmpl="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")"
52 gen_template_fallback "${tmpl}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "$@"
62 local order="$1"; shift
66 local atomicname="${atomic}_${pfx}${name}${sfx}${order}"
69 local template="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")"
[all …]
H A Dgen-atomic-instrumented.sh37 local order="$1"; shift
39 if [ "${order}" = "_release" ]; then
41 elif [ -z "${order}" ] && ! meta_in "$meta" "slv"; then
59 local order="$1"; shift
63 local atomicname="${atomic}_${pfx}${name}${sfx}${order}"
67 local checks="$(gen_params_checks "${meta}" "${order}" "$@")"
71 gen_kerneldoc "" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "${atomic}" "${int}" "$@"
88 local order="$1"; shift
92 case "$order" in
110 raw_${xchg}${order}(__ai_ptr, __ai_oldp, __VA_ARGS__); \\
[all …]
/openbmc/linux/arch/arm64/kvm/hyp/nvhe/
H A Dpage_alloc.c35 unsigned short order) in __find_buddy_nocheck() argument
39 addr ^= (PAGE_SIZE << order); in __find_buddy_nocheck()
54 unsigned short order) in __find_buddy_avail() argument
56 struct hyp_page *buddy = __find_buddy_nocheck(pool, p, order); in __find_buddy_avail()
58 if (!buddy || buddy->order != order || buddy->refcount) in __find_buddy_avail()
97 unsigned short order = p->order; in __hyp_attach_page() local
100 memset(hyp_page_to_virt(p), 0, PAGE_SIZE << p->order); in __hyp_attach_page()
112 p->order = HYP_NO_ORDER; in __hyp_attach_page()
113 for (; (order + 1) <= pool->max_order; order++) { in __hyp_attach_page()
114 buddy = __find_buddy_avail(pool, p, order); in __hyp_attach_page()
[all …]
/openbmc/linux/include/trace/events/
H A Dcompaction.h168 int order,
172 TP_ARGS(order, gfp_mask, prio),
175 __field(int, order)
181 __entry->order = order;
187 __entry->order,
195 int order,
198 TP_ARGS(zone, order, ret),
203 __field(int, order)
210 __entry->order = order;
217 __entry->order,
[all …]
H A Dvmscan.h68 TP_PROTO(int nid, int zid, int order),
70 TP_ARGS(nid, zid, order),
75 __field( int, order )
81 __entry->order = order;
86 __entry->order)
91 TP_PROTO(int nid, int zid, int order, gfp_t gfp_flags),
93 TP_ARGS(nid, zid, order, gfp_flags),
98 __field( int, order )
105 __entry->order = order;
111 __entry->order,
[all …]
H A Dkmem.h138 TP_PROTO(struct page *page, unsigned int order),
140 TP_ARGS(page, order),
144 __field( unsigned int, order )
149 __entry->order = order;
155 __entry->order)
179 TP_PROTO(struct page *page, unsigned int order,
182 TP_ARGS(page, order, gfp_flags, migratetype),
186 __field( unsigned int, order )
193 __entry->order = order;
201 __entry->order,
[all …]
/openbmc/linux/mm/
H A Dpage_alloc.c229 static void __free_pages_ok(struct page *page, unsigned int order,
303 static bool page_contains_unaccepted(struct page *page, unsigned int order);
304 static void accept_page(struct page *page, unsigned int order);
305 static bool cond_accept_memory(struct zone *zone, unsigned int order);
331 _deferred_grow_zone(struct zone *zone, unsigned int order) in _deferred_grow_zone() argument
333 return deferred_grow_zone(zone, order); in _deferred_grow_zone()
520 static inline unsigned int order_to_pindex(int migratetype, int order) in order_to_pindex() argument
525 if (order > PAGE_ALLOC_COSTLY_ORDER) { in order_to_pindex()
526 VM_BUG_ON(order != pageblock_order); in order_to_pindex()
533 VM_BUG_ON(order > PAGE_ALLOC_COSTLY_ORDE in order_to_pindex()
541 int order = pindex / MIGRATE_PCPTYPES; pindex_to_order() local
553 pcp_allowed_order(unsigned int order) pcp_allowed_order() argument
564 free_the_page(struct page * page,unsigned int order) free_the_page() argument
584 prep_compound_page(struct page * page,unsigned int order) prep_compound_page() argument
608 set_buddy_order(struct page * page,unsigned int order) set_buddy_order() argument
627 compaction_capture(struct capture_control * capc,struct page * page,int order,int migratetype) compaction_capture() argument
658 compaction_capture(struct capture_control * capc,struct page * page,int order,int migratetype) compaction_capture() argument
666 add_to_free_list(struct page * page,struct zone * zone,unsigned int order,int migratetype) add_to_free_list() argument
676 add_to_free_list_tail(struct page * page,struct zone * zone,unsigned int order,int migratetype) add_to_free_list_tail() argument
690 move_to_free_list(struct page * page,struct zone * zone,unsigned int order,int migratetype) move_to_free_list() argument
698 del_page_from_free_list(struct page * page,struct zone * zone,unsigned int order) del_page_from_free_list() argument
727 buddy_merge_likely(unsigned long pfn,unsigned long buddy_pfn,struct page * page,unsigned int order) buddy_merge_likely() argument
768 __free_one_page(struct page * page,unsigned long pfn,struct zone * zone,unsigned int order,int migratetype,fpi_t fpi_flags) __free_one_page() argument
861 split_free_page(struct page * free_page,unsigned int order,unsigned long split_pfn_offset) split_free_page() argument
1080 free_pages_prepare(struct page * page,unsigned int order,fpi_t fpi_flags) free_pages_prepare() argument
1207 unsigned int order; free_pcppages_bulk() local
1263 free_one_page(struct zone * zone,struct page * page,unsigned long pfn,unsigned int order,int migratetype,fpi_t fpi_flags) free_one_page() argument
1277 __free_pages_ok(struct page * page,unsigned int order,fpi_t fpi_flags) __free_pages_ok() argument
1306 __free_pages_core(struct page * page,unsigned int order) __free_pages_core() argument
1457 check_new_pages(struct page * page,unsigned int order) check_new_pages() argument
1499 post_alloc_hook(struct page * page,unsigned int order,gfp_t gfp_flags) post_alloc_hook() argument
1559 prep_new_page(struct page * page,unsigned int order,gfp_t gfp_flags,unsigned int alloc_flags) prep_new_page() argument
1584 __rmqueue_smallest(struct zone * zone,unsigned int order,int migratetype) __rmqueue_smallest() argument
1624 __rmqueue_cma_fallback(struct zone * zone,unsigned int order) __rmqueue_cma_fallback() argument
1630 __rmqueue_cma_fallback(struct zone * zone,unsigned int order) __rmqueue_cma_fallback() argument
1644 unsigned int order; move_freepages() local
1720 can_steal_fallback(unsigned int order,int start_mt) can_steal_fallback() argument
1867 find_suitable_fallback(struct free_area * area,unsigned int order,int migratetype,bool only_stealable,bool * can_steal) find_suitable_fallback() argument
1948 int order; unreserve_highatomic_pageblock() local
2023 __rmqueue_fallback(struct zone * zone,int order,int start_migratetype,unsigned int alloc_flags) __rmqueue_fallback() argument
2104 __rmqueue(struct zone * zone,unsigned int order,int migratetype,unsigned int alloc_flags) __rmqueue() argument
2141 rmqueue_bulk(struct zone * zone,unsigned int order,unsigned long count,struct list_head * list,int migratetype,unsigned int alloc_flags) rmqueue_bulk() argument
2333 free_unref_page_prepare(struct page * page,unsigned long pfn,unsigned int order) free_unref_page_prepare() argument
2394 free_unref_page_commit(struct zone * zone,struct per_cpu_pages * pcp,struct page * page,int migratetype,unsigned int order) free_unref_page_commit() argument
2422 free_unref_page(struct page * page,unsigned int order) free_unref_page() argument
2554 split_page(struct page * page,unsigned int order) split_page() argument
2568 __isolate_free_page(struct page * page,unsigned int order) __isolate_free_page() argument
2620 __putback_isolated_page(struct page * page,unsigned int order,int mt) __putback_isolated_page() argument
2660 rmqueue_buddy(struct zone * preferred_zone,struct zone * zone,unsigned int order,unsigned int alloc_flags,int migratetype) rmqueue_buddy() argument
2701 __rmqueue_pcplist(struct zone * zone,unsigned int order,int migratetype,unsigned int alloc_flags,struct per_cpu_pages * pcp,struct list_head * list) __rmqueue_pcplist() argument
2742 rmqueue_pcplist(struct zone * preferred_zone,struct zone * zone,unsigned int order,int migratetype,unsigned int alloc_flags) rmqueue_pcplist() argument
2789 rmqueue(struct zone * preferred_zone,struct zone * zone,unsigned int order,gfp_t gfp_flags,unsigned int alloc_flags,int migratetype) rmqueue() argument
2823 should_fail_alloc_page(gfp_t gfp_mask,unsigned int order) should_fail_alloc_page() argument
2830 __zone_watermark_unusable_free(struct zone * z,unsigned int order,unsigned int alloc_flags) __zone_watermark_unusable_free() argument
2857 __zone_watermark_ok(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx,unsigned int alloc_flags,long free_pages) __zone_watermark_ok() argument
2935 zone_watermark_ok(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx,unsigned int alloc_flags) zone_watermark_ok() argument
2942 zone_watermark_fast(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx,unsigned int alloc_flags,gfp_t gfp_mask) zone_watermark_fast() argument
2987 zone_watermark_ok_safe(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx) zone_watermark_ok_safe() argument
3070 get_page_from_freelist(gfp_t gfp_mask,unsigned int order,int alloc_flags,const struct alloc_context * ac) get_page_from_freelist() argument
3273 __alloc_pages_cpuset_fallback(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac) __alloc_pages_cpuset_fallback() argument
3293 __alloc_pages_may_oom(gfp_t gfp_mask,unsigned int order,const struct alloc_context * ac,unsigned long * did_some_progress) __alloc_pages_may_oom() argument
3388 __alloc_pages_direct_compact(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,enum compact_priority prio,enum compact_result * compact_result) __alloc_pages_direct_compact() argument
3447 should_compact_retry(struct alloc_context * ac,int order,int alloc_flags,enum compact_result compact_result,enum compact_priority * compact_priority,int * compaction_retries) should_compact_retry() argument
3513 __alloc_pages_direct_compact(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,enum compact_priority prio,enum compact_result * compact_result) __alloc_pages_direct_compact() argument
3522 should_compact_retry(struct alloc_context * ac,unsigned int order,int alloc_flags,enum compact_result compact_result,enum compact_priority * compact_priority,int * compaction_retries) should_compact_retry() argument
3634 __perform_reclaim(gfp_t gfp_mask,unsigned int order,const struct alloc_context * ac) __perform_reclaim() argument
3660 __alloc_pages_direct_reclaim(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,unsigned long * did_some_progress) __alloc_pages_direct_reclaim() argument
3693 wake_all_kswapds(unsigned int order,gfp_t gfp_mask,const struct alloc_context * ac) wake_all_kswapds() argument
3713 gfp_to_alloc_flags(gfp_t gfp_mask,unsigned int order) gfp_to_alloc_flags() argument
3814 should_reclaim_retry(gfp_t gfp_mask,unsigned order,struct alloc_context * ac,int alloc_flags,bool did_some_progress,int * no_progress_loops) should_reclaim_retry() argument
3919 __alloc_pages_slowpath(gfp_t gfp_mask,unsigned int order,struct alloc_context * ac) __alloc_pages_slowpath() argument
4197 prepare_alloc_pages(gfp_t gfp_mask,unsigned int order,int preferred_nid,nodemask_t * nodemask,struct alloc_context * ac,gfp_t * alloc_gfp,unsigned int * alloc_flags) prepare_alloc_pages() argument
4413 __alloc_pages(gfp_t gfp,unsigned int order,int preferred_nid,nodemask_t * nodemask) __alloc_pages() argument
4478 __folio_alloc(gfp_t gfp,unsigned int order,int preferred_nid,nodemask_t * nodemask) __folio_alloc() argument
4492 __get_free_pages(gfp_t gfp_mask,unsigned int order) __get_free_pages() argument
4529 __free_pages(struct page * page,unsigned int order) __free_pages() argument
4542 free_pages(unsigned long addr,unsigned int order) free_pages() argument
4678 make_alloc_exact(unsigned long addr,unsigned int order,size_t size) make_alloc_exact() argument
4715 unsigned int order = get_order(size); alloc_pages_exact() local
4740 unsigned int order = get_order(size); alloc_pages_exact_nid() local
6132 int order; alloc_contig_range() local
6417 unsigned int order; __offline_isolated_pages() local
6460 unsigned int order; is_free_buddy_page() local
6517 unsigned int order; take_page_off_buddy() local
6607 page_contains_unaccepted(struct page * page,unsigned int order) page_contains_unaccepted() argument
6615 accept_page(struct page * page,unsigned int order) accept_page() argument
6653 cond_accept_memory(struct zone * zone,unsigned int order) cond_accept_memory() argument
6709 page_contains_unaccepted(struct page * page,unsigned int order) page_contains_unaccepted() argument
6714 accept_page(struct page * page,unsigned int order) accept_page() argument
6718 cond_accept_memory(struct zone * zone,unsigned int order) cond_accept_memory() argument
[all...]
H A Dcompaction.c53 #define block_start_pfn(pfn, order) round_down(pfn, 1UL << (order)) argument
54 #define block_end_pfn(pfn, order) ALIGN((pfn) + 1, 1UL << (order)) argument
57 * Page order with-respect-to which proactive compaction
87 unsigned int i, order, nr_pages; in split_map_pages() local
94 order = page_private(page); in split_map_pages()
95 nr_pages = 1 << order; in split_map_pages()
97 post_alloc_hook(page, order, __GFP_MOVABLE); in split_map_pages()
98 if (order) in split_map_pages()
153 defer_compaction(struct zone * zone,int order) defer_compaction() argument
168 compaction_deferred(struct zone * zone,int order) compaction_deferred() argument
191 compaction_defer_reset(struct zone * zone,int order,bool alloc_success) compaction_defer_reset() argument
205 compaction_restarting(struct zone * zone,int order) compaction_restarting() argument
596 unsigned int order; isolate_freepages_block() local
627 const unsigned int order = compound_order(page); isolate_freepages_block() local
1019 const unsigned int order = compound_order(page); isolate_migratepages_block() local
1451 next_search_order(struct compact_control * cc,int order) next_search_order() argument
1477 int order; fast_isolate_freepages() local
1852 int order; fast_find_migrateblock() local
2072 is_via_compact_memory(int order) is_via_compact_memory() argument
2171 unsigned int order; __compact_finished() local
2278 __compaction_suitable(struct zone * zone,int order,int highest_zoneidx,unsigned long wmark_target) __compaction_suitable() argument
2307 compaction_suitable(struct zone * zone,int order,int highest_zoneidx) compaction_suitable() argument
2350 compaction_zonelist_suitable(struct alloc_context * ac,int order,int alloc_flags) compaction_zonelist_suitable() argument
2618 compact_zone_order(struct zone * zone,int order,gfp_t gfp_mask,enum compact_priority prio,unsigned int alloc_flags,int highest_zoneidx,struct page ** capture) compact_zone_order() argument
2683 try_to_compact_pages(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,enum compact_priority prio,struct page ** capture) try_to_compact_pages() argument
3011 wakeup_kcompactd(pg_data_t * pgdat,int order,int highest_zoneidx) wakeup_kcompactd() argument
[all...]
/openbmc/linux/drivers/gpu/drm/ttm/
H A Dttm_pool.c80 unsigned int order) in ttm_pool_alloc_page() argument
91 if (order) in ttm_pool_alloc_page()
96 p = alloc_pages_node(pool->nid, gfp_flags, order); in ttm_pool_alloc_page()
98 p->private = order; in ttm_pool_alloc_page()
106 if (order) in ttm_pool_alloc_page()
109 vaddr = dma_alloc_attrs(pool->dev, (1ULL << order) * PAGE_SIZE, in ttm_pool_alloc_page()
122 dma->vaddr = (unsigned long)vaddr | order; in ttm_pool_alloc_page()
133 unsigned int order, struct page *p) in ttm_pool_free_page() argument
144 set_pages_wb(p, 1 << order); in ttm_pool_free_page()
148 __free_pages(p, order); in ttm_pool_free_page()
[all …]
/openbmc/linux/lib/
H A Dtest_xarray.c72 unsigned order, void *entry, gfp_t gfp) in xa_store_order() argument
74 XA_STATE_ORDER(xas, xa, index, order); in xa_store_order()
177 unsigned int order; in check_xa_mark_1() local
207 for (order = 2; order < max_order; order++) { in check_xa_mark_1()
208 unsigned long base = round_down(index, 1UL << order); in check_xa_mark_1()
209 unsigned long next = base + (1UL << order); in check_xa_mark_1()
217 xa_store_order(xa, index, order, xa_mk_index(index), in check_xa_mark_1()
328 unsigned int order; in check_xa_shrink() local
353 for (order = 0; order < max_order; order++) { in check_xa_shrink()
354 unsigned long max = (1UL << order) - 1; in check_xa_shrink()
[all …]
/openbmc/linux/tools/testing/radix-tree/
H A Dmultiorder.c16 unsigned order) in item_insert_order() argument
18 XA_STATE_ORDER(xas, xa, index, order); in item_insert_order()
19 struct item *item = item_create(index, order); in item_insert_order()
42 int order[NUM_ENTRIES] = {1, 1, 2, 3, 4, 1, 0, 1, 3, 0, 7}; in multiorder_iteration() local
47 err = item_insert_order(xa, index[i], order[i]); in multiorder_iteration()
53 if (j <= (index[i] | ((1 << order[i]) - 1))) in multiorder_iteration()
58 int height = order[i] / XA_CHUNK_SHIFT; in multiorder_iteration()
60 unsigned long mask = (1UL << order[i]) - 1; in multiorder_iteration()
66 assert(item->order == order[i]); in multiorder_iteration()
82 int order[MT_NUM_ENTRIES] = {1, 0, 2, 4, 3, 1, 3, 0, 7}; in multiorder_tagged_iteration() local
[all …]
/openbmc/linux/drivers/media/pci/cx18/
H A Dcx18-mailbox.c231 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument
240 mb = &order->mb; in epu_dma_done()
247 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done()
253 mdl_ack = order->mdl_ack; in epu_dma_done()
277 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done()
324 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument
327 char *str = order->str; in epu_debug()
329 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug()
335 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument
337 switch (order->rpu) { in epu_cmd()
[all …]
/openbmc/linux/drivers/gpu/drm/lib/
H A Ddrm_random.c16 void drm_random_reorder(unsigned int *order, unsigned int count, in drm_random_reorder() argument
24 swap(order[i], order[j]); in drm_random_reorder()
31 unsigned int *order, i; in drm_random_order() local
33 order = kmalloc_array(count, sizeof(*order), GFP_KERNEL); in drm_random_order()
34 if (!order) in drm_random_order()
35 return order; in drm_random_order()
38 order[i] = i; in drm_random_order()
40 drm_random_reorder(order, count, state); in drm_random_order()
41 return order; in drm_random_order()
/openbmc/linux/arch/riscv/kvm/
H A Dtlb.c22 unsigned long order) in kvm_riscv_local_hfence_gvma_vmid_gpa() argument
26 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_vmid_gpa()
33 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa()
38 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa()
50 unsigned long order) in kvm_riscv_local_hfence_gvma_gpa() argument
54 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_gpa()
61 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa()
66 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa()
81 unsigned long order) in kvm_riscv_local_hfence_vvma_asid_gva() argument
85 if (PTRS_PER_PTE < (gvsz >> order)) { in kvm_riscv_local_hfence_vvma_asid_gva()
[all …]
/openbmc/linux/mm/kmsan/
H A Dinit.c114 bool kmsan_memblock_free_pages(struct page *page, unsigned int order) in kmsan_memblock_free_pages() argument
118 if (!held_back[order].shadow) { in kmsan_memblock_free_pages()
119 held_back[order].shadow = page; in kmsan_memblock_free_pages()
122 if (!held_back[order].origin) { in kmsan_memblock_free_pages()
123 held_back[order].origin = page; in kmsan_memblock_free_pages()
126 shadow = held_back[order].shadow; in kmsan_memblock_free_pages()
127 origin = held_back[order].origin; in kmsan_memblock_free_pages()
128 kmsan_setup_meta(page, shadow, origin, order); in kmsan_memblock_free_pages()
130 held_back[order].shadow = NULL; in kmsan_memblock_free_pages()
131 held_back[order].origin = NULL; in kmsan_memblock_free_pages()
[all …]
/openbmc/linux/Documentation/trace/postprocess/
H A Dtrace-vmscan-postprocess.pl317 my $order = $1;
318 $perprocesspid{$process_pid}->{MM_VMSCAN_DIRECT_RECLAIM_BEGIN_PERORDER}[$order]++;
319 $perprocesspid{$process_pid}->{STATE_DIRECT_ORDER} = $order;
328 my $order = $perprocesspid{$process_pid}->{STATE_DIRECT_ORDER};
330 $perprocesspid{$process_pid}->{HIGH_DIRECT_RECLAIM_LATENCY}[$index] = "$order-$latency";
341 my $order = $2;
342 $perprocesspid{$process_pid}->{STATE_KSWAPD_ORDER} = $order;
347 $perprocesspid{$process_pid}->{MM_VMSCAN_KSWAPD_WAKE_PERORDER}[$order]++;
350 $perprocesspid{$process_pid}->{HIGH_KSWAPD_REWAKEUP_PERORDER}[$order]++;
360 my $order = $perprocesspid{$process_pid}->{STATE_KSWAPD_ORDER};
[all …]
/openbmc/linux/kernel/bpf/
H A Dcgroup_iter.c54 int order; member
77 if (p->order == BPF_CGROUP_ITER_DESCENDANTS_PRE) in cgroup_iter_seq_start()
79 else if (p->order == BPF_CGROUP_ITER_DESCENDANTS_POST) in cgroup_iter_seq_start()
110 if (p->order == BPF_CGROUP_ITER_DESCENDANTS_PRE) in cgroup_iter_seq_next()
112 else if (p->order == BPF_CGROUP_ITER_DESCENDANTS_POST) in cgroup_iter_seq_next()
114 else if (p->order == BPF_CGROUP_ITER_ANCESTORS_UP) in cgroup_iter_seq_next()
176 p->order = aux->cgroup.order; in BTF_ID_LIST_GLOBAL_SINGLE()
200 int order = linfo->cgroup.order; in bpf_iter_attach_cgroup() local
203 if (order != BPF_CGROUP_ITER_DESCENDANTS_PRE && in bpf_iter_attach_cgroup()
204 order != BPF_CGROUP_ITER_DESCENDANTS_POST && in bpf_iter_attach_cgroup()
[all …]
/openbmc/linux/drivers/gpu/drm/tests/
H A Ddrm_buddy_test.c21 static inline u64 get_size(int order, u64 chunk_size) in get_size() argument
23 return (1 << order) * chunk_size; in get_size()
228 unsigned int order; in check_mm() local
244 order = drm_buddy_block_order(root); in check_mm()
247 if (order != mm->max_order) { in check_mm()
268 block = list_first_entry_or_null(&mm->free_list[order], in check_mm()
271 kunit_err(test, "root mismatch at order=%u\n", order); in check_mm()
331 int order, top; in drm_test_buddy_alloc_pathological() local
358 for (order = top; order--;) { in drm_test_buddy_alloc_pathological()
359 size = get_size(order, PAGE_SIZE); in drm_test_buddy_alloc_pathological()
[all …]
/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/subdev/therm/
H A Dgk104.c34 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_enable() local
38 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable()
39 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_enable()
42 nvkm_mask(dev, 0x20200 + order[i].offset, 0xff00, 0x4500); in gk104_clkgate_enable()
50 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable()
51 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_enable()
54 nvkm_mask(dev, 0x20200 + order[i].offset, 0x00ff, 0x0045); in gk104_clkgate_enable()
63 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_fini() local
67 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_fini()
68 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_fini()
[all …]
/openbmc/linux/include/linux/
H A Dgfp.h171 static inline void arch_free_page(struct page *page, int order) { } in arch_free_page() argument
174 static inline void arch_alloc_page(struct page *page, int order) { } in arch_alloc_page() argument
177 struct page *__alloc_pages(gfp_t gfp, unsigned int order, int preferred_nid,
179 struct folio *__folio_alloc(gfp_t gfp, unsigned int order, int preferred_nid,
232 __alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node() argument
237 return __alloc_pages(gfp_mask, order, nid, NULL); in __alloc_pages_node()
241 struct folio *__folio_alloc_node(gfp_t gfp, unsigned int order, int nid) in __folio_alloc_node() argument
246 return __folio_alloc(gfp, order, nid, NULL); in __folio_alloc_node()
255 unsigned int order) in alloc_pages_node() argument
260 return __alloc_pages_node(nid, gfp_mask, order); in alloc_pages_node()
[all …]
H A Dcompaction.h61 * Number of free order-0 pages that should be available above given watermark
65 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument
69 * free scanner may have up to 1 << order pages on its list and then in compact_gap()
70 * try to split an (order - 1) free page. At that point, a gap of in compact_gap()
71 * 1 << order might not be enough, so it's safer to require twice that in compact_gap()
80 return 2UL << order; in compact_gap()
90 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order);
91 extern int fragmentation_index(struct zone *zone, unsigned int order);
93 unsigned int order, unsigned int alloc_flags,
97 extern bool compaction_suitable(struct zone *zone, int order,
110 compaction_suitable(struct zone * zone,int order,int highest_zoneidx) compaction_suitable() argument
124 wakeup_kcompactd(pg_data_t * pgdat,int order,int highest_zoneidx) wakeup_kcompactd() argument
[all...]
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/
H A Ddr_buddy.c75 unsigned int *order) in dr_buddy_find_free_seg() argument
99 *order = order_iter; in dr_buddy_find_free_seg()
120 unsigned int order, in mlx5dr_buddy_alloc_mem() argument
126 err = dr_buddy_find_free_seg(buddy, order, &seg, &order_iter); in mlx5dr_buddy_alloc_mem()
137 while (order_iter > order) { in mlx5dr_buddy_alloc_mem()
144 seg <<= order; in mlx5dr_buddy_alloc_mem()
151 unsigned int seg, unsigned int order) in mlx5dr_buddy_free_mem() argument
153 seg >>= order; in mlx5dr_buddy_free_mem()
158 while (test_bit(seg ^ 1, buddy->bitmap[order])) { in mlx5dr_buddy_free_mem()
159 bitmap_clear(buddy->bitmap[order], seg ^ 1, 1); in mlx5dr_buddy_free_mem()
[all …]
/openbmc/linux/arch/riscv/mm/
H A Dhugetlbpage.c35 unsigned long order; in huge_pte_alloc() local
68 for_each_napot_order(order) { in huge_pte_alloc()
69 if (napot_cont_size(order) == sz) { in huge_pte_alloc()
70 pte = pte_alloc_huge(mm, pmd, addr & napot_cont_mask(order)); in huge_pte_alloc()
88 unsigned long order; in huge_pte_offset() local
119 for_each_napot_order(order) { in huge_pte_offset()
120 if (napot_cont_size(order) == sz) { in huge_pte_offset()
121 pte = pte_offset_huge(pmd, addr & napot_cont_mask(order)); in huge_pte_offset()
186 unsigned long order; in arch_make_huge_pte() local
188 for_each_napot_order(order) { in arch_make_huge_pte()
270 unsigned long order; huge_ptep_set_access_flags() local
314 unsigned long order; huge_ptep_set_wrprotect() local
369 unsigned long order; is_napot_size() local
384 unsigned long order; napot_hugetlbpages_init() local
[all...]
/openbmc/linux/drivers/gpu/drm/i915/selftests/
H A Di915_syncmap.c274 unsigned int pass, order; in igt_syncmap_join_above() local
296 for (order = 0; order < 64; order += SHIFT) { in igt_syncmap_join_above()
297 u64 context = BIT_ULL(order); in igt_syncmap_join_above()
335 unsigned int step, order, idx; in igt_syncmap_join_below() local
345 for (order = 64 - SHIFT; order > 0; order -= SHIFT) { in igt_syncmap_join_below()
346 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below()
354 context, order, step, sync->height, sync->prefix); in igt_syncmap_join_below()
362 for (order = SHIFT; order < 64; order += SHIFT) { in igt_syncmap_join_below()
363 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below()
367 context, order, step); in igt_syncmap_join_below()
[all …]
H A Di915_random.c70 void i915_random_reorder(unsigned int *order, unsigned int count, in i915_random_reorder() argument
73 i915_prandom_shuffle(order, sizeof(*order), count, state); in i915_random_reorder()
78 unsigned int *order, i; in i915_random_order() local
80 order = kmalloc_array(count, sizeof(*order), in i915_random_order()
82 if (!order) in i915_random_order()
83 return order; in i915_random_order()
86 order[i] = i; in i915_random_order()
88 i915_random_reorder(order, count, state); in i915_random_order()
89 return order; in i915_random_order()

12345678910>>...99