Lines Matching full:last
102 tb->itree.last = tb->itree.start + tb->size - 1; in tb_record()
126 #define PAGE_FOR_EACH_TB(start, last, pagedesc, T, N) \ argument
127 for (T = foreach_tb_first(start, last), \
128 N = foreach_tb_next(T, start, last); \
130 T = N, N = foreach_tb_next(N, start, last))
135 tb_page_addr_t last) in foreach_tb_first() argument
137 IntervalTreeNode *n = interval_tree_iter_first(&tb_root, start, last); in foreach_tb_first()
143 tb_page_addr_t last) in foreach_tb_next() argument
148 n = interval_tree_iter_next(&tb->itree, start, last); in foreach_tb_next()
314 #define PAGE_FOR_EACH_TB(start, last, pagedesc, tb, n) \ argument
598 * Lock a range of pages ([@start,@last]) as well as the pages of all
603 tb_page_addr_t last) in page_collection_lock() argument
610 last >>= TARGET_PAGE_BITS; in page_collection_lock()
611 g_assert(start <= last); in page_collection_lock()
621 for (index = start; index <= last; index++) { in page_collection_lock()
1018 tb_page_addr_t last) in tb_invalidate_phys_range() argument
1025 PAGE_FOR_EACH_TB(start, last, unused, tb, n) { in tb_invalidate_phys_range()
1037 tb_page_addr_t start, last; in tb_invalidate_phys_page() local
1040 last = addr | ~TARGET_PAGE_MASK; in tb_invalidate_phys_page()
1041 tb_invalidate_phys_range(NULL, start, last); in tb_invalidate_phys_page()
1056 tb_page_addr_t last; in tb_invalidate_phys_page_unwind() local
1070 last = addr | ~TARGET_PAGE_MASK; in tb_invalidate_phys_page_unwind()
1074 PAGE_FOR_EACH_TB(addr, last, unused, tb, n) { in tb_invalidate_phys_page_unwind()
1108 tb_page_addr_t last, in tb_invalidate_phys_page_range__locked() argument
1117 tcg_debug_assert(((start ^ last) & TARGET_PAGE_MASK) == 0); in tb_invalidate_phys_page_range__locked()
1124 * We remove all the TBs in the range [start, last]. in tb_invalidate_phys_page_range__locked()
1127 PAGE_FOR_EACH_TB(start, last, p, tb, n) { in tb_invalidate_phys_page_range__locked()
1139 if (!(tb_last < start || tb_start > last)) { in tb_invalidate_phys_page_range__locked()
1172 * [start;last]. NOTE: start and end may refer to *different* physical pages.
1178 tb_page_addr_t last) in tb_invalidate_phys_range() argument
1183 pages = page_collection_lock(start, last); in tb_invalidate_phys_range()
1185 index_last = last >> TARGET_PAGE_BITS; in tb_invalidate_phys_range()
1196 page_last = MIN(page_last, last); in tb_invalidate_phys_range()
1214 ram_addr_t last = start + len - 1; in tb_invalidate_phys_range_fast() local
1215 struct page_collection *pages = page_collection_lock(start, last); in tb_invalidate_phys_range_fast()
1218 start, last, ra); in tb_invalidate_phys_range_fast()