Lines Matching full:last
102 tb->itree.last = tb->itree.start + tb->size - 1; in tb_record()
126 #define PAGE_FOR_EACH_TB(start, last, pagedesc, T, N) \ argument
127 for (T = foreach_tb_first(start, last), \
128 N = foreach_tb_next(T, start, last); \
130 T = N, N = foreach_tb_next(N, start, last))
135 tb_page_addr_t last) in foreach_tb_first() argument
137 IntervalTreeNode *n = interval_tree_iter_first(&tb_root, start, last); in foreach_tb_first()
143 tb_page_addr_t last) in foreach_tb_next() argument
148 n = interval_tree_iter_next(&tb->itree, start, last); in foreach_tb_next()
314 #define PAGE_FOR_EACH_TB(start, last, pagedesc, tb, n) \ argument
598 * Lock a range of pages ([@start,@last]) as well as the pages of all
603 tb_page_addr_t last) in page_collection_lock() argument
610 last >>= TARGET_PAGE_BITS; in page_collection_lock()
611 g_assert(start <= last); in page_collection_lock()
621 for (index = start; index <= last; index++) { in page_collection_lock()
1010 tb_page_addr_t last) in tb_invalidate_phys_range() argument
1017 PAGE_FOR_EACH_TB(start, last, unused, tb, n) { in tb_invalidate_phys_range()
1029 tb_page_addr_t start, last; in tb_invalidate_phys_page() local
1032 last = addr | ~TARGET_PAGE_MASK; in tb_invalidate_phys_page()
1033 tb_invalidate_phys_range(NULL, start, last); in tb_invalidate_phys_page()
1048 tb_page_addr_t last; in tb_invalidate_phys_page_unwind() local
1062 last = addr | ~TARGET_PAGE_MASK; in tb_invalidate_phys_page_unwind()
1066 PAGE_FOR_EACH_TB(addr, last, unused, tb, n) { in tb_invalidate_phys_page_unwind()
1100 tb_page_addr_t last, in tb_invalidate_phys_page_range__locked() argument
1109 tcg_debug_assert(((start ^ last) & TARGET_PAGE_MASK) == 0); in tb_invalidate_phys_page_range__locked()
1116 * We remove all the TBs in the range [start, last]. in tb_invalidate_phys_page_range__locked()
1119 PAGE_FOR_EACH_TB(start, last, p, tb, n) { in tb_invalidate_phys_page_range__locked()
1131 if (!(tb_last < start || tb_start > last)) { in tb_invalidate_phys_page_range__locked()
1164 * [start;last]. NOTE: start and end may refer to *different* physical pages.
1170 tb_page_addr_t last) in tb_invalidate_phys_range() argument
1175 pages = page_collection_lock(start, last); in tb_invalidate_phys_range()
1177 index_last = last >> TARGET_PAGE_BITS; in tb_invalidate_phys_range()
1188 page_last = MIN(page_last, last); in tb_invalidate_phys_range()
1206 ram_addr_t last = start + len - 1; in tb_invalidate_phys_range_fast() local
1207 struct page_collection *pages = page_collection_lock(start, last); in tb_invalidate_phys_range_fast()
1210 start, last, ra); in tb_invalidate_phys_range_fast()