Lines Matching refs:ei

173                                      uint32_t wi, uint32_t *ei)  in split_tlb_entry_spec_way()  argument
187 *ei = (v >> 12) & (is32 ? 0x7 : 0x3); in split_tlb_entry_spec_way()
193 *ei = (v >> eibase) & 0x3; in split_tlb_entry_spec_way()
200 *ei = (v >> eibase) & 0x3; in split_tlb_entry_spec_way()
202 *ei = (v >> 27) & 0x1; in split_tlb_entry_spec_way()
209 *ei = (v >> eibase) & 0x7; in split_tlb_entry_spec_way()
211 *ei = (v >> 28) & 0x1; in split_tlb_entry_spec_way()
216 *ei = 0; in split_tlb_entry_spec_way()
228 uint32_t *vpn, uint32_t *wi, uint32_t *ei) in split_tlb_entry_spec() argument
233 split_tlb_entry_spec_way(env, v, dtlb, vpn, *wi, ei); in split_tlb_entry_spec()
241 *ei = (v >> 29) & 0x7; in split_tlb_entry_spec()
247 unsigned wi, unsigned ei) in xtensa_tlb_get_entry() argument
251 assert(wi < tlb->nways && ei < tlb->way_size[wi]); in xtensa_tlb_get_entry()
253 env->dtlb[wi] + ei : in xtensa_tlb_get_entry()
254 env->itlb[wi] + ei; in xtensa_tlb_get_entry()
262 uint32_t ei; in get_tlb_entry() local
264 if (split_tlb_entry_spec(env, v, dtlb, &vpn, &wi, &ei)) { in get_tlb_entry()
268 return xtensa_tlb_get_entry(env, dtlb, wi, ei); in get_tlb_entry()
276 unsigned wi, unsigned ei, uint32_t vpn, in xtensa_tlb_set_entry_mmu() argument
286 unsigned wi, unsigned ei, in xtensa_tlb_set_entry() argument
290 xtensa_tlb_entry *entry = xtensa_tlb_get_entry(env, dtlb, wi, ei); in xtensa_tlb_set_entry()
297 xtensa_tlb_set_entry_mmu(env, entry, dtlb, wi, ei, vpn, pte); in xtensa_tlb_set_entry()
302 __func__, dtlb, wi, ei); in xtensa_tlb_set_entry()
336 unsigned wi, ei; in reset_tlb_mmu_all_ways() local
339 for (ei = 0; ei < tlb->way_size[wi]; ++ei) { in reset_tlb_mmu_all_ways()
340 entry[wi][ei].asid = 0; in reset_tlb_mmu_all_ways()
341 entry[wi][ei].variable = true; in reset_tlb_mmu_all_ways()
384 uint32_t ei; in reset_tlb_mmu_ways56() local
385 for (ei = 0; ei < 8; ++ei) { in reset_tlb_mmu_ways56()
386 entry[6][ei].vaddr = ei << 29; in reset_tlb_mmu_ways56()
387 entry[6][ei].paddr = ei << 29; in reset_tlb_mmu_ways56()
388 entry[6][ei].asid = 1; in reset_tlb_mmu_ways56()
389 entry[6][ei].attr = 3; in reset_tlb_mmu_ways56()
397 unsigned ei; in reset_tlb_region_way0() local
399 for (ei = 0; ei < 8; ++ei) { in reset_tlb_region_way0()
400 entry[0][ei].vaddr = ei << 29; in reset_tlb_region_way0()
401 entry[0][ei].paddr = ei << 29; in reset_tlb_region_way0()
402 entry[0][ei].asid = 1; in reset_tlb_region_way0()
403 entry[0][ei].attr = 2; in reset_tlb_region_way0()
404 entry[0][ei].variable = true; in reset_tlb_region_way0()
472 uint32_t ei; in xtensa_tlb_lookup() local
473 split_tlb_entry_spec_way(env, addr, dtlb, &vpn, wi, &ei); in xtensa_tlb_lookup()
474 if (entry[wi][ei].vaddr == vpn && entry[wi][ei].asid) { in xtensa_tlb_lookup()
475 unsigned ring = get_ring(env, entry[wi][ei].asid); in xtensa_tlb_lookup()
483 *pei = ei; in xtensa_tlb_lookup()
535 uint32_t ei; in HELPER() local
537 int res = xtensa_tlb_lookup(env, v, dtlb, &wi, &ei, &ring); in HELPER()
561 uint32_t ei; in HELPER() local
562 if (split_tlb_entry_spec(env, v, dtlb, &vpn, &wi, &ei)) { in HELPER()
563 xtensa_tlb_set_entry(env, dtlb, wi, ei, vpn, p); in HELPER()
809 uint32_t ei; in get_physical_addr_mmu() local
815 int ret = xtensa_tlb_lookup(env, vaddr, dtlb, &wi, &ei, &ring); in get_physical_addr_mmu()
821 split_tlb_entry_spec_way(env, vaddr, dtlb, &vpn, wi, &ei); in get_physical_addr_mmu()
825 xtensa_tlb_set_entry(env, dtlb, wi, ei, vpn, pte); in get_physical_addr_mmu()
830 xtensa_tlb_set_entry_mmu(env, &tmp_entry, dtlb, wi, ei, vpn, pte); in get_physical_addr_mmu()
840 entry = xtensa_tlb_get_entry(env, dtlb, wi, ei); in get_physical_addr_mmu()
908 uint32_t ei = (vaddr >> 29) & 0x7; in get_physical_addr_region() local
910 xtensa_tlb_get_entry(env, dtlb, wi, ei); in get_physical_addr_region()
1082 unsigned wi, ei; in dump_tlb() local
1102 for (ei = 0; ei < conf->way_size[wi]; ++ei) { in dump_tlb()
1104 xtensa_tlb_get_entry(env, dtlb, wi, ei); in dump_tlb()