1/* 2 * Memory access templates for MemoryRegionCache 3 * 4 * Copyright (c) 2018 Red Hat, Inc. 5 * 6 * This library is free software; you can redistribute it and/or 7 * modify it under the terms of the GNU Lesser General Public 8 * License as published by the Free Software Foundation; either 9 * version 2.1 of the License, or (at your option) any later version. 10 * 11 * This library is distributed in the hope that it will be useful, 12 * but WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 14 * Lesser General Public License for more details. 15 * 16 * You should have received a copy of the GNU Lesser General Public 17 * License along with this library; if not, see <http://www.gnu.org/licenses/>. 18 */ 19 20#define ADDRESS_SPACE_LD_CACHED(size) \ 21 glue(glue(address_space_ld, size), glue(ENDIANNESS, _cached)) 22#define ADDRESS_SPACE_LD_CACHED_SLOW(size) \ 23 glue(glue(address_space_ld, size), glue(ENDIANNESS, _cached_slow)) 24#define LD_P(size) \ 25 glue(glue(ld, size), glue(ENDIANNESS, _p)) 26 27static inline uint32_t ADDRESS_SPACE_LD_CACHED(l)(MemoryRegionCache *cache, 28 hwaddr addr, MemTxAttrs attrs, MemTxResult *result) 29{ 30 assert(addr < cache->len && 4 <= cache->len - addr); 31 fuzz_dma_read_cb(cache->xlat + addr, 4, cache->mrs.mr); 32 if (likely(cache->ptr)) { 33 return LD_P(l)(cache->ptr + addr); 34 } else { 35 return ADDRESS_SPACE_LD_CACHED_SLOW(l)(cache, addr, attrs, result); 36 } 37} 38 39static inline uint64_t ADDRESS_SPACE_LD_CACHED(q)(MemoryRegionCache *cache, 40 hwaddr addr, MemTxAttrs attrs, MemTxResult *result) 41{ 42 assert(addr < cache->len && 8 <= cache->len - addr); 43 fuzz_dma_read_cb(cache->xlat + addr, 8, cache->mrs.mr); 44 if (likely(cache->ptr)) { 45 return LD_P(q)(cache->ptr + addr); 46 } else { 47 return ADDRESS_SPACE_LD_CACHED_SLOW(q)(cache, addr, attrs, result); 48 } 49} 50 51static inline uint32_t ADDRESS_SPACE_LD_CACHED(uw)(MemoryRegionCache *cache, 52 hwaddr addr, MemTxAttrs attrs, MemTxResult *result) 53{ 54 assert(addr < cache->len && 2 <= cache->len - addr); 55 fuzz_dma_read_cb(cache->xlat + addr, 2, cache->mrs.mr); 56 if (likely(cache->ptr)) { 57 return LD_P(uw)(cache->ptr + addr); 58 } else { 59 return ADDRESS_SPACE_LD_CACHED_SLOW(uw)(cache, addr, attrs, result); 60 } 61} 62 63#undef ADDRESS_SPACE_LD_CACHED 64#undef ADDRESS_SPACE_LD_CACHED_SLOW 65#undef LD_P 66 67#define ADDRESS_SPACE_ST_CACHED(size) \ 68 glue(glue(address_space_st, size), glue(ENDIANNESS, _cached)) 69#define ADDRESS_SPACE_ST_CACHED_SLOW(size) \ 70 glue(glue(address_space_st, size), glue(ENDIANNESS, _cached_slow)) 71#define ST_P(size) \ 72 glue(glue(st, size), glue(ENDIANNESS, _p)) 73 74static inline void ADDRESS_SPACE_ST_CACHED(l)(MemoryRegionCache *cache, 75 hwaddr addr, uint32_t val, MemTxAttrs attrs, MemTxResult *result) 76{ 77 assert(addr < cache->len && 4 <= cache->len - addr); 78 if (likely(cache->ptr)) { 79 ST_P(l)(cache->ptr + addr, val); 80 } else { 81 ADDRESS_SPACE_ST_CACHED_SLOW(l)(cache, addr, val, attrs, result); 82 } 83} 84 85static inline void ADDRESS_SPACE_ST_CACHED(w)(MemoryRegionCache *cache, 86 hwaddr addr, uint32_t val, MemTxAttrs attrs, MemTxResult *result) 87{ 88 assert(addr < cache->len && 2 <= cache->len - addr); 89 if (likely(cache->ptr)) { 90 ST_P(w)(cache->ptr + addr, val); 91 } else { 92 ADDRESS_SPACE_ST_CACHED_SLOW(w)(cache, addr, val, attrs, result); 93 } 94} 95 96static inline void ADDRESS_SPACE_ST_CACHED(q)(MemoryRegionCache *cache, 97 hwaddr addr, uint64_t val, MemTxAttrs attrs, MemTxResult *result) 98{ 99 assert(addr < cache->len && 8 <= cache->len - addr); 100 if (likely(cache->ptr)) { 101 ST_P(q)(cache->ptr + addr, val); 102 } else { 103 ADDRESS_SPACE_ST_CACHED_SLOW(q)(cache, addr, val, attrs, result); 104 } 105} 106 107#undef ADDRESS_SPACE_ST_CACHED 108#undef ADDRESS_SPACE_ST_CACHED_SLOW 109#undef ST_P 110 111#undef ENDIANNESS 112