Message ID | 20241114160131.48616-30-richard.henderson@linaro.org (mailing list archive) |
---|---|
State | New |
Headers | show |
Series | accel/tcg: Convert victim tlb to IntervalTree | expand |
On 11/14/24 08:01, Richard Henderson wrote: > Signed-off-by: Richard Henderson <richard.henderson@linaro.org> > --- > accel/tcg/cputlb.c | 151 ++++++++++++++++++++++----------------------- > 1 file changed, 74 insertions(+), 77 deletions(-) > > diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c > index a33bebf55a..8f459be5a8 100644 > --- a/accel/tcg/cputlb.c > +++ b/accel/tcg/cputlb.c > @@ -1684,10 +1684,7 @@ bool tlb_plugin_lookup(CPUState *cpu, vaddr addr, int mmu_idx, > */ > > typedef struct MMULookupPageData { > - CPUTLBEntryFull *full; > - void *haddr; > vaddr addr; > - int flags; > int size; > TLBLookupOutput o; > } MMULookupPageData; > @@ -1724,10 +1721,6 @@ static void mmu_lookup1(CPUState *cpu, MMULookupPageData *data, MemOp memop, > }; > > tlb_lookup_nofail(cpu, &data->o, &i); > - > - data->full = &data->o.full; > - data->flags = data->o.flags; > - data->haddr = data->o.haddr; > } > > /** > @@ -1743,24 +1736,22 @@ static void mmu_lookup1(CPUState *cpu, MMULookupPageData *data, MemOp memop, > static void mmu_watch_or_dirty(CPUState *cpu, MMULookupPageData *data, > MMUAccessType access_type, uintptr_t ra) > { > - CPUTLBEntryFull *full = data->full; > - vaddr addr = data->addr; > - int flags = data->flags; > - int size = data->size; > + int flags = data->o.flags; > > /* On watchpoint hit, this will longjmp out. */ > if (flags & TLB_WATCHPOINT) { > int wp = access_type == MMU_DATA_STORE ? BP_MEM_WRITE : BP_MEM_READ; > - cpu_check_watchpoint(cpu, addr, size, full->attrs, wp, ra); > + cpu_check_watchpoint(cpu, data->addr, data->size, > + data->o.full.attrs, wp, ra); > flags &= ~TLB_WATCHPOINT; > } > > /* Note that notdirty is only set for writes. */ > if (flags & TLB_NOTDIRTY) { > - notdirty_write(cpu, addr, size, full, ra); > + notdirty_write(cpu, data->addr, data->size, &data->o.full, ra); > flags &= ~TLB_NOTDIRTY; > } > - data->flags = flags; > + data->o.flags = flags; > } > > /** > @@ -1795,7 +1786,7 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi, > if (likely(!crosspage)) { > mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra); > > - flags = l->page[0].flags; > + flags = l->page[0].o.flags; > if (unlikely(flags & (TLB_WATCHPOINT | TLB_NOTDIRTY))) { > mmu_watch_or_dirty(cpu, &l->page[0], type, ra); > } > @@ -1812,7 +1803,7 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi, > mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra); > mmu_lookup1(cpu, &l->page[1], 0, l->mmu_idx, type, ra); > > - flags = l->page[0].flags | l->page[1].flags; > + flags = l->page[0].o.flags | l->page[1].o.flags; > if (unlikely(flags & (TLB_WATCHPOINT | TLB_NOTDIRTY))) { > mmu_watch_or_dirty(cpu, &l->page[0], type, ra); > mmu_watch_or_dirty(cpu, &l->page[1], type, ra); > @@ -2029,7 +2020,7 @@ static Int128 do_ld16_mmio_beN(CPUState *cpu, CPUTLBEntryFull *full, > */ > static uint64_t do_ld_bytes_beN(MMULookupPageData *p, uint64_t ret_be) > { > - uint8_t *haddr = p->haddr; > + uint8_t *haddr = p->o.haddr; > int i, size = p->size; > > for (i = 0; i < size; i++) { > @@ -2047,7 +2038,7 @@ static uint64_t do_ld_bytes_beN(MMULookupPageData *p, uint64_t ret_be) > */ > static uint64_t do_ld_parts_beN(MMULookupPageData *p, uint64_t ret_be) > { > - void *haddr = p->haddr; > + void *haddr = p->o.haddr; > int size = p->size; > > do { > @@ -2097,7 +2088,7 @@ static uint64_t do_ld_parts_beN(MMULookupPageData *p, uint64_t ret_be) > static uint64_t do_ld_whole_be4(MMULookupPageData *p, uint64_t ret_be) > { > int o = p->addr & 3; > - uint32_t x = load_atomic4(p->haddr - o); > + uint32_t x = load_atomic4(p->o.haddr - o); > > x = cpu_to_be32(x); > x <<= o * 8; > @@ -2117,7 +2108,7 @@ static uint64_t do_ld_whole_be8(CPUState *cpu, uintptr_t ra, > MMULookupPageData *p, uint64_t ret_be) > { > int o = p->addr & 7; > - uint64_t x = load_atomic8_or_exit(cpu, ra, p->haddr - o); > + uint64_t x = load_atomic8_or_exit(cpu, ra, p->o.haddr - o); > > x = cpu_to_be64(x); > x <<= o * 8; > @@ -2137,7 +2128,7 @@ static Int128 do_ld_whole_be16(CPUState *cpu, uintptr_t ra, > MMULookupPageData *p, uint64_t ret_be) > { > int o = p->addr & 15; > - Int128 x, y = load_atomic16_or_exit(cpu, ra, p->haddr - o); > + Int128 x, y = load_atomic16_or_exit(cpu, ra, p->o.haddr - o); > int size = p->size; > > if (!HOST_BIG_ENDIAN) { > @@ -2160,8 +2151,8 @@ static uint64_t do_ld_beN(CPUState *cpu, MMULookupPageData *p, > MemOp atom; > unsigned tmp, half_size; > > - if (unlikely(p->flags & TLB_MMIO)) { > - return do_ld_mmio_beN(cpu, p->full, ret_be, p->addr, p->size, > + if (unlikely(p->o.flags & TLB_MMIO)) { > + return do_ld_mmio_beN(cpu, &p->o.full, ret_be, p->addr, p->size, > mmu_idx, type, ra); > } > > @@ -2210,8 +2201,9 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p, > uint64_t b; > MemOp atom; > > - if (unlikely(p->flags & TLB_MMIO)) { > - return do_ld16_mmio_beN(cpu, p->full, a, p->addr, size, mmu_idx, ra); > + if (unlikely(p->o.flags & TLB_MMIO)) { > + return do_ld16_mmio_beN(cpu, &p->o.full, a, p->addr, > + size, mmu_idx, ra); > } > > /* > @@ -2223,7 +2215,7 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p, > case MO_ATOM_SUBALIGN: > p->size = size - 8; > a = do_ld_parts_beN(p, a); > - p->haddr += size - 8; > + p->o.haddr += size - 8; > p->size = 8; > b = do_ld_parts_beN(p, 0); > break; > @@ -2242,7 +2234,7 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p, > case MO_ATOM_NONE: > p->size = size - 8; > a = do_ld_bytes_beN(p, a); > - b = ldq_be_p(p->haddr + size - 8); > + b = ldq_be_p(p->o.haddr + size - 8); > break; > > default: > @@ -2255,10 +2247,11 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p, > static uint8_t do_ld_1(CPUState *cpu, MMULookupPageData *p, int mmu_idx, > MMUAccessType type, uintptr_t ra) > { > - if (unlikely(p->flags & TLB_MMIO)) { > - return do_ld_mmio_beN(cpu, p->full, 0, p->addr, 1, mmu_idx, type, ra); > + if (unlikely(p->o.flags & TLB_MMIO)) { > + return do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 1, > + mmu_idx, type, ra); > } else { > - return *(uint8_t *)p->haddr; > + return *(uint8_t *)p->o.haddr; > } > } > > @@ -2267,14 +2260,15 @@ static uint16_t do_ld_2(CPUState *cpu, MMULookupPageData *p, int mmu_idx, > { > uint16_t ret; > > - if (unlikely(p->flags & TLB_MMIO)) { > - ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 2, mmu_idx, type, ra); > + if (unlikely(p->o.flags & TLB_MMIO)) { > + ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 2, > + mmu_idx, type, ra); > if ((memop & MO_BSWAP) == MO_LE) { > ret = bswap16(ret); > } > } else { > /* Perform the load host endian, then swap if necessary. */ > - ret = load_atom_2(cpu, ra, p->haddr, memop); > + ret = load_atom_2(cpu, ra, p->o.haddr, memop); > if (memop & MO_BSWAP) { > ret = bswap16(ret); > } > @@ -2287,14 +2281,15 @@ static uint32_t do_ld_4(CPUState *cpu, MMULookupPageData *p, int mmu_idx, > { > uint32_t ret; > > - if (unlikely(p->flags & TLB_MMIO)) { > - ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 4, mmu_idx, type, ra); > + if (unlikely(p->o.flags & TLB_MMIO)) { > + ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 4, > + mmu_idx, type, ra); > if ((memop & MO_BSWAP) == MO_LE) { > ret = bswap32(ret); > } > } else { > /* Perform the load host endian. */ > - ret = load_atom_4(cpu, ra, p->haddr, memop); > + ret = load_atom_4(cpu, ra, p->o.haddr, memop); > if (memop & MO_BSWAP) { > ret = bswap32(ret); > } > @@ -2307,14 +2302,15 @@ static uint64_t do_ld_8(CPUState *cpu, MMULookupPageData *p, int mmu_idx, > { > uint64_t ret; > > - if (unlikely(p->flags & TLB_MMIO)) { > - ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 8, mmu_idx, type, ra); > + if (unlikely(p->o.flags & TLB_MMIO)) { > + ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 8, > + mmu_idx, type, ra); > if ((memop & MO_BSWAP) == MO_LE) { > ret = bswap64(ret); > } > } else { > /* Perform the load host endian. */ > - ret = load_atom_8(cpu, ra, p->haddr, memop); > + ret = load_atom_8(cpu, ra, p->o.haddr, memop); > if (memop & MO_BSWAP) { > ret = bswap64(ret); > } > @@ -2414,15 +2410,15 @@ static Int128 do_ld16_mmu(CPUState *cpu, vaddr addr, > cpu_req_mo(TCG_MO_LD_LD | TCG_MO_ST_LD); > crosspage = mmu_lookup(cpu, addr, oi, ra, MMU_DATA_LOAD, &l); > if (likely(!crosspage)) { > - if (unlikely(l.page[0].flags & TLB_MMIO)) { > - ret = do_ld16_mmio_beN(cpu, l.page[0].full, 0, addr, 16, > + if (unlikely(l.page[0].o.flags & TLB_MMIO)) { > + ret = do_ld16_mmio_beN(cpu, &l.page[0].o.full, 0, addr, 16, > l.mmu_idx, ra); > if ((l.memop & MO_BSWAP) == MO_LE) { > ret = bswap128(ret); > } > } else { > /* Perform the load host endian. */ > - ret = load_atom_16(cpu, ra, l.page[0].haddr, l.memop); > + ret = load_atom_16(cpu, ra, l.page[0].o.haddr, l.memop); > if (l.memop & MO_BSWAP) { > ret = bswap128(ret); > } > @@ -2568,10 +2564,10 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p, > MemOp atom; > unsigned tmp, half_size; > > - if (unlikely(p->flags & TLB_MMIO)) { > - return do_st_mmio_leN(cpu, p->full, val_le, p->addr, > + if (unlikely(p->o.flags & TLB_MMIO)) { > + return do_st_mmio_leN(cpu, &p->o.full, val_le, p->addr, > p->size, mmu_idx, ra); > - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { > + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { > return val_le >> (p->size * 8); > } > > @@ -2582,7 +2578,7 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p, > atom = mop & MO_ATOM_MASK; > switch (atom) { > case MO_ATOM_SUBALIGN: > - return store_parts_leN(p->haddr, p->size, val_le); > + return store_parts_leN(p->o.haddr, p->size, val_le); > > case MO_ATOM_IFALIGN_PAIR: > case MO_ATOM_WITHIN16_PAIR: > @@ -2593,9 +2589,9 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p, > ? p->size == half_size > : p->size >= half_size) { > if (!HAVE_al8_fast && p->size <= 4) { > - return store_whole_le4(p->haddr, p->size, val_le); > + return store_whole_le4(p->o.haddr, p->size, val_le); > } else if (HAVE_al8) { > - return store_whole_le8(p->haddr, p->size, val_le); > + return store_whole_le8(p->o.haddr, p->size, val_le); > } else { > cpu_loop_exit_atomic(cpu, ra); > } > @@ -2605,7 +2601,7 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p, > case MO_ATOM_IFALIGN: > case MO_ATOM_WITHIN16: > case MO_ATOM_NONE: > - return store_bytes_leN(p->haddr, p->size, val_le); > + return store_bytes_leN(p->o.haddr, p->size, val_le); > > default: > g_assert_not_reached(); > @@ -2622,10 +2618,10 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, > int size = p->size; > MemOp atom; > > - if (unlikely(p->flags & TLB_MMIO)) { > - return do_st16_mmio_leN(cpu, p->full, val_le, p->addr, > + if (unlikely(p->o.flags & TLB_MMIO)) { > + return do_st16_mmio_leN(cpu, &p->o.full, val_le, p->addr, > size, mmu_idx, ra); > - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { > + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { > return int128_gethi(val_le) >> ((size - 8) * 8); > } > > @@ -2636,8 +2632,8 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, > atom = mop & MO_ATOM_MASK; > switch (atom) { > case MO_ATOM_SUBALIGN: > - store_parts_leN(p->haddr, 8, int128_getlo(val_le)); > - return store_parts_leN(p->haddr + 8, p->size - 8, > + store_parts_leN(p->o.haddr, 8, int128_getlo(val_le)); > + return store_parts_leN(p->o.haddr + 8, p->size - 8, > int128_gethi(val_le)); > > case MO_ATOM_WITHIN16_PAIR: > @@ -2645,7 +2641,7 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, > if (!HAVE_CMPXCHG128) { > cpu_loop_exit_atomic(cpu, ra); > } > - return store_whole_le16(p->haddr, p->size, val_le); > + return store_whole_le16(p->o.haddr, p->size, val_le); > > case MO_ATOM_IFALIGN_PAIR: > /* > @@ -2655,8 +2651,8 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, > case MO_ATOM_IFALIGN: > case MO_ATOM_WITHIN16: > case MO_ATOM_NONE: > - stq_le_p(p->haddr, int128_getlo(val_le)); > - return store_bytes_leN(p->haddr + 8, p->size - 8, > + stq_le_p(p->o.haddr, int128_getlo(val_le)); > + return store_bytes_leN(p->o.haddr + 8, p->size - 8, > int128_gethi(val_le)); > > default: > @@ -2667,69 +2663,69 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, > static void do_st_1(CPUState *cpu, MMULookupPageData *p, uint8_t val, > int mmu_idx, uintptr_t ra) > { > - if (unlikely(p->flags & TLB_MMIO)) { > - do_st_mmio_leN(cpu, p->full, val, p->addr, 1, mmu_idx, ra); > - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { > + if (unlikely(p->o.flags & TLB_MMIO)) { > + do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 1, mmu_idx, ra); > + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { > /* nothing */ > } else { > - *(uint8_t *)p->haddr = val; > + *(uint8_t *)p->o.haddr = val; > } > } > > static void do_st_2(CPUState *cpu, MMULookupPageData *p, uint16_t val, > int mmu_idx, MemOp memop, uintptr_t ra) > { > - if (unlikely(p->flags & TLB_MMIO)) { > + if (unlikely(p->o.flags & TLB_MMIO)) { > if ((memop & MO_BSWAP) != MO_LE) { > val = bswap16(val); > } > - do_st_mmio_leN(cpu, p->full, val, p->addr, 2, mmu_idx, ra); > - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { > + do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 2, mmu_idx, ra); > + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { > /* nothing */ > } else { > /* Swap to host endian if necessary, then store. */ > if (memop & MO_BSWAP) { > val = bswap16(val); > } > - store_atom_2(cpu, ra, p->haddr, memop, val); > + store_atom_2(cpu, ra, p->o.haddr, memop, val); > } > } > > static void do_st_4(CPUState *cpu, MMULookupPageData *p, uint32_t val, > int mmu_idx, MemOp memop, uintptr_t ra) > { > - if (unlikely(p->flags & TLB_MMIO)) { > + if (unlikely(p->o.flags & TLB_MMIO)) { > if ((memop & MO_BSWAP) != MO_LE) { > val = bswap32(val); > } > - do_st_mmio_leN(cpu, p->full, val, p->addr, 4, mmu_idx, ra); > - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { > + do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 4, mmu_idx, ra); > + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { > /* nothing */ > } else { > /* Swap to host endian if necessary, then store. */ > if (memop & MO_BSWAP) { > val = bswap32(val); > } > - store_atom_4(cpu, ra, p->haddr, memop, val); > + store_atom_4(cpu, ra, p->o.haddr, memop, val); > } > } > > static void do_st_8(CPUState *cpu, MMULookupPageData *p, uint64_t val, > int mmu_idx, MemOp memop, uintptr_t ra) > { > - if (unlikely(p->flags & TLB_MMIO)) { > + if (unlikely(p->o.flags & TLB_MMIO)) { > if ((memop & MO_BSWAP) != MO_LE) { > val = bswap64(val); > } > - do_st_mmio_leN(cpu, p->full, val, p->addr, 8, mmu_idx, ra); > - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { > + do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 8, mmu_idx, ra); > + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { > /* nothing */ > } else { > /* Swap to host endian if necessary, then store. */ > if (memop & MO_BSWAP) { > val = bswap64(val); > } > - store_atom_8(cpu, ra, p->haddr, memop, val); > + store_atom_8(cpu, ra, p->o.haddr, memop, val); > } > } > > @@ -2822,19 +2818,20 @@ static void do_st16_mmu(CPUState *cpu, vaddr addr, Int128 val, > cpu_req_mo(TCG_MO_LD_ST | TCG_MO_ST_ST); > crosspage = mmu_lookup(cpu, addr, oi, ra, MMU_DATA_STORE, &l); > if (likely(!crosspage)) { > - if (unlikely(l.page[0].flags & TLB_MMIO)) { > + if (unlikely(l.page[0].o.flags & TLB_MMIO)) { > if ((l.memop & MO_BSWAP) != MO_LE) { > val = bswap128(val); > } > - do_st16_mmio_leN(cpu, l.page[0].full, val, addr, 16, l.mmu_idx, ra); > - } else if (unlikely(l.page[0].flags & TLB_DISCARD_WRITE)) { > + do_st16_mmio_leN(cpu, &l.page[0].o.full, val, addr, > + 16, l.mmu_idx, ra); > + } else if (unlikely(l.page[0].o.flags & TLB_DISCARD_WRITE)) { > /* nothing */ > } else { > /* Swap to host endian if necessary, then store. */ > if (l.memop & MO_BSWAP) { > val = bswap128(val); > } > - store_atom_16(cpu, ra, l.page[0].haddr, l.memop, val); > + store_atom_16(cpu, ra, l.page[0].o.haddr, l.memop, val); > } > return; > } Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c index a33bebf55a..8f459be5a8 100644 --- a/accel/tcg/cputlb.c +++ b/accel/tcg/cputlb.c @@ -1684,10 +1684,7 @@ bool tlb_plugin_lookup(CPUState *cpu, vaddr addr, int mmu_idx, */ typedef struct MMULookupPageData { - CPUTLBEntryFull *full; - void *haddr; vaddr addr; - int flags; int size; TLBLookupOutput o; } MMULookupPageData; @@ -1724,10 +1721,6 @@ static void mmu_lookup1(CPUState *cpu, MMULookupPageData *data, MemOp memop, }; tlb_lookup_nofail(cpu, &data->o, &i); - - data->full = &data->o.full; - data->flags = data->o.flags; - data->haddr = data->o.haddr; } /** @@ -1743,24 +1736,22 @@ static void mmu_lookup1(CPUState *cpu, MMULookupPageData *data, MemOp memop, static void mmu_watch_or_dirty(CPUState *cpu, MMULookupPageData *data, MMUAccessType access_type, uintptr_t ra) { - CPUTLBEntryFull *full = data->full; - vaddr addr = data->addr; - int flags = data->flags; - int size = data->size; + int flags = data->o.flags; /* On watchpoint hit, this will longjmp out. */ if (flags & TLB_WATCHPOINT) { int wp = access_type == MMU_DATA_STORE ? BP_MEM_WRITE : BP_MEM_READ; - cpu_check_watchpoint(cpu, addr, size, full->attrs, wp, ra); + cpu_check_watchpoint(cpu, data->addr, data->size, + data->o.full.attrs, wp, ra); flags &= ~TLB_WATCHPOINT; } /* Note that notdirty is only set for writes. */ if (flags & TLB_NOTDIRTY) { - notdirty_write(cpu, addr, size, full, ra); + notdirty_write(cpu, data->addr, data->size, &data->o.full, ra); flags &= ~TLB_NOTDIRTY; } - data->flags = flags; + data->o.flags = flags; } /** @@ -1795,7 +1786,7 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi, if (likely(!crosspage)) { mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra); - flags = l->page[0].flags; + flags = l->page[0].o.flags; if (unlikely(flags & (TLB_WATCHPOINT | TLB_NOTDIRTY))) { mmu_watch_or_dirty(cpu, &l->page[0], type, ra); } @@ -1812,7 +1803,7 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi, mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra); mmu_lookup1(cpu, &l->page[1], 0, l->mmu_idx, type, ra); - flags = l->page[0].flags | l->page[1].flags; + flags = l->page[0].o.flags | l->page[1].o.flags; if (unlikely(flags & (TLB_WATCHPOINT | TLB_NOTDIRTY))) { mmu_watch_or_dirty(cpu, &l->page[0], type, ra); mmu_watch_or_dirty(cpu, &l->page[1], type, ra); @@ -2029,7 +2020,7 @@ static Int128 do_ld16_mmio_beN(CPUState *cpu, CPUTLBEntryFull *full, */ static uint64_t do_ld_bytes_beN(MMULookupPageData *p, uint64_t ret_be) { - uint8_t *haddr = p->haddr; + uint8_t *haddr = p->o.haddr; int i, size = p->size; for (i = 0; i < size; i++) { @@ -2047,7 +2038,7 @@ static uint64_t do_ld_bytes_beN(MMULookupPageData *p, uint64_t ret_be) */ static uint64_t do_ld_parts_beN(MMULookupPageData *p, uint64_t ret_be) { - void *haddr = p->haddr; + void *haddr = p->o.haddr; int size = p->size; do { @@ -2097,7 +2088,7 @@ static uint64_t do_ld_parts_beN(MMULookupPageData *p, uint64_t ret_be) static uint64_t do_ld_whole_be4(MMULookupPageData *p, uint64_t ret_be) { int o = p->addr & 3; - uint32_t x = load_atomic4(p->haddr - o); + uint32_t x = load_atomic4(p->o.haddr - o); x = cpu_to_be32(x); x <<= o * 8; @@ -2117,7 +2108,7 @@ static uint64_t do_ld_whole_be8(CPUState *cpu, uintptr_t ra, MMULookupPageData *p, uint64_t ret_be) { int o = p->addr & 7; - uint64_t x = load_atomic8_or_exit(cpu, ra, p->haddr - o); + uint64_t x = load_atomic8_or_exit(cpu, ra, p->o.haddr - o); x = cpu_to_be64(x); x <<= o * 8; @@ -2137,7 +2128,7 @@ static Int128 do_ld_whole_be16(CPUState *cpu, uintptr_t ra, MMULookupPageData *p, uint64_t ret_be) { int o = p->addr & 15; - Int128 x, y = load_atomic16_or_exit(cpu, ra, p->haddr - o); + Int128 x, y = load_atomic16_or_exit(cpu, ra, p->o.haddr - o); int size = p->size; if (!HOST_BIG_ENDIAN) { @@ -2160,8 +2151,8 @@ static uint64_t do_ld_beN(CPUState *cpu, MMULookupPageData *p, MemOp atom; unsigned tmp, half_size; - if (unlikely(p->flags & TLB_MMIO)) { - return do_ld_mmio_beN(cpu, p->full, ret_be, p->addr, p->size, + if (unlikely(p->o.flags & TLB_MMIO)) { + return do_ld_mmio_beN(cpu, &p->o.full, ret_be, p->addr, p->size, mmu_idx, type, ra); } @@ -2210,8 +2201,9 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p, uint64_t b; MemOp atom; - if (unlikely(p->flags & TLB_MMIO)) { - return do_ld16_mmio_beN(cpu, p->full, a, p->addr, size, mmu_idx, ra); + if (unlikely(p->o.flags & TLB_MMIO)) { + return do_ld16_mmio_beN(cpu, &p->o.full, a, p->addr, + size, mmu_idx, ra); } /* @@ -2223,7 +2215,7 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p, case MO_ATOM_SUBALIGN: p->size = size - 8; a = do_ld_parts_beN(p, a); - p->haddr += size - 8; + p->o.haddr += size - 8; p->size = 8; b = do_ld_parts_beN(p, 0); break; @@ -2242,7 +2234,7 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p, case MO_ATOM_NONE: p->size = size - 8; a = do_ld_bytes_beN(p, a); - b = ldq_be_p(p->haddr + size - 8); + b = ldq_be_p(p->o.haddr + size - 8); break; default: @@ -2255,10 +2247,11 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p, static uint8_t do_ld_1(CPUState *cpu, MMULookupPageData *p, int mmu_idx, MMUAccessType type, uintptr_t ra) { - if (unlikely(p->flags & TLB_MMIO)) { - return do_ld_mmio_beN(cpu, p->full, 0, p->addr, 1, mmu_idx, type, ra); + if (unlikely(p->o.flags & TLB_MMIO)) { + return do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 1, + mmu_idx, type, ra); } else { - return *(uint8_t *)p->haddr; + return *(uint8_t *)p->o.haddr; } } @@ -2267,14 +2260,15 @@ static uint16_t do_ld_2(CPUState *cpu, MMULookupPageData *p, int mmu_idx, { uint16_t ret; - if (unlikely(p->flags & TLB_MMIO)) { - ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 2, mmu_idx, type, ra); + if (unlikely(p->o.flags & TLB_MMIO)) { + ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 2, + mmu_idx, type, ra); if ((memop & MO_BSWAP) == MO_LE) { ret = bswap16(ret); } } else { /* Perform the load host endian, then swap if necessary. */ - ret = load_atom_2(cpu, ra, p->haddr, memop); + ret = load_atom_2(cpu, ra, p->o.haddr, memop); if (memop & MO_BSWAP) { ret = bswap16(ret); } @@ -2287,14 +2281,15 @@ static uint32_t do_ld_4(CPUState *cpu, MMULookupPageData *p, int mmu_idx, { uint32_t ret; - if (unlikely(p->flags & TLB_MMIO)) { - ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 4, mmu_idx, type, ra); + if (unlikely(p->o.flags & TLB_MMIO)) { + ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 4, + mmu_idx, type, ra); if ((memop & MO_BSWAP) == MO_LE) { ret = bswap32(ret); } } else { /* Perform the load host endian. */ - ret = load_atom_4(cpu, ra, p->haddr, memop); + ret = load_atom_4(cpu, ra, p->o.haddr, memop); if (memop & MO_BSWAP) { ret = bswap32(ret); } @@ -2307,14 +2302,15 @@ static uint64_t do_ld_8(CPUState *cpu, MMULookupPageData *p, int mmu_idx, { uint64_t ret; - if (unlikely(p->flags & TLB_MMIO)) { - ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 8, mmu_idx, type, ra); + if (unlikely(p->o.flags & TLB_MMIO)) { + ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 8, + mmu_idx, type, ra); if ((memop & MO_BSWAP) == MO_LE) { ret = bswap64(ret); } } else { /* Perform the load host endian. */ - ret = load_atom_8(cpu, ra, p->haddr, memop); + ret = load_atom_8(cpu, ra, p->o.haddr, memop); if (memop & MO_BSWAP) { ret = bswap64(ret); } @@ -2414,15 +2410,15 @@ static Int128 do_ld16_mmu(CPUState *cpu, vaddr addr, cpu_req_mo(TCG_MO_LD_LD | TCG_MO_ST_LD); crosspage = mmu_lookup(cpu, addr, oi, ra, MMU_DATA_LOAD, &l); if (likely(!crosspage)) { - if (unlikely(l.page[0].flags & TLB_MMIO)) { - ret = do_ld16_mmio_beN(cpu, l.page[0].full, 0, addr, 16, + if (unlikely(l.page[0].o.flags & TLB_MMIO)) { + ret = do_ld16_mmio_beN(cpu, &l.page[0].o.full, 0, addr, 16, l.mmu_idx, ra); if ((l.memop & MO_BSWAP) == MO_LE) { ret = bswap128(ret); } } else { /* Perform the load host endian. */ - ret = load_atom_16(cpu, ra, l.page[0].haddr, l.memop); + ret = load_atom_16(cpu, ra, l.page[0].o.haddr, l.memop); if (l.memop & MO_BSWAP) { ret = bswap128(ret); } @@ -2568,10 +2564,10 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p, MemOp atom; unsigned tmp, half_size; - if (unlikely(p->flags & TLB_MMIO)) { - return do_st_mmio_leN(cpu, p->full, val_le, p->addr, + if (unlikely(p->o.flags & TLB_MMIO)) { + return do_st_mmio_leN(cpu, &p->o.full, val_le, p->addr, p->size, mmu_idx, ra); - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { return val_le >> (p->size * 8); } @@ -2582,7 +2578,7 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p, atom = mop & MO_ATOM_MASK; switch (atom) { case MO_ATOM_SUBALIGN: - return store_parts_leN(p->haddr, p->size, val_le); + return store_parts_leN(p->o.haddr, p->size, val_le); case MO_ATOM_IFALIGN_PAIR: case MO_ATOM_WITHIN16_PAIR: @@ -2593,9 +2589,9 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p, ? p->size == half_size : p->size >= half_size) { if (!HAVE_al8_fast && p->size <= 4) { - return store_whole_le4(p->haddr, p->size, val_le); + return store_whole_le4(p->o.haddr, p->size, val_le); } else if (HAVE_al8) { - return store_whole_le8(p->haddr, p->size, val_le); + return store_whole_le8(p->o.haddr, p->size, val_le); } else { cpu_loop_exit_atomic(cpu, ra); } @@ -2605,7 +2601,7 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p, case MO_ATOM_IFALIGN: case MO_ATOM_WITHIN16: case MO_ATOM_NONE: - return store_bytes_leN(p->haddr, p->size, val_le); + return store_bytes_leN(p->o.haddr, p->size, val_le); default: g_assert_not_reached(); @@ -2622,10 +2618,10 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, int size = p->size; MemOp atom; - if (unlikely(p->flags & TLB_MMIO)) { - return do_st16_mmio_leN(cpu, p->full, val_le, p->addr, + if (unlikely(p->o.flags & TLB_MMIO)) { + return do_st16_mmio_leN(cpu, &p->o.full, val_le, p->addr, size, mmu_idx, ra); - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { return int128_gethi(val_le) >> ((size - 8) * 8); } @@ -2636,8 +2632,8 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, atom = mop & MO_ATOM_MASK; switch (atom) { case MO_ATOM_SUBALIGN: - store_parts_leN(p->haddr, 8, int128_getlo(val_le)); - return store_parts_leN(p->haddr + 8, p->size - 8, + store_parts_leN(p->o.haddr, 8, int128_getlo(val_le)); + return store_parts_leN(p->o.haddr + 8, p->size - 8, int128_gethi(val_le)); case MO_ATOM_WITHIN16_PAIR: @@ -2645,7 +2641,7 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, if (!HAVE_CMPXCHG128) { cpu_loop_exit_atomic(cpu, ra); } - return store_whole_le16(p->haddr, p->size, val_le); + return store_whole_le16(p->o.haddr, p->size, val_le); case MO_ATOM_IFALIGN_PAIR: /* @@ -2655,8 +2651,8 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, case MO_ATOM_IFALIGN: case MO_ATOM_WITHIN16: case MO_ATOM_NONE: - stq_le_p(p->haddr, int128_getlo(val_le)); - return store_bytes_leN(p->haddr + 8, p->size - 8, + stq_le_p(p->o.haddr, int128_getlo(val_le)); + return store_bytes_leN(p->o.haddr + 8, p->size - 8, int128_gethi(val_le)); default: @@ -2667,69 +2663,69 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, static void do_st_1(CPUState *cpu, MMULookupPageData *p, uint8_t val, int mmu_idx, uintptr_t ra) { - if (unlikely(p->flags & TLB_MMIO)) { - do_st_mmio_leN(cpu, p->full, val, p->addr, 1, mmu_idx, ra); - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { + if (unlikely(p->o.flags & TLB_MMIO)) { + do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 1, mmu_idx, ra); + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { /* nothing */ } else { - *(uint8_t *)p->haddr = val; + *(uint8_t *)p->o.haddr = val; } } static void do_st_2(CPUState *cpu, MMULookupPageData *p, uint16_t val, int mmu_idx, MemOp memop, uintptr_t ra) { - if (unlikely(p->flags & TLB_MMIO)) { + if (unlikely(p->o.flags & TLB_MMIO)) { if ((memop & MO_BSWAP) != MO_LE) { val = bswap16(val); } - do_st_mmio_leN(cpu, p->full, val, p->addr, 2, mmu_idx, ra); - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { + do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 2, mmu_idx, ra); + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { /* nothing */ } else { /* Swap to host endian if necessary, then store. */ if (memop & MO_BSWAP) { val = bswap16(val); } - store_atom_2(cpu, ra, p->haddr, memop, val); + store_atom_2(cpu, ra, p->o.haddr, memop, val); } } static void do_st_4(CPUState *cpu, MMULookupPageData *p, uint32_t val, int mmu_idx, MemOp memop, uintptr_t ra) { - if (unlikely(p->flags & TLB_MMIO)) { + if (unlikely(p->o.flags & TLB_MMIO)) { if ((memop & MO_BSWAP) != MO_LE) { val = bswap32(val); } - do_st_mmio_leN(cpu, p->full, val, p->addr, 4, mmu_idx, ra); - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { + do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 4, mmu_idx, ra); + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { /* nothing */ } else { /* Swap to host endian if necessary, then store. */ if (memop & MO_BSWAP) { val = bswap32(val); } - store_atom_4(cpu, ra, p->haddr, memop, val); + store_atom_4(cpu, ra, p->o.haddr, memop, val); } } static void do_st_8(CPUState *cpu, MMULookupPageData *p, uint64_t val, int mmu_idx, MemOp memop, uintptr_t ra) { - if (unlikely(p->flags & TLB_MMIO)) { + if (unlikely(p->o.flags & TLB_MMIO)) { if ((memop & MO_BSWAP) != MO_LE) { val = bswap64(val); } - do_st_mmio_leN(cpu, p->full, val, p->addr, 8, mmu_idx, ra); - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { + do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 8, mmu_idx, ra); + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { /* nothing */ } else { /* Swap to host endian if necessary, then store. */ if (memop & MO_BSWAP) { val = bswap64(val); } - store_atom_8(cpu, ra, p->haddr, memop, val); + store_atom_8(cpu, ra, p->o.haddr, memop, val); } } @@ -2822,19 +2818,20 @@ static void do_st16_mmu(CPUState *cpu, vaddr addr, Int128 val, cpu_req_mo(TCG_MO_LD_ST | TCG_MO_ST_ST); crosspage = mmu_lookup(cpu, addr, oi, ra, MMU_DATA_STORE, &l); if (likely(!crosspage)) { - if (unlikely(l.page[0].flags & TLB_MMIO)) { + if (unlikely(l.page[0].o.flags & TLB_MMIO)) { if ((l.memop & MO_BSWAP) != MO_LE) { val = bswap128(val); } - do_st16_mmio_leN(cpu, l.page[0].full, val, addr, 16, l.mmu_idx, ra); - } else if (unlikely(l.page[0].flags & TLB_DISCARD_WRITE)) { + do_st16_mmio_leN(cpu, &l.page[0].o.full, val, addr, + 16, l.mmu_idx, ra); + } else if (unlikely(l.page[0].o.flags & TLB_DISCARD_WRITE)) { /* nothing */ } else { /* Swap to host endian if necessary, then store. */ if (l.memop & MO_BSWAP) { val = bswap128(val); } - store_atom_16(cpu, ra, l.page[0].haddr, l.memop, val); + store_atom_16(cpu, ra, l.page[0].o.haddr, l.memop, val); } return; }
Signed-off-by: Richard Henderson <richard.henderson@linaro.org> --- accel/tcg/cputlb.c | 151 ++++++++++++++++++++++----------------------- 1 file changed, 74 insertions(+), 77 deletions(-)