Message ID | 20240715095702.1222213-5-ethan84@andestech.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | Support RISC-V IOPMP | expand |
On Mon, Jul 15, 2024 at 7:58 PM Ethan Chen via <qemu-devel@nongnu.org> wrote: > > Support basic functions of IOPMP specification v0.9.1 rapid-k model. > The specification url: > https://github.com/riscv-non-isa/iopmp-spec/releases/tag/v0.9.1 > > The IOPMP checks whether memory access from a device or CPU is valid. > This implementation uses an IOMMU to modify the address space accessed > by the device. > > For device access with IOMMUAccessFlags specifying read or write > (IOMMU_RO or IOMMU_WO), the IOPMP checks the permission in > iopmp_translate. If the access is valid, the target address space is > downstream_as. If the access is blocked, it will be redirected to > blocked_rwx_as. > > For CPU access with IOMMUAccessFlags not specifying read or write > (IOMMU_NONE), the IOPMP translates the access to the corresponding > address space based on the permission. If the access has full permission > (rwx), the target address space is downstream_as. If the access has > limited permission, the target address space is blocked_ followed by > the lacked permissions. > > The operation of a blocked region can trigger an IOPMP interrupt, a bus > error, or it can respond with success and fabricated data, depending on > the value of the IOPMP ERR_CFG register. > > Signed-off-by: Ethan Chen <ethan84@andestech.com> > --- > hw/misc/Kconfig | 3 + > hw/misc/meson.build | 1 + > hw/misc/riscv_iopmp.c | 1154 +++++++++++++++++++++++++++++++++ > hw/misc/trace-events | 3 + > include/hw/misc/riscv_iopmp.h | 168 +++++ > 5 files changed, 1329 insertions(+) > create mode 100644 hw/misc/riscv_iopmp.c > create mode 100644 include/hw/misc/riscv_iopmp.h > > diff --git a/hw/misc/Kconfig b/hw/misc/Kconfig > index 1e08785b83..427f0c702e 100644 > --- a/hw/misc/Kconfig > +++ b/hw/misc/Kconfig > @@ -213,4 +213,7 @@ config IOSB > config XLNX_VERSAL_TRNG > bool > > +config RISCV_IOPMP > + bool > + > source macio/Kconfig > diff --git a/hw/misc/meson.build b/hw/misc/meson.build > index 2ca8717be2..d9006e1d81 100644 > --- a/hw/misc/meson.build > +++ b/hw/misc/meson.build > @@ -34,6 +34,7 @@ system_ss.add(when: 'CONFIG_SIFIVE_E_PRCI', if_true: files('sifive_e_prci.c')) > system_ss.add(when: 'CONFIG_SIFIVE_E_AON', if_true: files('sifive_e_aon.c')) > system_ss.add(when: 'CONFIG_SIFIVE_U_OTP', if_true: files('sifive_u_otp.c')) > system_ss.add(when: 'CONFIG_SIFIVE_U_PRCI', if_true: files('sifive_u_prci.c')) > +specific_ss.add(when: 'CONFIG_RISCV_IOPMP', if_true: files('riscv_iopmp.c')) > > subdir('macio') > > diff --git a/hw/misc/riscv_iopmp.c b/hw/misc/riscv_iopmp.c > new file mode 100644 > index 0000000000..db43e3c73f > --- /dev/null > +++ b/hw/misc/riscv_iopmp.c > @@ -0,0 +1,1154 @@ > +/* > + * QEMU RISC-V IOPMP (Input Output Physical Memory Protection) > + * > + * Copyright (c) 2023-2024 Andes Tech. Corp. > + * > + * SPDX-License-Identifier: GPL-2.0-or-later > + * > + * This program is free software; you can redistribute it and/or modify it > + * under the terms and conditions of the GNU General Public License, > + * version 2 or later, as published by the Free Software Foundation. > + * > + * This program is distributed in the hope it will be useful, but WITHOUT > + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or > + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for > + * more details. > + * > + * You should have received a copy of the GNU General Public License along with > + * this program. If not, see <http://www.gnu.org/licenses/>. > + */ > + > +#include "qemu/osdep.h" > +#include "qemu/log.h" > +#include "qapi/error.h" > +#include "trace.h" > +#include "exec/exec-all.h" > +#include "exec/address-spaces.h" > +#include "hw/qdev-properties.h" > +#include "hw/sysbus.h" > +#include "hw/misc/riscv_iopmp.h" > +#include "memory.h" > +#include "hw/irq.h" > +#include "hw/registerfields.h" > +#include "trace.h" > + > +#define TYPE_IOPMP_IOMMU_MEMORY_REGION "iopmp-iommu-memory-region" > + > +REG32(VERSION, 0x00) > + FIELD(VERSION, VENDOR, 0, 24) > + FIELD(VERSION, SPECVER , 24, 8) > +REG32(IMP, 0x04) > + FIELD(IMP, IMPID, 0, 32) > +REG32(HWCFG0, 0x08) > + FIELD(HWCFG0, MODEL, 0, 4) > + FIELD(HWCFG0, TOR_EN, 4, 1) > + FIELD(HWCFG0, SPS_EN, 5, 1) > + FIELD(HWCFG0, USER_CFG_EN, 6, 1) > + FIELD(HWCFG0, PRIENT_PROG, 7, 1) > + FIELD(HWCFG0, RRID_TRANSL_EN, 8, 1) > + FIELD(HWCFG0, RRID_TRANSL_PROG, 9, 1) > + FIELD(HWCFG0, CHK_X, 10, 1) > + FIELD(HWCFG0, NO_X, 11, 1) > + FIELD(HWCFG0, NO_W, 12, 1) > + FIELD(HWCFG0, STALL_EN, 13, 1) > + FIELD(HWCFG0, PEIS, 14, 1) > + FIELD(HWCFG0, PEES, 15, 1) > + FIELD(HWCFG0, MFR_EN, 16, 1) > + FIELD(HWCFG0, MD_NUM, 24, 7) > + FIELD(HWCFG0, ENABLE, 31, 1) > +REG32(HWCFG1, 0x0C) > + FIELD(HWCFG1, RRID_NUM, 0, 16) > + FIELD(HWCFG1, ENTRY_NUM, 16, 16) > +REG32(HWCFG2, 0x10) > + FIELD(HWCFG2, PRIO_ENTRY, 0, 16) > + FIELD(HWCFG2, RRID_TRANSL, 16, 16) > +REG32(ENTRYOFFSET, 0x14) > + FIELD(ENTRYOFFSET, OFFSET, 0, 32) > +REG32(MDSTALL, 0x30) > + FIELD(MDSTALL, EXEMPT, 0, 1) > + FIELD(MDSTALL, MD, 1, 31) > +REG32(MDSTALLH, 0x34) > + FIELD(MDSTALLH, MD, 0, 32) > +REG32(RRIDSCP, 0x38) > + FIELD(RRIDSCP, RRID, 0, 16) > + FIELD(RRIDSCP, OP, 30, 2) > +REG32(MDLCK, 0x40) > + FIELD(MDLCK, L, 0, 1) > + FIELD(MDLCK, MD, 1, 31) > +REG32(MDLCKH, 0x44) > + FIELD(MDLCKH, MDH, 0, 32) > +REG32(MDCFGLCK, 0x48) > + FIELD(MDCFGLCK, L, 0, 1) > + FIELD(MDCFGLCK, F, 1, 7) > +REG32(ENTRYLCK, 0x4C) > + FIELD(ENTRYLCK, L, 0, 1) > + FIELD(ENTRYLCK, F, 1, 16) > +REG32(ERR_CFG, 0x60) > + FIELD(ERR_CFG, L, 0, 1) > + FIELD(ERR_CFG, IE, 1, 1) > + FIELD(ERR_CFG, IRE, 2, 1) > + FIELD(ERR_CFG, IWE, 3, 1) > + FIELD(ERR_CFG, IXE, 4, 1) > + FIELD(ERR_CFG, RRE, 5, 1) > + FIELD(ERR_CFG, RWE, 6, 1) > + FIELD(ERR_CFG, RXE, 7, 1) > +REG32(ERR_REQINFO, 0x64) > + FIELD(ERR_REQINFO, V, 0, 1) > + FIELD(ERR_REQINFO, TTYPE, 1, 2) > + FIELD(ERR_REQINFO, ETYPE, 4, 3) > + FIELD(ERR_REQINFO, SVC, 7, 1) > +REG32(ERR_REQADDR, 0x68) > + FIELD(ERR_REQADDR, ADDR, 0, 32) > +REG32(ERR_REQADDRH, 0x6C) > + FIELD(ERR_REQADDRH, ADDRH, 0, 32) > +REG32(ERR_REQID, 0x70) > + FIELD(ERR_REQID, RRID, 0, 16) > + FIELD(ERR_REQID, EID, 16, 16) > +REG32(ERR_MFR, 0x74) > + FIELD(ERR_MFR, SVW, 0, 16) > + FIELD(ERR_MFR, SVI, 16, 12) > + FIELD(ERR_MFR, SVS, 31, 1) > +REG32(MDCFG0, 0x800) > + FIELD(MDCFG0, T, 0, 16) > +REG32(SRCMD_EN0, 0x1000) > + FIELD(SRCMD_EN0, L, 0, 1) > + FIELD(SRCMD_EN0, MD, 1, 31) > +REG32(SRCMD_ENH0, 0x1004) > + FIELD(SRCMD_ENH0, MDH, 0, 32) > +REG32(SRCMD_R0, 0x1008) > + FIELD(SRCMD_R0, MD, 1, 31) > +REG32(SRCMD_RH0, 0x100C) > + FIELD(SRCMD_RH0, MDH, 0, 32) > +REG32(SRCMD_W0, 0x1010) > + FIELD(SRCMD_W0, MD, 1, 31) > +REG32(SRCMD_WH0, 0x1014) > + FIELD(SRCMD_WH0, MDH, 0, 32) > + > +FIELD(ENTRY_ADDR, ADDR, 0, 32) > +FIELD(ENTRY_ADDRH, ADDRH, 0, 32) > + > +FIELD(ENTRY_CFG, R, 0, 1) > +FIELD(ENTRY_CFG, W, 1, 1) > +FIELD(ENTRY_CFG, X, 2, 1) > +FIELD(ENTRY_CFG, A, 3, 2) > +FIELD(ENTRY_CFG, SIRE, 5, 1) > +FIELD(ENTRY_CFG, SIWE, 6, 1) > +FIELD(ENTRY_CFG, SIXE, 7, 1) > +FIELD(ENTRY_CFG, SERE, 8, 1) > +FIELD(ENTRY_CFG, SEWE, 9, 1) > +FIELD(ENTRY_CFG, SEXE, 10, 1) > + > +FIELD(ENTRY_USER_CFG, IM, 0, 32) > + > +/* Offsets to SRCMD_EN(i) */ > +#define SRCMD_EN_OFFSET 0x0 > +#define SRCMD_ENH_OFFSET 0x4 > +#define SRCMD_R_OFFSET 0x8 > +#define SRCMD_RH_OFFSET 0xC > +#define SRCMD_W_OFFSET 0x10 > +#define SRCMD_WH_OFFSET 0x14 > + > +/* Offsets to ENTRY_ADDR(i) */ > +#define ENTRY_ADDR_OFFSET 0x0 > +#define ENTRY_ADDRH_OFFSET 0x4 > +#define ENTRY_CFG_OFFSET 0x8 > +#define ENTRY_USER_CFG_OFFSET 0xC > + > +/* Memmap for parallel IOPMPs */ > +typedef struct iopmp_protection_memmap { > + MemMapEntry entry; > + IopmpState *iopmp_s; > + QLIST_ENTRY(iopmp_protection_memmap) list; > +} iopmp_protection_memmap; > +QLIST_HEAD(, iopmp_protection_memmap) > + iopmp_protection_memmaps = QLIST_HEAD_INITIALIZER(iopmp_protection_memmaps); > + > +static void iopmp_iommu_notify(IopmpState *s) > +{ > + IOMMUTLBEvent event = { > + .entry = { > + .iova = 0, > + .translated_addr = 0, > + .addr_mask = -1ULL, > + .perm = IOMMU_NONE, > + }, > + .type = IOMMU_NOTIFIER_UNMAP, > + }; > + > + for (int i = 0; i < s->rrid_num; i++) { > + memory_region_notify_iommu(&s->iommu, i, event); > + } > +} > + > +static void iopmp_decode_napot(uint64_t a, uint64_t *sa, > + uint64_t *ea) > +{ > + /* > + * aaaa...aaa0 8-byte NAPOT range > + * aaaa...aa01 16-byte NAPOT range > + * aaaa...a011 32-byte NAPOT range > + * ... > + * aa01...1111 2^XLEN-byte NAPOT range > + * a011...1111 2^(XLEN+1)-byte NAPOT range > + * 0111...1111 2^(XLEN+2)-byte NAPOT range > + * 1111...1111 Reserved > + */ > + > + a = (a << 2) | 0x3; > + *sa = a & (a + 1); > + *ea = a | (a + 1); > +} > + > +static void iopmp_update_rule(IopmpState *s, uint32_t entry_index) > +{ > + uint8_t this_cfg = s->regs.entry[entry_index].cfg_reg; > + uint64_t this_addr = s->regs.entry[entry_index].addr_reg | > + ((uint64_t)s->regs.entry[entry_index].addrh_reg << 32); > + uint64_t prev_addr = 0u; > + uint64_t sa = 0u; > + uint64_t ea = 0u; > + > + if (entry_index >= 1u) { > + prev_addr = s->regs.entry[entry_index - 1].addr_reg | > + ((uint64_t)s->regs.entry[entry_index - 1].addrh_reg << 32); > + } > + > + switch (FIELD_EX32(this_cfg, ENTRY_CFG, A)) { > + case IOPMP_AMATCH_OFF: > + sa = 0u; > + ea = -1; > + break; > + > + case IOPMP_AMATCH_TOR: > + sa = (prev_addr) << 2; /* shift up from [xx:0] to [xx+2:2] */ > + ea = ((this_addr) << 2) - 1u; > + if (sa > ea) { > + sa = ea = 0u; > + } > + break; > + > + case IOPMP_AMATCH_NA4: > + sa = this_addr << 2; /* shift up from [xx:0] to [xx+2:2] */ > + ea = (sa + 4u) - 1u; > + break; > + > + case IOPMP_AMATCH_NAPOT: > + iopmp_decode_napot(this_addr, &sa, &ea); > + break; > + > + default: > + sa = 0u; > + ea = 0u; > + break; > + } > + > + s->entry_addr[entry_index].sa = sa; > + s->entry_addr[entry_index].ea = ea; > + iopmp_iommu_notify(s); > +} > + > +static uint64_t iopmp_read(void *opaque, hwaddr addr, unsigned size) > +{ > + IopmpState *s = IOPMP(opaque); > + uint32_t rz = 0; > + uint32_t offset, idx; > + > + switch (addr) { > + case A_VERSION: > + rz = VENDER_VIRT << R_VERSION_VENDOR_SHIFT | > + SPECVER_0_9_1 << R_VERSION_SPECVER_SHIFT; It would be better to use the FIELD_DP32() macro instead of the manual shifts > + break; > + case A_IMP: > + rz = IMPID_0_9_1; > + break; > + case A_HWCFG0: > + rz = s->model << R_HWCFG0_MODEL_SHIFT | > + 1 << R_HWCFG0_TOR_EN_SHIFT | > + 0 << R_HWCFG0_SPS_EN_SHIFT | > + 0 << R_HWCFG0_USER_CFG_EN_SHIFT | > + s->prient_prog << R_HWCFG0_PRIENT_PROG_SHIFT | > + 0 << R_HWCFG0_RRID_TRANSL_EN_SHIFT | > + 0 << R_HWCFG0_RRID_TRANSL_PROG_SHIFT | > + 1 << R_HWCFG0_CHK_X_SHIFT | > + 0 << R_HWCFG0_NO_X_SHIFT | > + 0 << R_HWCFG0_NO_W_SHIFT | > + 0 << R_HWCFG0_STALL_EN_SHIFT | > + 0 << R_HWCFG0_PEIS_SHIFT | > + 0 << R_HWCFG0_PEES_SHIFT | > + 0 << R_HWCFG0_MFR_EN_SHIFT | > + s->md_num << R_HWCFG0_MD_NUM_SHIFT | > + s->enable << R_HWCFG0_ENABLE_SHIFT ; > + break; > + case A_HWCFG1: > + rz = s->rrid_num << R_HWCFG1_RRID_NUM_SHIFT | > + s->entry_num << R_HWCFG1_ENTRY_NUM_SHIFT; > + break; > + case A_HWCFG2: > + rz = s->prio_entry << R_HWCFG2_PRIO_ENTRY_SHIFT; > + break; > + case A_ENTRYOFFSET: > + rz = s->entry_offset; > + break; > + case A_ERR_CFG: > + rz = s->regs.err_cfg; > + break; > + case A_MDLCK: > + rz = s->regs.mdlck; > + break; > + case A_MDLCKH: > + rz = s->regs.mdlckh; > + break; > + case A_MDCFGLCK: > + rz = s->regs.mdcfglck; > + break; > + case A_ENTRYLCK: > + rz = s->regs.entrylck; > + break; > + case A_ERR_REQADDR: > + rz = s->regs.err_reqaddr & UINT32_MAX; > + break; > + case A_ERR_REQADDRH: > + rz = s->regs.err_reqaddr >> 32; > + break; > + case A_ERR_REQID: > + rz = s->regs.err_reqid; > + break; > + case A_ERR_REQINFO: > + rz = s->regs.err_reqinfo; > + break; > + > + default: > + if (addr >= A_MDCFG0 && > + addr < A_MDCFG0 + 4 * (s->md_num - 1)) { > + offset = addr - A_MDCFG0; > + idx = offset >> 2; > + if (idx == 0 && offset == 0) { > + rz = s->regs.mdcfg[idx]; > + } else { > + /* Only MDCFG0 is implemented in rapid-k model */ > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > + __func__, (int)addr); > + } > + } else if (addr >= A_SRCMD_EN0 && > + addr < A_SRCMD_WH0 + 32 * (s->rrid_num - 1)) { > + offset = addr - A_SRCMD_EN0; > + idx = offset >> 5; > + offset &= 0x1f; > + > + switch (offset) { > + case SRCMD_EN_OFFSET: > + rz = s->regs.srcmd_en[idx]; > + break; > + case SRCMD_ENH_OFFSET: > + rz = s->regs.srcmd_enh[idx]; > + break; > + default: > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > + __func__, (int)addr); > + break; > + } > + } else if (addr >= s->entry_offset && > + addr < s->entry_offset + ENTRY_USER_CFG_OFFSET + > + 16 * (s->entry_num - 1)) { > + offset = addr - s->entry_offset; > + idx = offset >> 4; > + offset &= 0xf; > + > + switch (offset) { > + case ENTRY_ADDR_OFFSET: > + rz = s->regs.entry[idx].addr_reg; > + break; > + case ENTRY_ADDRH_OFFSET: > + rz = s->regs.entry[idx].addrh_reg; > + break; > + case ENTRY_CFG_OFFSET: > + rz = s->regs.entry[idx].cfg_reg; > + break; > + case ENTRY_USER_CFG_OFFSET: > + /* Does not support user customized permission */ > + rz = 0; > + break; > + default: > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > + __func__, (int)addr); > + break; > + } > + } else { > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > + __func__, (int)addr); > + } > + break; > + } > + trace_iopmp_read(addr, rz); > + return rz; > +} > + > +static void > +iopmp_write(void *opaque, hwaddr addr, uint64_t value, unsigned size) > +{ > + IopmpState *s = IOPMP(opaque); > + uint32_t offset, idx; > + uint32_t value32 = value; > + > + trace_iopmp_write(addr, value32); > + > + switch (addr) { > + case A_VERSION: /* RO */ > + break; > + case A_IMP: /* RO */ > + break; > + case A_HWCFG0: > + if (FIELD_EX32(value32, HWCFG0, PRIENT_PROG)) { > + /* W1C */ > + s->prient_prog = 0; > + } > + if (FIELD_EX32(value32, HWCFG0, ENABLE)) { > + /* W1S */ > + s->enable = 1; > + iopmp_iommu_notify(s); > + } > + break; > + case A_HWCFG1: /* RO */ > + break; > + case A_HWCFG2: > + if (s->prient_prog) { > + s->prio_entry = FIELD_EX32(value32, HWCFG2, PRIO_ENTRY); > + } > + break; > + case A_ERR_CFG: > + if (!FIELD_EX32(s->regs.err_cfg, ERR_CFG, L)) { > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, L, > + FIELD_EX32(value32, ERR_CFG, L)); > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IE, > + FIELD_EX32(value32, ERR_CFG, IE)); > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IRE, > + FIELD_EX32(value32, ERR_CFG, IRE)); > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RRE, > + FIELD_EX32(value32, ERR_CFG, RRE)); > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IWE, > + FIELD_EX32(value32, ERR_CFG, IWE)); > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RWE, > + FIELD_EX32(value32, ERR_CFG, RWE)); > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IXE, > + FIELD_EX32(value32, ERR_CFG, IXE)); > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RXE, > + FIELD_EX32(value32, ERR_CFG, RXE)); > + } > + break; > + case A_MDLCK: > + if (!FIELD_EX32(s->regs.mdlck, MDLCK, L)) { > + s->regs.mdlck = value32; > + } > + break; > + case A_MDLCKH: > + if (!FIELD_EX32(s->regs.mdlck, MDLCK, L)) { > + s->regs.mdlckh = value32; > + } > + break; > + case A_MDCFGLCK: > + if (!FIELD_EX32(s->regs.mdcfglck, MDCFGLCK, L)) { > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, > + FIELD_EX32(value32, MDCFGLCK, F)); > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, > + FIELD_EX32(value32, MDCFGLCK, L)); > + } > + break; > + case A_ENTRYLCK: > + if (!(FIELD_EX32(s->regs.entrylck, ENTRYLCK, L))) { > + s->regs.entrylck = FIELD_DP32(s->regs.entrylck, ENTRYLCK, F, > + FIELD_EX32(value32, ENTRYLCK, F)); > + s->regs.entrylck = FIELD_DP32(s->regs.entrylck, ENTRYLCK, L, > + FIELD_EX32(value32, ENTRYLCK, L)); > + } > + case A_ERR_REQADDR: /* RO */ > + break; > + case A_ERR_REQADDRH: /* RO */ > + break; > + case A_ERR_REQID: /* RO */ > + break; > + case A_ERR_REQINFO: > + if (FIELD_EX32(value32, ERR_REQINFO, V)) { > + s->regs.err_reqinfo = FIELD_DP32(s->regs.err_reqinfo, > + ERR_REQINFO, V, 0); > + qemu_set_irq(s->irq, 0); > + } > + break; > + > + default: > + if (addr >= A_MDCFG0 && > + addr < A_MDCFG0 + 4 * (s->md_num - 1)) { > + offset = addr - A_MDCFG0; > + idx = offset >> 2; > + /* RO in rapid-k model */ > + if (idx > 0) { > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > + __func__, (int)addr); > + } > + } else if (addr >= A_SRCMD_EN0 && > + addr < A_SRCMD_WH0 + 32 * (s->rrid_num - 1)) { > + offset = addr - A_SRCMD_EN0; > + idx = offset >> 5; > + offset &= 0x1f; > + > + if (offset % 4) { > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > + __func__, (int)addr); > + } else if (FIELD_EX32(s->regs.srcmd_en[idx], SRCMD_EN0, L) > + == 0) { > + switch (offset) { > + case SRCMD_EN_OFFSET: > + s->regs.srcmd_en[idx] = > + FIELD_DP32(s->regs.srcmd_en[idx], SRCMD_EN0, L, > + FIELD_EX32(value32, SRCMD_EN0, L)); > + > + /* MD field is protected by mdlck */ > + value32 = (value32 & ~s->regs.mdlck) | > + (s->regs.srcmd_en[idx] & s->regs.mdlck); > + s->regs.srcmd_en[idx] = > + FIELD_DP32(s->regs.srcmd_en[idx], SRCMD_EN0, MD, > + FIELD_EX32(value32, SRCMD_EN0, MD)); > + break; > + case SRCMD_ENH_OFFSET: > + value32 = (value32 & ~s->regs.mdlckh) | > + (s->regs.srcmd_enh[idx] & s->regs.mdlckh); > + s->regs.srcmd_enh[idx] = > + FIELD_DP32(s->regs.srcmd_enh[idx], SRCMD_ENH0, MDH, > + value32); > + break; > + default: > + break; > + } > + } > + } else if (addr >= s->entry_offset && > + addr < s->entry_offset + ENTRY_USER_CFG_OFFSET > + + 16 * (s->entry_num - 1)) { > + offset = addr - s->entry_offset; > + idx = offset >> 4; > + offset &= 0xf; > + > + /* index < ENTRYLCK_F is protected */ > + if (idx >= FIELD_EX32(s->regs.entrylck, ENTRYLCK, F)) { > + switch (offset) { > + case ENTRY_ADDR_OFFSET: > + s->regs.entry[idx].addr_reg = value32; > + break; > + case ENTRY_ADDRH_OFFSET: > + s->regs.entry[idx].addrh_reg = value32; > + break; > + case ENTRY_CFG_OFFSET: > + s->regs.entry[idx].cfg_reg = value32; > + break; > + case ENTRY_USER_CFG_OFFSET: > + /* Does not support user customized permission */ > + break; > + default: > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > + __func__, (int)addr); > + break; > + } > + iopmp_update_rule(s, idx); > + if (idx + 1 < s->entry_num && > + FIELD_EX32(s->regs.entry[idx + 1].cfg_reg, ENTRY_CFG, A) == > + IOPMP_AMATCH_TOR) { > + iopmp_update_rule(s, idx + 1); > + } > + } > + } else { > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", __func__, > + (int)addr); > + } > + } > +} > + > +/* Match entry in memory domain */ > +static int match_entry_md(IopmpState *s, int md_idx, hwaddr start_addr, > + hwaddr end_addr, int *entry_idx, > + int *prior_entry_in_tlb) > +{ > + int entry_idx_s, entry_idx_e; > + int result = ENTRY_NO_HIT; > + int i = 0; > + hwaddr tlb_sa = start_addr & ~(TARGET_PAGE_SIZE - 1); > + hwaddr tlb_ea = tlb_sa + TARGET_PAGE_SIZE - 1; > + > + entry_idx_s = md_idx * s->regs.mdcfg[0]; > + entry_idx_e = (md_idx + 1) * s->regs.mdcfg[0]; > + > + if (entry_idx_s >= s->entry_num) { > + return result; > + } > + if (entry_idx_e > s->entry_num) { > + entry_idx_e = s->entry_num; > + } > + i = entry_idx_s; > + for (i = entry_idx_s; i < entry_idx_e; i++) { > + if (FIELD_EX32(s->regs.entry[i].cfg_reg, ENTRY_CFG, A) == > + IOPMP_AMATCH_OFF) { > + continue; > + } > + if (start_addr >= s->entry_addr[i].sa && > + start_addr <= s->entry_addr[i].ea) { > + /* Check end address */ > + if (end_addr >= s->entry_addr[i].sa && > + end_addr <= s->entry_addr[i].ea) { > + *entry_idx = i; > + return ENTRY_HIT; > + } else if (i >= s->prio_entry) { > + /* Continue for non-prio_entry */ > + continue; > + } else { > + *entry_idx = i; > + return ENTRY_PAR_HIT; > + } > + } else if (end_addr >= s->entry_addr[i].sa && > + end_addr <= s->entry_addr[i].ea) { > + /* Only end address matches the entry */ > + if (i >= s->prio_entry) { > + continue; > + } else { > + *entry_idx = i; > + return ENTRY_PAR_HIT; > + } > + } else if (start_addr < s->entry_addr[i].sa && > + end_addr > s->entry_addr[i].ea) { > + if (i >= s->prio_entry) { > + continue; > + } else { > + *entry_idx = i; > + return ENTRY_PAR_HIT; > + } > + } > + if (prior_entry_in_tlb != NULL) { > + if ((s->entry_addr[i].sa >= tlb_sa && > + s->entry_addr[i].sa <= tlb_ea) || > + (s->entry_addr[i].ea >= tlb_sa && > + s->entry_addr[i].ea <= tlb_ea)) { > + /* > + * TLB should not use the cached result when the tlb contains > + * higher priority entry > + */ > + *prior_entry_in_tlb = 1; > + } > + } > + } > + return result; > +} > + > +static int match_entry(IopmpState *s, int rrid, hwaddr start_addr, > + hwaddr end_addr, int *match_md_idx, > + int *match_entry_idx, int *prior_entry_in_tlb) > +{ > + int cur_result = ENTRY_NO_HIT; > + int result = ENTRY_NO_HIT; > + /* Remove lock bit */ > + uint64_t srcmd_en = ((uint64_t)s->regs.srcmd_en[rrid] | > + ((uint64_t)s->regs.srcmd_enh[rrid] << 32)) >> 1; > + > + for (int md_idx = 0; md_idx < s->md_num; md_idx++) { > + if (srcmd_en & (1ULL << md_idx)) { > + cur_result = match_entry_md(s, md_idx, start_addr, end_addr, > + match_entry_idx, prior_entry_in_tlb); > + if (cur_result == ENTRY_HIT || cur_result == ENTRY_PAR_HIT) { > + *match_md_idx = md_idx; > + return cur_result; > + } > + } > + } > + return result; > +} > + > +static void iopmp_error_reaction(IopmpState *s, uint32_t id, hwaddr start, > + uint32_t info) > +{ > + if (!FIELD_EX32(s->regs.err_reqinfo, ERR_REQINFO, V)) { > + s->regs.err_reqinfo = info; > + s->regs.err_reqinfo = FIELD_DP32(s->regs.err_reqinfo, ERR_REQINFO, V, > + 1); > + s->regs.err_reqid = id; > + /* addr[LEN+2:2] */ > + s->regs.err_reqaddr = start >> 2; > + > + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_READ && > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IRE)) { > + qemu_set_irq(s->irq, 1); > + } > + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_WRITE && > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IWE)) { > + qemu_set_irq(s->irq, 1); > + } > + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_FETCH && > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IXE)) { > + qemu_set_irq(s->irq, 1); > + } > + } > +} > + > +static IOMMUTLBEntry iopmp_translate(IOMMUMemoryRegion *iommu, hwaddr addr, > + IOMMUAccessFlags flags, int iommu_idx) > +{ > + int rrid = iommu_idx; > + IopmpState *s = IOPMP(container_of(iommu, IopmpState, iommu)); > + hwaddr start_addr, end_addr; > + int entry_idx = -1; > + int md_idx = -1; > + int result; > + uint32_t error_info = 0; > + uint32_t error_id = 0; > + int prior_entry_in_tlb = 0; > + iopmp_permission iopmp_perm; > + IOMMUTLBEntry entry = { > + .target_as = &s->downstream_as, > + .iova = addr, > + .translated_addr = addr, > + .addr_mask = 0, > + .perm = IOMMU_NONE, > + }; > + > + if (!s->enable) { > + /* Bypass IOPMP */ > + entry.addr_mask = -1ULL, > + entry.perm = IOMMU_RW; > + return entry; > + } > + > + /* unknown RRID */ > + if (rrid >= s->rrid_num) { > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > + ERR_REQINFO_ETYPE_RRID); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > + iopmp_error_reaction(s, error_id, addr, error_info); > + entry.target_as = &s->blocked_rwx_as; > + entry.perm = IOMMU_RW; > + return entry; > + } > + > + if (s->transaction_state[rrid].supported == true) { > + start_addr = s->transaction_state[rrid].start_addr; > + end_addr = s->transaction_state[rrid].end_addr; > + } else { > + /* No transaction information, use the same address */ > + start_addr = addr; > + end_addr = addr; > + } > + > + result = match_entry(s, rrid, start_addr, end_addr, &md_idx, &entry_idx, > + &prior_entry_in_tlb); > + if (result == ENTRY_HIT) { > + entry.addr_mask = s->entry_addr[entry_idx].ea - > + s->entry_addr[entry_idx].sa; > + if (prior_entry_in_tlb) { > + /* Make TLB repeat iommu translation on every access */ I don't follow this, if we have a prior entry in the TLB cache we don't cache the accesses? > + entry.addr_mask = 0; > + } > + iopmp_perm = s->regs.entry[entry_idx].cfg_reg & IOPMP_RWX; > + if (flags) { > + if ((iopmp_perm & flags) == 0) { > + /* Permission denied */ > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > + ERR_REQINFO_ETYPE_READ + flags - 1); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > + iopmp_error_reaction(s, error_id, start_addr, error_info); > + entry.target_as = &s->blocked_rwx_as; > + entry.perm = IOMMU_RW; > + } else { > + entry.target_as = &s->downstream_as; > + entry.perm = iopmp_perm; > + } > + } else { > + /* CPU access with IOMMU_NONE flag */ > + if (iopmp_perm & IOPMP_XO) { > + if ((iopmp_perm & IOPMP_RW) == IOPMP_RW) { > + entry.target_as = &s->downstream_as; > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_RO) { > + entry.target_as = &s->blocked_w_as; > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_WO) { > + entry.target_as = &s->blocked_r_as; > + } else { > + entry.target_as = &s->blocked_rw_as; > + } > + } else { > + if ((iopmp_perm & IOPMP_RW) == IOMMU_RW) { > + entry.target_as = &s->blocked_x_as; > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_RO) { > + entry.target_as = &s->blocked_wx_as; > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_WO) { > + entry.target_as = &s->blocked_rx_as; > + } else { > + entry.target_as = &s->blocked_rwx_as; > + } > + } > + entry.perm = IOMMU_RW; > + } > + } else { > + if (flags) { > + if (result == ENTRY_PAR_HIT) { > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > + ERR_REQINFO_ETYPE_PARHIT); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > + iopmp_error_reaction(s, error_id, start_addr, error_info); > + } else { > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > + ERR_REQINFO_ETYPE_NOHIT); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > + iopmp_error_reaction(s, error_id, start_addr, error_info); > + } > + } > + /* CPU access with IOMMU_NONE flag no_hit or par_hit */ > + entry.target_as = &s->blocked_rwx_as; > + entry.perm = IOMMU_RW; > + } > + return entry; > +} > + > +static const MemoryRegionOps iopmp_ops = { > + .read = iopmp_read, > + .write = iopmp_write, > + .endianness = DEVICE_NATIVE_ENDIAN, > + .valid = {.min_access_size = 4, .max_access_size = 4} > +}; > + > +static MemTxResult iopmp_permssion_write(void *opaque, hwaddr addr, > + uint64_t value, unsigned size, > + MemTxAttrs attrs) > +{ > + IopmpState *s = IOPMP(opaque); > + return address_space_write(&s->downstream_as, addr, attrs, &value, size); > +} > + > +static MemTxResult iopmp_permssion_read(void *opaque, hwaddr addr, > + uint64_t *pdata, unsigned size, > + MemTxAttrs attrs) > +{ > + IopmpState *s = IOPMP(opaque); > + return address_space_read(&s->downstream_as, addr, attrs, pdata, size); > +} > + > +static MemTxResult iopmp_handle_block(void *opaque, hwaddr addr, > + uint64_t *data, unsigned size, > + MemTxAttrs attrs, > + iopmp_access_type access_type) { > + IopmpState *s = IOPMP(opaque); > + int md_idx, entry_idx; > + uint32_t error_info = 0; > + uint32_t error_id = 0; > + int rrid = attrs.requester_id; > + int result; > + hwaddr start_addr, end_addr; > + start_addr = addr; > + end_addr = addr; > + result = match_entry(s, rrid, start_addr, end_addr, &md_idx, &entry_idx, > + NULL); > + > + if (result == ENTRY_HIT) { > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > + access_type); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, access_type); > + iopmp_error_reaction(s, error_id, start_addr, error_info); > + } else if (result == ENTRY_PAR_HIT) { > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > + ERR_REQINFO_ETYPE_PARHIT); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, > + access_type); > + iopmp_error_reaction(s, error_id, start_addr, error_info); > + } else { > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > + ERR_REQINFO_ETYPE_NOHIT); > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, access_type); > + iopmp_error_reaction(s, error_id, start_addr, error_info); > + } > + > + if (access_type == IOPMP_ACCESS_READ) { > + > + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RRE)) { > + case RRE_ERROR: > + return MEMTX_ERROR; > + break; > + case RRE_SUCCESS_VALUE: > + *data = s->fabricated_v; > + return MEMTX_OK; > + break; > + default: > + break; > + } > + return MEMTX_OK; > + } else if (access_type == IOPMP_ACCESS_WRITE) { > + > + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RWE)) { > + case RWE_ERROR: > + return MEMTX_ERROR; > + break; > + case RWE_SUCCESS: > + return MEMTX_OK; > + break; > + default: > + break; > + } > + return MEMTX_OK; > + } else { > + > + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RXE)) { > + case RXE_ERROR: > + return MEMTX_ERROR; > + break; > + case RXE_SUCCESS_VALUE: > + *data = s->fabricated_v; > + return MEMTX_OK; > + break; > + default: > + break; > + } > + return MEMTX_OK; > + } > + return MEMTX_OK; > +} > + > +static MemTxResult iopmp_block_write(void *opaque, hwaddr addr, uint64_t value, > + unsigned size, MemTxAttrs attrs) > +{ > + return iopmp_handle_block(opaque, addr, &value, size, attrs, > + IOPMP_ACCESS_WRITE); > +} > + > +static MemTxResult iopmp_block_read(void *opaque, hwaddr addr, uint64_t *pdata, > + unsigned size, MemTxAttrs attrs) > +{ > + return iopmp_handle_block(opaque, addr, pdata, size, attrs, > + IOPMP_ACCESS_READ); > +} > + > +static MemTxResult iopmp_block_fetch(void *opaque, hwaddr addr, uint64_t *pdata, > + unsigned size, MemTxAttrs attrs) > +{ > + return iopmp_handle_block(opaque, addr, pdata, size, attrs, > + IOPMP_ACCESS_FETCH); > +} > + > +static const MemoryRegionOps iopmp_block_rw_ops = { > + .fetch_with_attrs = iopmp_permssion_read, > + .read_with_attrs = iopmp_block_read, > + .write_with_attrs = iopmp_block_write, > + .endianness = DEVICE_NATIVE_ENDIAN, > + .valid = {.min_access_size = 1, .max_access_size = 8}, > +}; > + > +static const MemoryRegionOps iopmp_block_w_ops = { > + .fetch_with_attrs = iopmp_permssion_read, > + .read_with_attrs = iopmp_permssion_read, > + .write_with_attrs = iopmp_block_write, > + .endianness = DEVICE_NATIVE_ENDIAN, > + .valid = {.min_access_size = 1, .max_access_size = 8}, > +}; > + > +static const MemoryRegionOps iopmp_block_r_ops = { > + .fetch_with_attrs = iopmp_permssion_read, > + .read_with_attrs = iopmp_block_read, > + .write_with_attrs = iopmp_permssion_write, > + .endianness = DEVICE_NATIVE_ENDIAN, > + .valid = {.min_access_size = 1, .max_access_size = 8}, > +}; > + > +static const MemoryRegionOps iopmp_block_rwx_ops = { > + .fetch_with_attrs = iopmp_block_fetch, > + .read_with_attrs = iopmp_block_read, > + .write_with_attrs = iopmp_block_write, > + .endianness = DEVICE_NATIVE_ENDIAN, > + .valid = {.min_access_size = 1, .max_access_size = 8}, > +}; > + > +static const MemoryRegionOps iopmp_block_wx_ops = { > + .fetch_with_attrs = iopmp_block_fetch, > + .read_with_attrs = iopmp_permssion_read, > + .write_with_attrs = iopmp_block_write, > + .endianness = DEVICE_NATIVE_ENDIAN, > + .valid = {.min_access_size = 1, .max_access_size = 8}, > +}; > + > +static const MemoryRegionOps iopmp_block_rx_ops = { > + .fetch_with_attrs = iopmp_block_fetch, > + .read_with_attrs = iopmp_block_read, > + .write_with_attrs = iopmp_permssion_write, > + .endianness = DEVICE_NATIVE_ENDIAN, > + .valid = {.min_access_size = 1, .max_access_size = 8}, > +}; > + > +static const MemoryRegionOps iopmp_block_x_ops = { > + .fetch_with_attrs = iopmp_block_fetch, > + .read_with_attrs = iopmp_permssion_read, > + .write_with_attrs = iopmp_permssion_write, > + .endianness = DEVICE_NATIVE_ENDIAN, > + .valid = {.min_access_size = 1, .max_access_size = 8}, > +}; > + > +static void iopmp_realize(DeviceState *dev, Error **errp) > +{ > + Object *obj = OBJECT(dev); > + SysBusDevice *sbd = SYS_BUS_DEVICE(dev); > + IopmpState *s = IOPMP(dev); > + uint64_t size; > + > + size = -1ULL; > + s->model = IOPMP_MODEL_RAPIDK; Should this be a property to allow other models in the future? > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, s->md_num); > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, 1); > + > + s->prient_prog = s->default_prient_prog; > + s->rrid_num = MIN(s->rrid_num, IOPMP_MAX_RRID_NUM); > + s->md_num = MIN(s->md_num, IOPMP_MAX_MD_NUM); > + s->entry_num = s->md_num * s->k; > + s->prio_entry = MIN(s->prio_entry, s->entry_num); > + > + s->regs.mdcfg = g_malloc0(s->md_num * sizeof(uint32_t)); > + s->regs.mdcfg[0] = s->k; > + > + s->regs.srcmd_en = g_malloc0(s->rrid_num * sizeof(uint32_t)); > + s->regs.srcmd_enh = g_malloc0(s->rrid_num * sizeof(uint32_t)); > + s->regs.entry = g_malloc0(s->entry_num * sizeof(iopmp_entry_t)); > + s->entry_addr = g_malloc0(s->entry_num * sizeof(iopmp_addr_t)); > + s->transaction_state = g_malloc0(s->rrid_num * > + sizeof(iopmp_transaction_state)); > + qemu_mutex_init(&s->iopmp_transaction_mutex); > + > + memory_region_init_iommu(&s->iommu, sizeof(s->iommu), > + TYPE_IOPMP_IOMMU_MEMORY_REGION, > + obj, "riscv-iopmp-sysbus-iommu", UINT64_MAX); > + memory_region_init_io(&s->mmio, obj, &iopmp_ops, > + s, "iopmp-regs", 0x100000); > + sysbus_init_mmio(sbd, &s->mmio); > + > + memory_region_init_io(&s->blocked_rw, NULL, &iopmp_block_rw_ops, > + s, "iopmp-blocked-rw", size); > + memory_region_init_io(&s->blocked_w, NULL, &iopmp_block_w_ops, > + s, "iopmp-blocked-w", size); > + memory_region_init_io(&s->blocked_r, NULL, &iopmp_block_r_ops, > + s, "iopmp-blocked-r", size); > + > + memory_region_init_io(&s->blocked_rwx, NULL, &iopmp_block_rwx_ops, > + s, "iopmp-blocked-rwx", size); > + memory_region_init_io(&s->blocked_wx, NULL, &iopmp_block_wx_ops, > + s, "iopmp-blocked-wx", size); > + memory_region_init_io(&s->blocked_rx, NULL, &iopmp_block_rx_ops, > + s, "iopmp-blocked-rx", size); > + memory_region_init_io(&s->blocked_x, NULL, &iopmp_block_x_ops, > + s, "iopmp-blocked-x", size); > + address_space_init(&s->blocked_rw_as, &s->blocked_rw, > + "iopmp-blocked-rw-as"); > + address_space_init(&s->blocked_w_as, &s->blocked_w, > + "iopmp-blocked-w-as"); > + address_space_init(&s->blocked_r_as, &s->blocked_r, > + "iopmp-blocked-r-as"); > + > + address_space_init(&s->blocked_rwx_as, &s->blocked_rwx, > + "iopmp-blocked-rwx-as"); > + address_space_init(&s->blocked_wx_as, &s->blocked_wx, > + "iopmp-blocked-wx-as"); > + address_space_init(&s->blocked_rx_as, &s->blocked_rx, > + "iopmp-blocked-rx-as"); > + address_space_init(&s->blocked_x_as, &s->blocked_x, > + "iopmp-blocked-x-as"); > +} > + > +static void iopmp_reset(DeviceState *dev) > +{ > + IopmpState *s = IOPMP(dev); > + > + qemu_set_irq(s->irq, 0); > + memset(s->regs.srcmd_en, 0, s->rrid_num * sizeof(uint32_t)); > + memset(s->regs.srcmd_enh, 0, s->rrid_num * sizeof(uint32_t)); > + memset(s->entry_addr, 0, s->entry_num * sizeof(iopmp_addr_t)); > + > + s->regs.mdlck = 0; > + s->regs.mdlckh = 0; > + s->regs.entrylck = 0; > + s->regs.mdstall = 0; > + s->regs.mdstallh = 0; > + s->regs.rridscp = 0; > + s->regs.err_cfg = 0; > + s->regs.err_reqaddr = 0; > + s->regs.err_reqid = 0; > + s->regs.err_reqinfo = 0; > + > + s->prient_prog = s->default_prient_prog; > + s->enable = 0; > + > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, s->md_num); > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, 1); > + s->regs.mdcfg[0] = s->k; > +} > + > +static int iopmp_attrs_to_index(IOMMUMemoryRegion *iommu, MemTxAttrs attrs) > +{ > + return attrs.requester_id; > +} > + > +static void iopmp_iommu_memory_region_class_init(ObjectClass *klass, void *data) > +{ > + IOMMUMemoryRegionClass *imrc = IOMMU_MEMORY_REGION_CLASS(klass); > + > + imrc->translate = iopmp_translate; > + imrc->attrs_to_index = iopmp_attrs_to_index; > +} > + > +static Property iopmp_property[] = { > + DEFINE_PROP_BOOL("prient_prog", IopmpState, default_prient_prog, true), > + DEFINE_PROP_UINT32("k", IopmpState, k, 6), > + DEFINE_PROP_UINT32("prio_entry", IopmpState, prio_entry, 48), > + DEFINE_PROP_UINT32("rrid_num", IopmpState, rrid_num, 16), > + DEFINE_PROP_UINT32("md_num", IopmpState, md_num, 8), > + DEFINE_PROP_UINT32("entry_offset", IopmpState, entry_offset, 0x4000), > + DEFINE_PROP_UINT32("fabricated_v", IopmpState, fabricated_v, 0x0), > + DEFINE_PROP_END_OF_LIST(), > +}; > + > +static void iopmp_class_init(ObjectClass *klass, void *data) > +{ > + DeviceClass *dc = DEVICE_CLASS(klass); > + device_class_set_props(dc, iopmp_property); > + dc->realize = iopmp_realize; > + dc->reset = iopmp_reset; > +} > + > +static void iopmp_init(Object *obj) > +{ > + IopmpState *s = IOPMP(obj); > + SysBusDevice *sbd = SYS_BUS_DEVICE(obj); > + > + sysbus_init_irq(sbd, &s->irq); > +} > + > +static const TypeInfo iopmp_info = { > + .name = TYPE_IOPMP, > + .parent = TYPE_SYS_BUS_DEVICE, > + .instance_size = sizeof(IopmpState), > + .instance_init = iopmp_init, > + .class_init = iopmp_class_init, > +}; > + > +static const TypeInfo > +iopmp_iommu_memory_region_info = { > + .name = TYPE_IOPMP_IOMMU_MEMORY_REGION, > + .parent = TYPE_IOMMU_MEMORY_REGION, > + .class_init = iopmp_iommu_memory_region_class_init, > +}; > + > +static void > +iopmp_register_types(void) > +{ > + type_register_static(&iopmp_info); > + type_register_static(&iopmp_iommu_memory_region_info); > +} > + > +type_init(iopmp_register_types); > diff --git a/hw/misc/trace-events b/hw/misc/trace-events > index 1be0717c0c..c148166d2d 100644 > --- a/hw/misc/trace-events > +++ b/hw/misc/trace-events > @@ -362,3 +362,6 @@ aspeed_sli_read(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx > aspeed_sliio_write(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx64 " of size %u: 0x%" PRIx32 > aspeed_sliio_read(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx64 " of size %u: 0x%" PRIx32 > > +# riscv_iopmp.c > +iopmp_read(uint64_t addr, uint32_t val) "addr 0x%"PRIx64" val 0x%x" > +iopmp_write(uint64_t addr, uint32_t val) "addr 0x%"PRIx64" val 0x%x" > diff --git a/include/hw/misc/riscv_iopmp.h b/include/hw/misc/riscv_iopmp.h > new file mode 100644 > index 0000000000..b8fe479108 > --- /dev/null > +++ b/include/hw/misc/riscv_iopmp.h > @@ -0,0 +1,168 @@ > +/* > + * QEMU RISC-V IOPMP (Input Output Physical Memory Protection) > + * > + * Copyright (c) 2023-2024 Andes Tech. Corp. > + * > + * SPDX-License-Identifier: GPL-2.0-or-later > + * > + * This program is free software; you can redistribute it and/or modify it > + * under the terms and conditions of the GNU General Public License, > + * version 2 or later, as published by the Free Software Foundation. > + * > + * This program is distributed in the hope it will be useful, but WITHOUT > + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or > + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for > + * more details. > + * > + * You should have received a copy of the GNU General Public License along with > + * this program. If not, see <http://www.gnu.org/licenses/>. > + */ > + > +#ifndef RISCV_IOPMP_H > +#define RISCV_IOPMP_H > + > +#include "hw/sysbus.h" > +#include "qemu/typedefs.h" > +#include "memory.h" > +#include "exec/hwaddr.h" > + > +#define TYPE_IOPMP "iopmp" > +#define IOPMP(obj) OBJECT_CHECK(IopmpState, (obj), TYPE_IOPMP) > + > +#define IOPMP_MAX_MD_NUM 63 > +#define IOPMP_MAX_RRID_NUM 65535 > +#define IOPMP_MAX_ENTRY_NUM 65535 > + > +#define VENDER_VIRT 0 > +#define SPECVER_0_9_1 91 > +#define IMPID_0_9_1 91 > + > +#define RRE_ERROR 0 > +#define RRE_SUCCESS_VALUE 1 > + > +#define RWE_ERROR 0 > +#define RWE_SUCCESS 1 > + > +#define RXE_ERROR 0 > +#define RXE_SUCCESS_VALUE 1 > + > +#define ERR_REQINFO_TTYPE_READ 1 > +#define ERR_REQINFO_TTYPE_WRITE 2 > +#define ERR_REQINFO_TTYPE_FETCH 3 > +#define ERR_REQINFO_ETYPE_NOERROR 0 > +#define ERR_REQINFO_ETYPE_READ 1 > +#define ERR_REQINFO_ETYPE_WRITE 2 > +#define ERR_REQINFO_ETYPE_FETCH 3 > +#define ERR_REQINFO_ETYPE_PARHIT 4 > +#define ERR_REQINFO_ETYPE_NOHIT 5 > +#define ERR_REQINFO_ETYPE_RRID 6 > +#define ERR_REQINFO_ETYPE_USER 7 > + > +#define IOPMP_MODEL_FULL 0 > +#define IOPMP_MODEL_RAPIDK 0x1 > +#define IOPMP_MODEL_DYNAMICK 0x2 > +#define IOPMP_MODEL_ISOLATION 0x3 > +#define IOPMP_MODEL_COMPACTK 0x4 > + > +#define ENTRY_NO_HIT 0 > +#define ENTRY_PAR_HIT 1 > +#define ENTRY_HIT 2 Why not an enum? Alistair > + > +/* The generic iopmp address space which downstream is system memory */ > +extern AddressSpace iopmp_container_as; > + > +typedef enum { > + IOPMP_AMATCH_OFF, /* Null (off) */ > + IOPMP_AMATCH_TOR, /* Top of Range */ > + IOPMP_AMATCH_NA4, /* Naturally aligned four-byte region */ > + IOPMP_AMATCH_NAPOT /* Naturally aligned power-of-two region */ > +} iopmp_am_t; > + > +typedef enum { > + IOPMP_ACCESS_READ = 1, > + IOPMP_ACCESS_WRITE = 2, > + IOPMP_ACCESS_FETCH = 3 > +} iopmp_access_type; > + > +typedef enum { > + IOPMP_NONE = 0, > + IOPMP_RO = 1, > + IOPMP_WO = 2, > + IOPMP_RW = 3, > + IOPMP_XO = 4, > + IOPMP_RX = 5, > + IOPMP_WX = 6, > + IOPMP_RWX = 7, > +} iopmp_permission; > + > +typedef struct { > + uint32_t addr_reg; > + uint32_t addrh_reg; > + uint32_t cfg_reg; > +} iopmp_entry_t; > + > +typedef struct { > + uint64_t sa; > + uint64_t ea; > +} iopmp_addr_t; > + > +typedef struct { > + uint32_t *srcmd_en; > + uint32_t *srcmd_enh; > + uint32_t *mdcfg; > + iopmp_entry_t *entry; > + uint32_t mdlck; > + uint32_t mdlckh; > + uint32_t entrylck; > + uint32_t mdcfglck; > + uint32_t mdstall; > + uint32_t mdstallh; > + uint32_t rridscp; > + uint32_t err_cfg; > + uint64_t err_reqaddr; > + uint32_t err_reqid; > + uint32_t err_reqinfo; > +} iopmp_regs; > + > + > +/* To detect partially hit */ > +typedef struct iopmp_transaction_state { > + bool running; > + bool supported; > + hwaddr start_addr; > + hwaddr end_addr; > +} iopmp_transaction_state; > + > +typedef struct IopmpState { > + SysBusDevice parent_obj; > + iopmp_addr_t *entry_addr; > + MemoryRegion mmio; > + IOMMUMemoryRegion iommu; > + IOMMUMemoryRegion *next_iommu; > + iopmp_regs regs; > + MemoryRegion *downstream; > + MemoryRegion blocked_r, blocked_w, blocked_x, blocked_rw, blocked_rx, > + blocked_wx, blocked_rwx; > + MemoryRegion stall_io; > + uint32_t model; > + uint32_t k; > + bool prient_prog; > + bool default_prient_prog; > + iopmp_transaction_state *transaction_state; > + QemuMutex iopmp_transaction_mutex; > + > + AddressSpace downstream_as; > + AddressSpace blocked_r_as, blocked_w_as, blocked_x_as, blocked_rw_as, > + blocked_rx_as, blocked_wx_as, blocked_rwx_as; > + qemu_irq irq; > + bool enable; > + > + uint32_t prio_entry; > + uint32_t rrid_num; > + uint32_t md_num; > + uint32_t entry_num; > + uint32_t entry_offset; > + uint32_t fabricated_v; > +} IopmpState; > + > +#endif > -- > 2.34.1 > >
On Thu, Aug 08, 2024 at 01:56:35PM +1000, Alistair Francis wrote: > [EXTERNAL MAIL] > > On Mon, Jul 15, 2024 at 7:58 PM Ethan Chen via <qemu-devel@nongnu.org> wrote: > > > > Support basic functions of IOPMP specification v0.9.1 rapid-k model. > > The specification url: > > https://github.com/riscv-non-isa/iopmp-spec/releases/tag/v0.9.1 > > > > The IOPMP checks whether memory access from a device or CPU is valid. > > This implementation uses an IOMMU to modify the address space accessed > > by the device. > > > > For device access with IOMMUAccessFlags specifying read or write > > (IOMMU_RO or IOMMU_WO), the IOPMP checks the permission in > > iopmp_translate. If the access is valid, the target address space is > > downstream_as. If the access is blocked, it will be redirected to > > blocked_rwx_as. > > > > For CPU access with IOMMUAccessFlags not specifying read or write > > (IOMMU_NONE), the IOPMP translates the access to the corresponding > > address space based on the permission. If the access has full permission > > (rwx), the target address space is downstream_as. If the access has > > limited permission, the target address space is blocked_ followed by > > the lacked permissions. > > > > The operation of a blocked region can trigger an IOPMP interrupt, a bus > > error, or it can respond with success and fabricated data, depending on > > the value of the IOPMP ERR_CFG register. > > > > Signed-off-by: Ethan Chen <ethan84@andestech.com> > > --- > > hw/misc/Kconfig | 3 + > > hw/misc/meson.build | 1 + > > hw/misc/riscv_iopmp.c | 1154 +++++++++++++++++++++++++++++++++ > > hw/misc/trace-events | 3 + > > include/hw/misc/riscv_iopmp.h | 168 +++++ > > 5 files changed, 1329 insertions(+) > > create mode 100644 hw/misc/riscv_iopmp.c > > create mode 100644 include/hw/misc/riscv_iopmp.h > > > > diff --git a/hw/misc/Kconfig b/hw/misc/Kconfig > > index 1e08785b83..427f0c702e 100644 > > --- a/hw/misc/Kconfig > > +++ b/hw/misc/Kconfig > > @@ -213,4 +213,7 @@ config IOSB > > config XLNX_VERSAL_TRNG > > bool > > > > +config RISCV_IOPMP > > + bool > > + > > source macio/Kconfig > > diff --git a/hw/misc/meson.build b/hw/misc/meson.build > > index 2ca8717be2..d9006e1d81 100644 > > --- a/hw/misc/meson.build > > +++ b/hw/misc/meson.build > > @@ -34,6 +34,7 @@ system_ss.add(when: 'CONFIG_SIFIVE_E_PRCI', if_true: files('sifive_e_prci.c')) > > system_ss.add(when: 'CONFIG_SIFIVE_E_AON', if_true: files('sifive_e_aon.c')) > > system_ss.add(when: 'CONFIG_SIFIVE_U_OTP', if_true: files('sifive_u_otp.c')) > > system_ss.add(when: 'CONFIG_SIFIVE_U_PRCI', if_true: files('sifive_u_prci.c')) > > +specific_ss.add(when: 'CONFIG_RISCV_IOPMP', if_true: files('riscv_iopmp.c')) > > > > subdir('macio') > > > > diff --git a/hw/misc/riscv_iopmp.c b/hw/misc/riscv_iopmp.c > > new file mode 100644 > > index 0000000000..db43e3c73f > > --- /dev/null > > +++ b/hw/misc/riscv_iopmp.c > > @@ -0,0 +1,1154 @@ > > +/* > > + * QEMU RISC-V IOPMP (Input Output Physical Memory Protection) > > + * > > + * Copyright (c) 2023-2024 Andes Tech. Corp. > > + * > > + * SPDX-License-Identifier: GPL-2.0-or-later > > + * > > + * This program is free software; you can redistribute it and/or modify it > > + * under the terms and conditions of the GNU General Public License, > > + * version 2 or later, as published by the Free Software Foundation. > > + * > > + * This program is distributed in the hope it will be useful, but WITHOUT > > + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or > > + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for > > + * more details. > > + * > > + * this program. If not, see <http://www.gnu.org/licenses/>. > > + */ > > + > > +#include "qemu/osdep.h" > > +#include "qemu/log.h" > > +#include "qapi/error.h" > > +#include "trace.h" > > +#include "exec/exec-all.h" > > +#include "exec/address-spaces.h" > > +#include "hw/qdev-properties.h" > > +#include "hw/sysbus.h" > > +#include "hw/misc/riscv_iopmp.h" > > +#include "memory.h" > > +#include "hw/irq.h" > > +#include "hw/registerfields.h" > > +#include "trace.h" > > + > > +#define TYPE_IOPMP_IOMMU_MEMORY_REGION "iopmp-iommu-memory-region" > > + > > +REG32(VERSION, 0x00) > > + FIELD(VERSION, VENDOR, 0, 24) > > + FIELD(VERSION, SPECVER , 24, 8) > > +REG32(IMP, 0x04) > > + FIELD(IMP, IMPID, 0, 32) > > +REG32(HWCFG0, 0x08) > > + FIELD(HWCFG0, MODEL, 0, 4) > > + FIELD(HWCFG0, TOR_EN, 4, 1) > > + FIELD(HWCFG0, SPS_EN, 5, 1) > > + FIELD(HWCFG0, USER_CFG_EN, 6, 1) > > + FIELD(HWCFG0, PRIENT_PROG, 7, 1) > > + FIELD(HWCFG0, RRID_TRANSL_EN, 8, 1) > > + FIELD(HWCFG0, RRID_TRANSL_PROG, 9, 1) > > + FIELD(HWCFG0, CHK_X, 10, 1) > > + FIELD(HWCFG0, NO_X, 11, 1) > > + FIELD(HWCFG0, NO_W, 12, 1) > > + FIELD(HWCFG0, STALL_EN, 13, 1) > > + FIELD(HWCFG0, PEIS, 14, 1) > > + FIELD(HWCFG0, PEES, 15, 1) > > + FIELD(HWCFG0, MFR_EN, 16, 1) > > + FIELD(HWCFG0, MD_NUM, 24, 7) > > + FIELD(HWCFG0, ENABLE, 31, 1) > > +REG32(HWCFG1, 0x0C) > > + FIELD(HWCFG1, RRID_NUM, 0, 16) > > + FIELD(HWCFG1, ENTRY_NUM, 16, 16) > > +REG32(HWCFG2, 0x10) > > + FIELD(HWCFG2, PRIO_ENTRY, 0, 16) > > + FIELD(HWCFG2, RRID_TRANSL, 16, 16) > > +REG32(ENTRYOFFSET, 0x14) > > + FIELD(ENTRYOFFSET, OFFSET, 0, 32) > > +REG32(MDSTALL, 0x30) > > + FIELD(MDSTALL, EXEMPT, 0, 1) > > + FIELD(MDSTALL, MD, 1, 31) > > +REG32(MDSTALLH, 0x34) > > + FIELD(MDSTALLH, MD, 0, 32) > > +REG32(RRIDSCP, 0x38) > > + FIELD(RRIDSCP, RRID, 0, 16) > > + FIELD(RRIDSCP, OP, 30, 2) > > +REG32(MDLCK, 0x40) > > + FIELD(MDLCK, L, 0, 1) > > + FIELD(MDLCK, MD, 1, 31) > > +REG32(MDLCKH, 0x44) > > + FIELD(MDLCKH, MDH, 0, 32) > > +REG32(MDCFGLCK, 0x48) > > + FIELD(MDCFGLCK, L, 0, 1) > > + FIELD(MDCFGLCK, F, 1, 7) > > +REG32(ENTRYLCK, 0x4C) > > + FIELD(ENTRYLCK, L, 0, 1) > > + FIELD(ENTRYLCK, F, 1, 16) > > +REG32(ERR_CFG, 0x60) > > + FIELD(ERR_CFG, L, 0, 1) > > + FIELD(ERR_CFG, IE, 1, 1) > > + FIELD(ERR_CFG, IRE, 2, 1) > > + FIELD(ERR_CFG, IWE, 3, 1) > > + FIELD(ERR_CFG, IXE, 4, 1) > > + FIELD(ERR_CFG, RRE, 5, 1) > > + FIELD(ERR_CFG, RWE, 6, 1) > > + FIELD(ERR_CFG, RXE, 7, 1) > > +REG32(ERR_REQINFO, 0x64) > > + FIELD(ERR_REQINFO, V, 0, 1) > > + FIELD(ERR_REQINFO, TTYPE, 1, 2) > > + FIELD(ERR_REQINFO, ETYPE, 4, 3) > > + FIELD(ERR_REQINFO, SVC, 7, 1) > > +REG32(ERR_REQADDR, 0x68) > > + FIELD(ERR_REQADDR, ADDR, 0, 32) > > +REG32(ERR_REQADDRH, 0x6C) > > + FIELD(ERR_REQADDRH, ADDRH, 0, 32) > > +REG32(ERR_REQID, 0x70) > > + FIELD(ERR_REQID, RRID, 0, 16) > > + FIELD(ERR_REQID, EID, 16, 16) > > +REG32(ERR_MFR, 0x74) > > + FIELD(ERR_MFR, SVW, 0, 16) > > + FIELD(ERR_MFR, SVI, 16, 12) > > + FIELD(ERR_MFR, SVS, 31, 1) > > +REG32(MDCFG0, 0x800) > > + FIELD(MDCFG0, T, 0, 16) > > +REG32(SRCMD_EN0, 0x1000) > > + FIELD(SRCMD_EN0, L, 0, 1) > > + FIELD(SRCMD_EN0, MD, 1, 31) > > +REG32(SRCMD_ENH0, 0x1004) > > + FIELD(SRCMD_ENH0, MDH, 0, 32) > > +REG32(SRCMD_R0, 0x1008) > > + FIELD(SRCMD_R0, MD, 1, 31) > > +REG32(SRCMD_RH0, 0x100C) > > + FIELD(SRCMD_RH0, MDH, 0, 32) > > +REG32(SRCMD_W0, 0x1010) > > + FIELD(SRCMD_W0, MD, 1, 31) > > +REG32(SRCMD_WH0, 0x1014) > > + FIELD(SRCMD_WH0, MDH, 0, 32) > > + > > +FIELD(ENTRY_ADDR, ADDR, 0, 32) > > +FIELD(ENTRY_ADDRH, ADDRH, 0, 32) > > + > > +FIELD(ENTRY_CFG, R, 0, 1) > > +FIELD(ENTRY_CFG, W, 1, 1) > > +FIELD(ENTRY_CFG, X, 2, 1) > > +FIELD(ENTRY_CFG, A, 3, 2) > > +FIELD(ENTRY_CFG, SIRE, 5, 1) > > +FIELD(ENTRY_CFG, SIWE, 6, 1) > > +FIELD(ENTRY_CFG, SIXE, 7, 1) > > +FIELD(ENTRY_CFG, SERE, 8, 1) > > +FIELD(ENTRY_CFG, SEWE, 9, 1) > > +FIELD(ENTRY_CFG, SEXE, 10, 1) > > + > > +FIELD(ENTRY_USER_CFG, IM, 0, 32) > > + > > +/* Offsets to SRCMD_EN(i) */ > > +#define SRCMD_EN_OFFSET 0x0 > > +#define SRCMD_ENH_OFFSET 0x4 > > +#define SRCMD_R_OFFSET 0x8 > > +#define SRCMD_RH_OFFSET 0xC > > +#define SRCMD_W_OFFSET 0x10 > > +#define SRCMD_WH_OFFSET 0x14 > > + > > +/* Offsets to ENTRY_ADDR(i) */ > > +#define ENTRY_ADDR_OFFSET 0x0 > > +#define ENTRY_ADDRH_OFFSET 0x4 > > +#define ENTRY_CFG_OFFSET 0x8 > > +#define ENTRY_USER_CFG_OFFSET 0xC > > + > > +/* Memmap for parallel IOPMPs */ > > +typedef struct iopmp_protection_memmap { > > + MemMapEntry entry; > > + IopmpState *iopmp_s; > > + QLIST_ENTRY(iopmp_protection_memmap) list; > > +} iopmp_protection_memmap; > > +QLIST_HEAD(, iopmp_protection_memmap) > > + iopmp_protection_memmaps = QLIST_HEAD_INITIALIZER(iopmp_protection_memmaps); > > + > > +static void iopmp_iommu_notify(IopmpState *s) > > +{ > > + IOMMUTLBEvent event = { > > + .entry = { > > + .iova = 0, > > + .translated_addr = 0, > > + .addr_mask = -1ULL, > > + .perm = IOMMU_NONE, > > + }, > > + .type = IOMMU_NOTIFIER_UNMAP, > > + }; > > + > > + for (int i = 0; i < s->rrid_num; i++) { > > + memory_region_notify_iommu(&s->iommu, i, event); > > + } > > +} > > + > > +static void iopmp_decode_napot(uint64_t a, uint64_t *sa, > > + uint64_t *ea) > > +{ > > + /* > > + * aaaa...aaa0 8-byte NAPOT range > > + * aaaa...aa01 16-byte NAPOT range > > + * aaaa...a011 32-byte NAPOT range > > + * ... > > + * aa01...1111 2^XLEN-byte NAPOT range > > + * a011...1111 2^(XLEN+1)-byte NAPOT range > > + * 0111...1111 2^(XLEN+2)-byte NAPOT range > > + * 1111...1111 Reserved > > + */ > > + > > + a = (a << 2) | 0x3; > > + *sa = a & (a + 1); > > + *ea = a | (a + 1); > > +} > > + > > +static void iopmp_update_rule(IopmpState *s, uint32_t entry_index) > > +{ > > + uint8_t this_cfg = s->regs.entry[entry_index].cfg_reg; > > + uint64_t this_addr = s->regs.entry[entry_index].addr_reg | > > + ((uint64_t)s->regs.entry[entry_index].addrh_reg << 32); > > + uint64_t prev_addr = 0u; > > + uint64_t sa = 0u; > > + uint64_t ea = 0u; > > + > > + if (entry_index >= 1u) { > > + prev_addr = s->regs.entry[entry_index - 1].addr_reg | > > + ((uint64_t)s->regs.entry[entry_index - 1].addrh_reg << 32); > > + } > > + > > + switch (FIELD_EX32(this_cfg, ENTRY_CFG, A)) { > > + case IOPMP_AMATCH_OFF: > > + sa = 0u; > > + ea = -1; > > + break; > > + > > + case IOPMP_AMATCH_TOR: > > + sa = (prev_addr) << 2; /* shift up from [xx:0] to [xx+2:2] */ > > + ea = ((this_addr) << 2) - 1u; > > + if (sa > ea) { > > + sa = ea = 0u; > > + } > > + break; > > + > > + case IOPMP_AMATCH_NA4: > > + sa = this_addr << 2; /* shift up from [xx:0] to [xx+2:2] */ > > + ea = (sa + 4u) - 1u; > > + break; > > + > > + case IOPMP_AMATCH_NAPOT: > > + iopmp_decode_napot(this_addr, &sa, &ea); > > + break; > > + > > + default: > > + sa = 0u; > > + ea = 0u; > > + break; > > + } > > + > > + s->entry_addr[entry_index].sa = sa; > > + s->entry_addr[entry_index].ea = ea; > > + iopmp_iommu_notify(s); > > +} > > + > > +static uint64_t iopmp_read(void *opaque, hwaddr addr, unsigned size) > > +{ > > + IopmpState *s = IOPMP(opaque); > > + uint32_t rz = 0; > > + uint32_t offset, idx; > > + > > + switch (addr) { > > + case A_VERSION: > > + rz = VENDER_VIRT << R_VERSION_VENDOR_SHIFT | > > + SPECVER_0_9_1 << R_VERSION_SPECVER_SHIFT; > > It would be better to use the FIELD_DP32() macro instead of the manual shifts It will be refined in next revision. > > > + break; > > + case A_IMP: > > + rz = IMPID_0_9_1; > > + break; > > + case A_HWCFG0: > > + rz = s->model << R_HWCFG0_MODEL_SHIFT | > > + 1 << R_HWCFG0_TOR_EN_SHIFT | > > + 0 << R_HWCFG0_SPS_EN_SHIFT | > > + 0 << R_HWCFG0_USER_CFG_EN_SHIFT | > > + s->prient_prog << R_HWCFG0_PRIENT_PROG_SHIFT | > > + 0 << R_HWCFG0_RRID_TRANSL_EN_SHIFT | > > + 0 << R_HWCFG0_RRID_TRANSL_PROG_SHIFT | > > + 1 << R_HWCFG0_CHK_X_SHIFT | > > + 0 << R_HWCFG0_NO_X_SHIFT | > > + 0 << R_HWCFG0_NO_W_SHIFT | > > + 0 << R_HWCFG0_STALL_EN_SHIFT | > > + 0 << R_HWCFG0_PEIS_SHIFT | > > + 0 << R_HWCFG0_PEES_SHIFT | > > + 0 << R_HWCFG0_MFR_EN_SHIFT | > > + s->md_num << R_HWCFG0_MD_NUM_SHIFT | > > + s->enable << R_HWCFG0_ENABLE_SHIFT ; > > + break; > > + case A_HWCFG1: > > + rz = s->rrid_num << R_HWCFG1_RRID_NUM_SHIFT | > > + s->entry_num << R_HWCFG1_ENTRY_NUM_SHIFT; > > + break; > > + case A_HWCFG2: > > + rz = s->prio_entry << R_HWCFG2_PRIO_ENTRY_SHIFT; > > + break; > > + case A_ENTRYOFFSET: > > + rz = s->entry_offset; > > + break; > > + case A_ERR_CFG: > > + rz = s->regs.err_cfg; > > + break; > > + case A_MDLCK: > > + rz = s->regs.mdlck; > > + break; > > + case A_MDLCKH: > > + rz = s->regs.mdlckh; > > + break; > > + case A_MDCFGLCK: > > + rz = s->regs.mdcfglck; > > + break; > > + case A_ENTRYLCK: > > + rz = s->regs.entrylck; > > + break; > > + case A_ERR_REQADDR: > > + rz = s->regs.err_reqaddr & UINT32_MAX; > > + break; > > + case A_ERR_REQADDRH: > > + rz = s->regs.err_reqaddr >> 32; > > + break; > > + case A_ERR_REQID: > > + rz = s->regs.err_reqid; > > + break; > > + case A_ERR_REQINFO: > > + rz = s->regs.err_reqinfo; > > + break; > > + > > + default: > > + if (addr >= A_MDCFG0 && > > + addr < A_MDCFG0 + 4 * (s->md_num - 1)) { > > + offset = addr - A_MDCFG0; > > + idx = offset >> 2; > > + if (idx == 0 && offset == 0) { > > + rz = s->regs.mdcfg[idx]; > > + } else { > > + /* Only MDCFG0 is implemented in rapid-k model */ > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + } > > + } else if (addr >= A_SRCMD_EN0 && > > + addr < A_SRCMD_WH0 + 32 * (s->rrid_num - 1)) { > > + offset = addr - A_SRCMD_EN0; > > + idx = offset >> 5; > > + offset &= 0x1f; > > + > > + switch (offset) { > > + case SRCMD_EN_OFFSET: > > + rz = s->regs.srcmd_en[idx]; > > + break; > > + case SRCMD_ENH_OFFSET: > > + rz = s->regs.srcmd_enh[idx]; > > + break; > > + default: > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + break; > > + } > > + } else if (addr >= s->entry_offset && > > + addr < s->entry_offset + ENTRY_USER_CFG_OFFSET + > > + 16 * (s->entry_num - 1)) { > > + offset = addr - s->entry_offset; > > + idx = offset >> 4; > > + offset &= 0xf; > > + > > + switch (offset) { > > + case ENTRY_ADDR_OFFSET: > > + rz = s->regs.entry[idx].addr_reg; > > + break; > > + case ENTRY_ADDRH_OFFSET: > > + rz = s->regs.entry[idx].addrh_reg; > > + break; > > + case ENTRY_CFG_OFFSET: > > + rz = s->regs.entry[idx].cfg_reg; > > + break; > > + case ENTRY_USER_CFG_OFFSET: > > + /* Does not support user customized permission */ > > + rz = 0; > > + break; > > + default: > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + break; > > + } > > + } else { > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + } > > + break; > > + } > > + trace_iopmp_read(addr, rz); > > + return rz; > > +} > > + > > +static void > > +iopmp_write(void *opaque, hwaddr addr, uint64_t value, unsigned size) > > +{ > > + IopmpState *s = IOPMP(opaque); > > + uint32_t offset, idx; > > + uint32_t value32 = value; > > + > > + trace_iopmp_write(addr, value32); > > + > > + switch (addr) { > > + case A_VERSION: /* RO */ > > + break; > > + case A_IMP: /* RO */ > > + break; > > + case A_HWCFG0: > > + if (FIELD_EX32(value32, HWCFG0, PRIENT_PROG)) { > > + /* W1C */ > > + s->prient_prog = 0; > > + } > > + if (FIELD_EX32(value32, HWCFG0, ENABLE)) { > > + /* W1S */ > > + s->enable = 1; > > + iopmp_iommu_notify(s); > > + } > > + break; > > + case A_HWCFG1: /* RO */ > > + break; > > + case A_HWCFG2: > > + if (s->prient_prog) { > > + s->prio_entry = FIELD_EX32(value32, HWCFG2, PRIO_ENTRY); > > + } > > + break; > > + case A_ERR_CFG: > > + if (!FIELD_EX32(s->regs.err_cfg, ERR_CFG, L)) { > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, L, > > + FIELD_EX32(value32, ERR_CFG, L)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IE, > > + FIELD_EX32(value32, ERR_CFG, IE)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IRE, > > + FIELD_EX32(value32, ERR_CFG, IRE)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RRE, > > + FIELD_EX32(value32, ERR_CFG, RRE)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IWE, > > + FIELD_EX32(value32, ERR_CFG, IWE)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RWE, > > + FIELD_EX32(value32, ERR_CFG, RWE)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IXE, > > + FIELD_EX32(value32, ERR_CFG, IXE)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RXE, > > + FIELD_EX32(value32, ERR_CFG, RXE)); > > + } > > + break; > > + case A_MDLCK: > > + if (!FIELD_EX32(s->regs.mdlck, MDLCK, L)) { > > + s->regs.mdlck = value32; > > + } > > + break; > > + case A_MDLCKH: > > + if (!FIELD_EX32(s->regs.mdlck, MDLCK, L)) { > > + s->regs.mdlckh = value32; > > + } > > + break; > > + case A_MDCFGLCK: > > + if (!FIELD_EX32(s->regs.mdcfglck, MDCFGLCK, L)) { > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, > > + FIELD_EX32(value32, MDCFGLCK, F)); > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, > > + FIELD_EX32(value32, MDCFGLCK, L)); > > + } > > + break; > > + case A_ENTRYLCK: > > + if (!(FIELD_EX32(s->regs.entrylck, ENTRYLCK, L))) { > > + s->regs.entrylck = FIELD_DP32(s->regs.entrylck, ENTRYLCK, F, > > + FIELD_EX32(value32, ENTRYLCK, F)); > > + s->regs.entrylck = FIELD_DP32(s->regs.entrylck, ENTRYLCK, L, > > + FIELD_EX32(value32, ENTRYLCK, L)); > > + } > > + case A_ERR_REQADDR: /* RO */ > > + break; > > + case A_ERR_REQADDRH: /* RO */ > > + break; > > + case A_ERR_REQID: /* RO */ > > + break; > > + case A_ERR_REQINFO: > > + if (FIELD_EX32(value32, ERR_REQINFO, V)) { > > + s->regs.err_reqinfo = FIELD_DP32(s->regs.err_reqinfo, > > + ERR_REQINFO, V, 0); > > + qemu_set_irq(s->irq, 0); > > + } > > + break; > > + > > + default: > > + if (addr >= A_MDCFG0 && > > + addr < A_MDCFG0 + 4 * (s->md_num - 1)) { > > + offset = addr - A_MDCFG0; > > + idx = offset >> 2; > > + /* RO in rapid-k model */ > > + if (idx > 0) { > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + } > > + } else if (addr >= A_SRCMD_EN0 && > > + addr < A_SRCMD_WH0 + 32 * (s->rrid_num - 1)) { > > + offset = addr - A_SRCMD_EN0; > > + idx = offset >> 5; > > + offset &= 0x1f; > > + > > + if (offset % 4) { > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + } else if (FIELD_EX32(s->regs.srcmd_en[idx], SRCMD_EN0, L) > > + == 0) { > > + switch (offset) { > > + case SRCMD_EN_OFFSET: > > + s->regs.srcmd_en[idx] = > > + FIELD_DP32(s->regs.srcmd_en[idx], SRCMD_EN0, L, > > + FIELD_EX32(value32, SRCMD_EN0, L)); > > + > > + /* MD field is protected by mdlck */ > > + value32 = (value32 & ~s->regs.mdlck) | > > + (s->regs.srcmd_en[idx] & s->regs.mdlck); > > + s->regs.srcmd_en[idx] = > > + FIELD_DP32(s->regs.srcmd_en[idx], SRCMD_EN0, MD, > > + FIELD_EX32(value32, SRCMD_EN0, MD)); > > + break; > > + case SRCMD_ENH_OFFSET: > > + value32 = (value32 & ~s->regs.mdlckh) | > > + (s->regs.srcmd_enh[idx] & s->regs.mdlckh); > > + s->regs.srcmd_enh[idx] = > > + FIELD_DP32(s->regs.srcmd_enh[idx], SRCMD_ENH0, MDH, > > + value32); > > + break; > > + default: > > + break; > > + } > > + } > > + } else if (addr >= s->entry_offset && > > + addr < s->entry_offset + ENTRY_USER_CFG_OFFSET > > + + 16 * (s->entry_num - 1)) { > > + offset = addr - s->entry_offset; > > + idx = offset >> 4; > > + offset &= 0xf; > > + > > + /* index < ENTRYLCK_F is protected */ > > + if (idx >= FIELD_EX32(s->regs.entrylck, ENTRYLCK, F)) { > > + switch (offset) { > > + case ENTRY_ADDR_OFFSET: > > + s->regs.entry[idx].addr_reg = value32; > > + break; > > + case ENTRY_ADDRH_OFFSET: > > + s->regs.entry[idx].addrh_reg = value32; > > + break; > > + case ENTRY_CFG_OFFSET: > > + s->regs.entry[idx].cfg_reg = value32; > > + break; > > + case ENTRY_USER_CFG_OFFSET: > > + /* Does not support user customized permission */ > > + break; > > + default: > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + break; > > + } > > + iopmp_update_rule(s, idx); > > + if (idx + 1 < s->entry_num && > > + FIELD_EX32(s->regs.entry[idx + 1].cfg_reg, ENTRY_CFG, A) == > > + IOPMP_AMATCH_TOR) { > > + iopmp_update_rule(s, idx + 1); > > + } > > + } > > + } else { > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", __func__, > > + (int)addr); > > + } > > + } > > +} > > + > > +/* Match entry in memory domain */ > > +static int match_entry_md(IopmpState *s, int md_idx, hwaddr start_addr, > > + hwaddr end_addr, int *entry_idx, > > + int *prior_entry_in_tlb) > > +{ > > + int entry_idx_s, entry_idx_e; > > + int result = ENTRY_NO_HIT; > > + int i = 0; > > + hwaddr tlb_sa = start_addr & ~(TARGET_PAGE_SIZE - 1); > > + hwaddr tlb_ea = tlb_sa + TARGET_PAGE_SIZE - 1; > > + > > + entry_idx_s = md_idx * s->regs.mdcfg[0]; > > + entry_idx_e = (md_idx + 1) * s->regs.mdcfg[0]; > > + > > + if (entry_idx_s >= s->entry_num) { > > + return result; > > + } > > + if (entry_idx_e > s->entry_num) { > > + entry_idx_e = s->entry_num; > > + } > > + i = entry_idx_s; > > + for (i = entry_idx_s; i < entry_idx_e; i++) { > > + if (FIELD_EX32(s->regs.entry[i].cfg_reg, ENTRY_CFG, A) == > > + IOPMP_AMATCH_OFF) { > > + continue; > > + } > > + if (start_addr >= s->entry_addr[i].sa && > > + start_addr <= s->entry_addr[i].ea) { > > + /* Check end address */ > > + if (end_addr >= s->entry_addr[i].sa && > > + end_addr <= s->entry_addr[i].ea) { > > + *entry_idx = i; > > + return ENTRY_HIT; > > + } else if (i >= s->prio_entry) { > > + /* Continue for non-prio_entry */ > > + continue; > > + } else { > > + *entry_idx = i; > > + return ENTRY_PAR_HIT; > > + } > > + } else if (end_addr >= s->entry_addr[i].sa && > > + end_addr <= s->entry_addr[i].ea) { > > + /* Only end address matches the entry */ > > + if (i >= s->prio_entry) { > > + continue; > > + } else { > > + *entry_idx = i; > > + return ENTRY_PAR_HIT; > > + } > > + } else if (start_addr < s->entry_addr[i].sa && > > + end_addr > s->entry_addr[i].ea) { > > + if (i >= s->prio_entry) { > > + continue; > > + } else { > > + *entry_idx = i; > > + return ENTRY_PAR_HIT; > > + } > > + } > > + if (prior_entry_in_tlb != NULL) { > > + if ((s->entry_addr[i].sa >= tlb_sa && > > + s->entry_addr[i].sa <= tlb_ea) || > > + (s->entry_addr[i].ea >= tlb_sa && > > + s->entry_addr[i].ea <= tlb_ea)) { > > + /* > > + * TLB should not use the cached result when the tlb contains > > + * higher priority entry > > + */ > > + *prior_entry_in_tlb = 1; > > + } > > + } > > + } > > + return result; > > +} > > + > > +static int match_entry(IopmpState *s, int rrid, hwaddr start_addr, > > + hwaddr end_addr, int *match_md_idx, > > + int *match_entry_idx, int *prior_entry_in_tlb) > > +{ > > + int cur_result = ENTRY_NO_HIT; > > + int result = ENTRY_NO_HIT; > > + /* Remove lock bit */ > > + uint64_t srcmd_en = ((uint64_t)s->regs.srcmd_en[rrid] | > > + ((uint64_t)s->regs.srcmd_enh[rrid] << 32)) >> 1; > > + > > + for (int md_idx = 0; md_idx < s->md_num; md_idx++) { > > + if (srcmd_en & (1ULL << md_idx)) { > > + cur_result = match_entry_md(s, md_idx, start_addr, end_addr, > > + match_entry_idx, prior_entry_in_tlb); > > + if (cur_result == ENTRY_HIT || cur_result == ENTRY_PAR_HIT) { > > + *match_md_idx = md_idx; > > + return cur_result; > > + } > > + } > > + } > > + return result; > > +} > > + > > +static void iopmp_error_reaction(IopmpState *s, uint32_t id, hwaddr start, > > + uint32_t info) > > +{ > > + if (!FIELD_EX32(s->regs.err_reqinfo, ERR_REQINFO, V)) { > > + s->regs.err_reqinfo = info; > > + s->regs.err_reqinfo = FIELD_DP32(s->regs.err_reqinfo, ERR_REQINFO, V, > > + 1); > > + s->regs.err_reqid = id; > > + /* addr[LEN+2:2] */ > > + s->regs.err_reqaddr = start >> 2; > > + > > + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_READ && > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IRE)) { > > + qemu_set_irq(s->irq, 1); > > + } > > + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_WRITE && > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IWE)) { > > + qemu_set_irq(s->irq, 1); > > + } > > + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_FETCH && > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IXE)) { > > + qemu_set_irq(s->irq, 1); > > + } > > + } > > +} > > + > > +static IOMMUTLBEntry iopmp_translate(IOMMUMemoryRegion *iommu, hwaddr addr, > > + IOMMUAccessFlags flags, int iommu_idx) > > +{ > > + int rrid = iommu_idx; > > + IopmpState *s = IOPMP(container_of(iommu, IopmpState, iommu)); > > + hwaddr start_addr, end_addr; > > + int entry_idx = -1; > > + int md_idx = -1; > > + int result; > > + uint32_t error_info = 0; > > + uint32_t error_id = 0; > > + int prior_entry_in_tlb = 0; > > + iopmp_permission iopmp_perm; > > + IOMMUTLBEntry entry = { > > + .target_as = &s->downstream_as, > > + .iova = addr, > > + .translated_addr = addr, > > + .addr_mask = 0, > > + .perm = IOMMU_NONE, > > + }; > > + > > + if (!s->enable) { > > + /* Bypass IOPMP */ > > + entry.addr_mask = -1ULL, > > + entry.perm = IOMMU_RW; > > + return entry; > > + } > > + > > + /* unknown RRID */ > > + if (rrid >= s->rrid_num) { > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + ERR_REQINFO_ETYPE_RRID); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > > + iopmp_error_reaction(s, error_id, addr, error_info); > > + entry.target_as = &s->blocked_rwx_as; > > + entry.perm = IOMMU_RW; > > + return entry; > > + } > > + > > + if (s->transaction_state[rrid].supported == true) { > > + start_addr = s->transaction_state[rrid].start_addr; > > + end_addr = s->transaction_state[rrid].end_addr; > > + } else { > > + /* No transaction information, use the same address */ > > + start_addr = addr; > > + end_addr = addr; > > + } > > + > > + result = match_entry(s, rrid, start_addr, end_addr, &md_idx, &entry_idx, > > + &prior_entry_in_tlb); > > + if (result == ENTRY_HIT) { > > + entry.addr_mask = s->entry_addr[entry_idx].ea - > > + s->entry_addr[entry_idx].sa; > > + if (prior_entry_in_tlb) { > > + /* Make TLB repeat iommu translation on every access */ > > I don't follow this, if we have a prior entry in the TLB cache we > don't cache the accesses? For the cached TLB result to be used, the highest-priority entry in the TLB must occupy the entire TLB page. If a lower-priority entry fills the entire TLB page, it is still necessary to check which entry the transaction hits on each access to the TLB page. > > > + entry.addr_mask = 0; > > + } > > + iopmp_perm = s->regs.entry[entry_idx].cfg_reg & IOPMP_RWX; > > + if (flags) { > > + if ((iopmp_perm & flags) == 0) { > > + /* Permission denied */ > > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + ERR_REQINFO_ETYPE_READ + flags - 1); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > + entry.target_as = &s->blocked_rwx_as; > > + entry.perm = IOMMU_RW; > > + } else { > > + entry.target_as = &s->downstream_as; > > + entry.perm = iopmp_perm; > > + } > > + } else { > > + /* CPU access with IOMMU_NONE flag */ > > + if (iopmp_perm & IOPMP_XO) { > > + if ((iopmp_perm & IOPMP_RW) == IOPMP_RW) { > > + entry.target_as = &s->downstream_as; > > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_RO) { > > + entry.target_as = &s->blocked_w_as; > > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_WO) { > > + entry.target_as = &s->blocked_r_as; > > + } else { > > + entry.target_as = &s->blocked_rw_as; > > + } > > + } else { > > + if ((iopmp_perm & IOPMP_RW) == IOMMU_RW) { > > + entry.target_as = &s->blocked_x_as; > > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_RO) { > > + entry.target_as = &s->blocked_wx_as; > > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_WO) { > > + entry.target_as = &s->blocked_rx_as; > > + } else { > > + entry.target_as = &s->blocked_rwx_as; > > + } > > + } > > + entry.perm = IOMMU_RW; > > + } > > + } else { > > + if (flags) { > > + if (result == ENTRY_PAR_HIT) { > > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + ERR_REQINFO_ETYPE_PARHIT); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > + } else { > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + ERR_REQINFO_ETYPE_NOHIT); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > + } > > + } > > + /* CPU access with IOMMU_NONE flag no_hit or par_hit */ > > + entry.target_as = &s->blocked_rwx_as; > > + entry.perm = IOMMU_RW; > > + } > > + return entry; > > +} > > + > > +static const MemoryRegionOps iopmp_ops = { > > + .read = iopmp_read, > > + .write = iopmp_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 4, .max_access_size = 4} > > +}; > > + > > +static MemTxResult iopmp_permssion_write(void *opaque, hwaddr addr, > > + uint64_t value, unsigned size, > > + MemTxAttrs attrs) > > +{ > > + IopmpState *s = IOPMP(opaque); > > + return address_space_write(&s->downstream_as, addr, attrs, &value, size); > > +} > > + > > +static MemTxResult iopmp_permssion_read(void *opaque, hwaddr addr, > > + uint64_t *pdata, unsigned size, > > + MemTxAttrs attrs) > > +{ > > + IopmpState *s = IOPMP(opaque); > > + return address_space_read(&s->downstream_as, addr, attrs, pdata, size); > > +} > > + > > +static MemTxResult iopmp_handle_block(void *opaque, hwaddr addr, > > + uint64_t *data, unsigned size, > > + MemTxAttrs attrs, > > + iopmp_access_type access_type) { > > + IopmpState *s = IOPMP(opaque); > > + int md_idx, entry_idx; > > + uint32_t error_info = 0; > > + uint32_t error_id = 0; > > + int rrid = attrs.requester_id; > > + int result; > > + hwaddr start_addr, end_addr; > > + start_addr = addr; > > + end_addr = addr; > > + result = match_entry(s, rrid, start_addr, end_addr, &md_idx, &entry_idx, > > + NULL); > > + > > + if (result == ENTRY_HIT) { > > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + access_type); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, access_type); > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > + } else if (result == ENTRY_PAR_HIT) { > > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + ERR_REQINFO_ETYPE_PARHIT); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, > > + access_type); > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > + } else { > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + ERR_REQINFO_ETYPE_NOHIT); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, access_type); > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > + } > > + > > + if (access_type == IOPMP_ACCESS_READ) { > > + > > + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RRE)) { > > + case RRE_ERROR: > > + return MEMTX_ERROR; > > + break; > > + case RRE_SUCCESS_VALUE: > > + *data = s->fabricated_v; > > + return MEMTX_OK; > > + break; > > + default: > > + break; > > + } > > + return MEMTX_OK; > > + } else if (access_type == IOPMP_ACCESS_WRITE) { > > + > > + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RWE)) { > > + case RWE_ERROR: > > + return MEMTX_ERROR; > > + break; > > + case RWE_SUCCESS: > > + return MEMTX_OK; > > + break; > > + default: > > + break; > > + } > > + return MEMTX_OK; > > + } else { > > + > > + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RXE)) { > > + case RXE_ERROR: > > + return MEMTX_ERROR; > > + break; > > + case RXE_SUCCESS_VALUE: > > + *data = s->fabricated_v; > > + return MEMTX_OK; > > + break; > > + default: > > + break; > > + } > > + return MEMTX_OK; > > + } > > + return MEMTX_OK; > > +} > > + > > +static MemTxResult iopmp_block_write(void *opaque, hwaddr addr, uint64_t value, > > + unsigned size, MemTxAttrs attrs) > > +{ > > + return iopmp_handle_block(opaque, addr, &value, size, attrs, > > + IOPMP_ACCESS_WRITE); > > +} > > + > > +static MemTxResult iopmp_block_read(void *opaque, hwaddr addr, uint64_t *pdata, > > + unsigned size, MemTxAttrs attrs) > > +{ > > + return iopmp_handle_block(opaque, addr, pdata, size, attrs, > > + IOPMP_ACCESS_READ); > > +} > > + > > +static MemTxResult iopmp_block_fetch(void *opaque, hwaddr addr, uint64_t *pdata, > > + unsigned size, MemTxAttrs attrs) > > +{ > > + return iopmp_handle_block(opaque, addr, pdata, size, attrs, > > + IOPMP_ACCESS_FETCH); > > +} > > + > > +static const MemoryRegionOps iopmp_block_rw_ops = { > > + .fetch_with_attrs = iopmp_permssion_read, > > + .read_with_attrs = iopmp_block_read, > > + .write_with_attrs = iopmp_block_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static const MemoryRegionOps iopmp_block_w_ops = { > > + .fetch_with_attrs = iopmp_permssion_read, > > + .read_with_attrs = iopmp_permssion_read, > > + .write_with_attrs = iopmp_block_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static const MemoryRegionOps iopmp_block_r_ops = { > > + .fetch_with_attrs = iopmp_permssion_read, > > + .read_with_attrs = iopmp_block_read, > > + .write_with_attrs = iopmp_permssion_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static const MemoryRegionOps iopmp_block_rwx_ops = { > > + .fetch_with_attrs = iopmp_block_fetch, > > + .read_with_attrs = iopmp_block_read, > > + .write_with_attrs = iopmp_block_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static const MemoryRegionOps iopmp_block_wx_ops = { > > + .fetch_with_attrs = iopmp_block_fetch, > > + .read_with_attrs = iopmp_permssion_read, > > + .write_with_attrs = iopmp_block_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static const MemoryRegionOps iopmp_block_rx_ops = { > > + .fetch_with_attrs = iopmp_block_fetch, > > + .read_with_attrs = iopmp_block_read, > > + .write_with_attrs = iopmp_permssion_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static const MemoryRegionOps iopmp_block_x_ops = { > > + .fetch_with_attrs = iopmp_block_fetch, > > + .read_with_attrs = iopmp_permssion_read, > > + .write_with_attrs = iopmp_permssion_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static void iopmp_realize(DeviceState *dev, Error **errp) > > +{ > > + Object *obj = OBJECT(dev); > > + SysBusDevice *sbd = SYS_BUS_DEVICE(dev); > > + IopmpState *s = IOPMP(dev); > > + uint64_t size; > > + > > + size = -1ULL; > > + s->model = IOPMP_MODEL_RAPIDK; > > Should this be a property to allow other models in the future? Yes, it will be refined in next revision. > > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, s->md_num); > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, 1); > > + > > + s->prient_prog = s->default_prient_prog; > > + s->rrid_num = MIN(s->rrid_num, IOPMP_MAX_RRID_NUM); > > + s->md_num = MIN(s->md_num, IOPMP_MAX_MD_NUM); > > + s->entry_num = s->md_num * s->k; > > + s->prio_entry = MIN(s->prio_entry, s->entry_num); > > + > > + s->regs.mdcfg = g_malloc0(s->md_num * sizeof(uint32_t)); > > + s->regs.mdcfg[0] = s->k; > > + > > + s->regs.srcmd_en = g_malloc0(s->rrid_num * sizeof(uint32_t)); > > + s->regs.srcmd_enh = g_malloc0(s->rrid_num * sizeof(uint32_t)); > > + s->regs.entry = g_malloc0(s->entry_num * sizeof(iopmp_entry_t)); > > + s->entry_addr = g_malloc0(s->entry_num * sizeof(iopmp_addr_t)); > > + s->transaction_state = g_malloc0(s->rrid_num * > > + sizeof(iopmp_transaction_state)); > > + qemu_mutex_init(&s->iopmp_transaction_mutex); > > + > > + memory_region_init_iommu(&s->iommu, sizeof(s->iommu), > > + TYPE_IOPMP_IOMMU_MEMORY_REGION, > > + obj, "riscv-iopmp-sysbus-iommu", UINT64_MAX); > > + memory_region_init_io(&s->mmio, obj, &iopmp_ops, > > + s, "iopmp-regs", 0x100000); > > + sysbus_init_mmio(sbd, &s->mmio); > > + > > + memory_region_init_io(&s->blocked_rw, NULL, &iopmp_block_rw_ops, > > + s, "iopmp-blocked-rw", size); > > + memory_region_init_io(&s->blocked_w, NULL, &iopmp_block_w_ops, > > + s, "iopmp-blocked-w", size); > > + memory_region_init_io(&s->blocked_r, NULL, &iopmp_block_r_ops, > > + s, "iopmp-blocked-r", size); > > + > > + memory_region_init_io(&s->blocked_rwx, NULL, &iopmp_block_rwx_ops, > > + s, "iopmp-blocked-rwx", size); > > + memory_region_init_io(&s->blocked_wx, NULL, &iopmp_block_wx_ops, > > + s, "iopmp-blocked-wx", size); > > + memory_region_init_io(&s->blocked_rx, NULL, &iopmp_block_rx_ops, > > + s, "iopmp-blocked-rx", size); > > + memory_region_init_io(&s->blocked_x, NULL, &iopmp_block_x_ops, > > + s, "iopmp-blocked-x", size); > > + address_space_init(&s->blocked_rw_as, &s->blocked_rw, > > + "iopmp-blocked-rw-as"); > > + address_space_init(&s->blocked_w_as, &s->blocked_w, > > + "iopmp-blocked-w-as"); > > + address_space_init(&s->blocked_r_as, &s->blocked_r, > > + "iopmp-blocked-r-as"); > > + > > + address_space_init(&s->blocked_rwx_as, &s->blocked_rwx, > > + "iopmp-blocked-rwx-as"); > > + address_space_init(&s->blocked_wx_as, &s->blocked_wx, > > + "iopmp-blocked-wx-as"); > > + address_space_init(&s->blocked_rx_as, &s->blocked_rx, > > + "iopmp-blocked-rx-as"); > > + address_space_init(&s->blocked_x_as, &s->blocked_x, > > + "iopmp-blocked-x-as"); > > +} > > + > > +static void iopmp_reset(DeviceState *dev) > > +{ > > + IopmpState *s = IOPMP(dev); > > + > > + qemu_set_irq(s->irq, 0); > > + memset(s->regs.srcmd_en, 0, s->rrid_num * sizeof(uint32_t)); > > + memset(s->regs.srcmd_enh, 0, s->rrid_num * sizeof(uint32_t)); > > + memset(s->entry_addr, 0, s->entry_num * sizeof(iopmp_addr_t)); > > + > > + s->regs.mdlck = 0; > > + s->regs.mdlckh = 0; > > + s->regs.entrylck = 0; > > + s->regs.mdstall = 0; > > + s->regs.mdstallh = 0; > > + s->regs.rridscp = 0; > > + s->regs.err_cfg = 0; > > + s->regs.err_reqaddr = 0; > > + s->regs.err_reqid = 0; > > + s->regs.err_reqinfo = 0; > > + > > + s->prient_prog = s->default_prient_prog; > > + s->enable = 0; > > + > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, s->md_num); > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, 1); > > + s->regs.mdcfg[0] = s->k; > > +} > > + > > +static int iopmp_attrs_to_index(IOMMUMemoryRegion *iommu, MemTxAttrs attrs) > > +{ > > + return attrs.requester_id; > > +} > > + > > +static void iopmp_iommu_memory_region_class_init(ObjectClass *klass, void *data) > > +{ > > + IOMMUMemoryRegionClass *imrc = IOMMU_MEMORY_REGION_CLASS(klass); > > + > > + imrc->translate = iopmp_translate; > > + imrc->attrs_to_index = iopmp_attrs_to_index; > > +} > > + > > +static Property iopmp_property[] = { > > + DEFINE_PROP_BOOL("prient_prog", IopmpState, default_prient_prog, true), > > + DEFINE_PROP_UINT32("k", IopmpState, k, 6), > > + DEFINE_PROP_UINT32("prio_entry", IopmpState, prio_entry, 48), > > + DEFINE_PROP_UINT32("rrid_num", IopmpState, rrid_num, 16), > > + DEFINE_PROP_UINT32("md_num", IopmpState, md_num, 8), > > + DEFINE_PROP_UINT32("entry_offset", IopmpState, entry_offset, 0x4000), > > + DEFINE_PROP_UINT32("fabricated_v", IopmpState, fabricated_v, 0x0), > > + DEFINE_PROP_END_OF_LIST(), > > +}; > > + > > +static void iopmp_class_init(ObjectClass *klass, void *data) > > +{ > > + DeviceClass *dc = DEVICE_CLASS(klass); > > + device_class_set_props(dc, iopmp_property); > > + dc->realize = iopmp_realize; > > + dc->reset = iopmp_reset; > > +} > > + > > +static void iopmp_init(Object *obj) > > +{ > > + IopmpState *s = IOPMP(obj); > > + SysBusDevice *sbd = SYS_BUS_DEVICE(obj); > > + > > + sysbus_init_irq(sbd, &s->irq); > > +} > > + > > +static const TypeInfo iopmp_info = { > > + .name = TYPE_IOPMP, > > + .parent = TYPE_SYS_BUS_DEVICE, > > + .instance_size = sizeof(IopmpState), > > + .instance_init = iopmp_init, > > + .class_init = iopmp_class_init, > > +}; > > + > > +static const TypeInfo > > +iopmp_iommu_memory_region_info = { > > + .name = TYPE_IOPMP_IOMMU_MEMORY_REGION, > > + .parent = TYPE_IOMMU_MEMORY_REGION, > > + .class_init = iopmp_iommu_memory_region_class_init, > > +}; > > + > > +static void > > +iopmp_register_types(void) > > +{ > > + type_register_static(&iopmp_info); > > + type_register_static(&iopmp_iommu_memory_region_info); > > +} > > + > > +type_init(iopmp_register_types); > > diff --git a/hw/misc/trace-events b/hw/misc/trace-events > > index 1be0717c0c..c148166d2d 100644 > > --- a/hw/misc/trace-events > > +++ b/hw/misc/trace-events > > @@ -362,3 +362,6 @@ aspeed_sli_read(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx > > aspeed_sliio_write(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx64 " of size %u: 0x%" PRIx32 > > aspeed_sliio_read(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx64 " of size %u: 0x%" PRIx32 > > > > +# riscv_iopmp.c > > +iopmp_read(uint64_t addr, uint32_t val) "addr 0x%"PRIx64" val 0x%x" > > +iopmp_write(uint64_t addr, uint32_t val) "addr 0x%"PRIx64" val 0x%x" > > diff --git a/include/hw/misc/riscv_iopmp.h b/include/hw/misc/riscv_iopmp.h > > new file mode 100644 > > index 0000000000..b8fe479108 > > --- /dev/null > > +++ b/include/hw/misc/riscv_iopmp.h > > @@ -0,0 +1,168 @@ > > +/* > > + * QEMU RISC-V IOPMP (Input Output Physical Memory Protection) > > + * > > + * Copyright (c) 2023-2024 Andes Tech. Corp. > > + * > > + * SPDX-License-Identifier: GPL-2.0-or-later > > + * > > + * This program is free software; you can redistribute it and/or modify it > > + * under the terms and conditions of the GNU General Public License, > > + * version 2 or later, as published by the Free Software Foundation. > > + * > > + * This program is distributed in the hope it will be useful, but WITHOUT > > + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or > > + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for > > + * more details. > > + * > > + * You should have received a copy of the GNU General Public License along with > > + * this program. If not, see <http://www.gnu.org/licenses/>. > > + */ > > + > > +#ifndef RISCV_IOPMP_H > > +#define RISCV_IOPMP_H > > + > > +#include "hw/sysbus.h" > > +#include "qemu/typedefs.h" > > +#include "memory.h" > > +#include "exec/hwaddr.h" > > + > > +#define TYPE_IOPMP "iopmp" > > +#define IOPMP(obj) OBJECT_CHECK(IopmpState, (obj), TYPE_IOPMP) > > + > > +#define IOPMP_MAX_MD_NUM 63 > > +#define IOPMP_MAX_RRID_NUM 65535 > > +#define IOPMP_MAX_ENTRY_NUM 65535 > > + > > +#define VENDER_VIRT 0 > > +#define SPECVER_0_9_1 91 > > +#define IMPID_0_9_1 91 > > + > > +#define RRE_ERROR 0 > > +#define RRE_SUCCESS_VALUE 1 > > + > > +#define RWE_ERROR 0 > > +#define RWE_SUCCESS 1 > > + > > +#define RXE_ERROR 0 > > +#define RXE_SUCCESS_VALUE 1 > > + > > +#define ERR_REQINFO_TTYPE_READ 1 > > +#define ERR_REQINFO_TTYPE_WRITE 2 > > +#define ERR_REQINFO_TTYPE_FETCH 3 > > +#define ERR_REQINFO_ETYPE_NOERROR 0 > > +#define ERR_REQINFO_ETYPE_READ 1 > > +#define ERR_REQINFO_ETYPE_WRITE 2 > > +#define ERR_REQINFO_ETYPE_FETCH 3 > > +#define ERR_REQINFO_ETYPE_PARHIT 4 > > +#define ERR_REQINFO_ETYPE_NOHIT 5 > > +#define ERR_REQINFO_ETYPE_RRID 6 > > +#define ERR_REQINFO_ETYPE_USER 7 > > + > > +#define IOPMP_MODEL_FULL 0 > > +#define IOPMP_MODEL_RAPIDK 0x1 > > +#define IOPMP_MODEL_DYNAMICK 0x2 > > +#define IOPMP_MODEL_ISOLATION 0x3 > > +#define IOPMP_MODEL_COMPACTK 0x4 > > + > > +#define ENTRY_NO_HIT 0 > > +#define ENTRY_PAR_HIT 1 > > +#define ENTRY_HIT 2 > > Why not an enum? > Thank you for your suggestion. There will be enums in next version. Thanks, Ethan Chen > Alistair > > > + > > +/* The generic iopmp address space which downstream is system memory */ > > +extern AddressSpace iopmp_container_as; > > + > > +typedef enum { > > + IOPMP_AMATCH_OFF, /* Null (off) */ > > + IOPMP_AMATCH_TOR, /* Top of Range */ > > + IOPMP_AMATCH_NA4, /* Naturally aligned four-byte region */ > > + IOPMP_AMATCH_NAPOT /* Naturally aligned power-of-two region */ > > +} iopmp_am_t; > > + > > +typedef enum { > > + IOPMP_ACCESS_READ = 1, > > + IOPMP_ACCESS_WRITE = 2, > > + IOPMP_ACCESS_FETCH = 3 > > +} iopmp_access_type; > > + > > +typedef enum { > > + IOPMP_NONE = 0, > > + IOPMP_RO = 1, > > + IOPMP_WO = 2, > > + IOPMP_RW = 3, > > + IOPMP_XO = 4, > > + IOPMP_RX = 5, > > + IOPMP_WX = 6, > > + IOPMP_RWX = 7, > > +} iopmp_permission; > > + > > +typedef struct { > > + uint32_t addr_reg; > > + uint32_t addrh_reg; > > + uint32_t cfg_reg; > > +} iopmp_entry_t; > > + > > +typedef struct { > > + uint64_t sa; > > + uint64_t ea; > > +} iopmp_addr_t; > > + > > +typedef struct { > > + uint32_t *srcmd_en; > > + uint32_t *srcmd_enh; > > + uint32_t *mdcfg; > > + iopmp_entry_t *entry; > > + uint32_t mdlck; > > + uint32_t mdlckh; > > + uint32_t entrylck; > > + uint32_t mdcfglck; > > + uint32_t mdstall; > > + uint32_t mdstallh; > > + uint32_t rridscp; > > + uint32_t err_cfg; > > + uint64_t err_reqaddr; > > + uint32_t err_reqid; > > + uint32_t err_reqinfo; > > +} iopmp_regs; > > + > > + > > +/* To detect partially hit */ > > +typedef struct iopmp_transaction_state { > > + bool running; > > + bool supported; > > + hwaddr start_addr; > > + hwaddr end_addr; > > +} iopmp_transaction_state; > > + > > +typedef struct IopmpState { > > + SysBusDevice parent_obj; > > + iopmp_addr_t *entry_addr; > > + MemoryRegion mmio; > > + IOMMUMemoryRegion iommu; > > + IOMMUMemoryRegion *next_iommu; > > + iopmp_regs regs; > > + MemoryRegion *downstream; > > + MemoryRegion blocked_r, blocked_w, blocked_x, blocked_rw, blocked_rx, > > + blocked_wx, blocked_rwx; > > + MemoryRegion stall_io; > > + uint32_t model; > > + uint32_t k; > > + bool prient_prog; > > + bool default_prient_prog; > > + iopmp_transaction_state *transaction_state; > > + QemuMutex iopmp_transaction_mutex; > > + > > + AddressSpace downstream_as; > > + AddressSpace blocked_r_as, blocked_w_as, blocked_x_as, blocked_rw_as, > > + blocked_rx_as, blocked_wx_as, blocked_rwx_as; > > + qemu_irq irq; > > + bool enable; > > + > > + uint32_t prio_entry; > > + uint32_t rrid_num; > > + uint32_t md_num; > > + uint32_t entry_num; > > + uint32_t entry_offset; > > + uint32_t fabricated_v; > > +} IopmpState; > > + > > +#endif > > -- > > 2.34.1 > > > >
On Thu, Aug 08, 2024 at 01:56:35PM +1000, Alistair Francis wrote: > [EXTERNAL MAIL] > > On Mon, Jul 15, 2024 at 7:58 PM Ethan Chen via <qemu-devel@nongnu.org> wrote: > > > > Support basic functions of IOPMP specification v0.9.1 rapid-k model. > > The specification url: > > https://github.com/riscv-non-isa/iopmp-spec/releases/tag/v0.9.1 > > > > The IOPMP checks whether memory access from a device or CPU is valid. > > This implementation uses an IOMMU to modify the address space accessed > > by the device. > > > > For device access with IOMMUAccessFlags specifying read or write > > (IOMMU_RO or IOMMU_WO), the IOPMP checks the permission in > > iopmp_translate. If the access is valid, the target address space is > > downstream_as. If the access is blocked, it will be redirected to > > blocked_rwx_as. > > > > For CPU access with IOMMUAccessFlags not specifying read or write > > (IOMMU_NONE), the IOPMP translates the access to the corresponding > > address space based on the permission. If the access has full permission > > (rwx), the target address space is downstream_as. If the access has > > limited permission, the target address space is blocked_ followed by > > the lacked permissions. > > > > The operation of a blocked region can trigger an IOPMP interrupt, a bus > > error, or it can respond with success and fabricated data, depending on > > the value of the IOPMP ERR_CFG register. > > > > Signed-off-by: Ethan Chen <ethan84@andestech.com> > > --- > > hw/misc/Kconfig | 3 + > > hw/misc/meson.build | 1 + > > hw/misc/riscv_iopmp.c | 1154 +++++++++++++++++++++++++++++++++ > > hw/misc/trace-events | 3 + > > include/hw/misc/riscv_iopmp.h | 168 +++++ > > 5 files changed, 1329 insertions(+) > > create mode 100644 hw/misc/riscv_iopmp.c > > create mode 100644 include/hw/misc/riscv_iopmp.h > > > > diff --git a/hw/misc/Kconfig b/hw/misc/Kconfig > > index 1e08785b83..427f0c702e 100644 > > --- a/hw/misc/Kconfig > > +++ b/hw/misc/Kconfig > > @@ -213,4 +213,7 @@ config IOSB > > config XLNX_VERSAL_TRNG > > bool > > > > +config RISCV_IOPMP > > + bool > > + > > source macio/Kconfig > > diff --git a/hw/misc/meson.build b/hw/misc/meson.build > > index 2ca8717be2..d9006e1d81 100644 > > --- a/hw/misc/meson.build > > +++ b/hw/misc/meson.build > > @@ -34,6 +34,7 @@ system_ss.add(when: 'CONFIG_SIFIVE_E_PRCI', if_true: files('sifive_e_prci.c')) > > system_ss.add(when: 'CONFIG_SIFIVE_E_AON', if_true: files('sifive_e_aon.c')) > > system_ss.add(when: 'CONFIG_SIFIVE_U_OTP', if_true: files('sifive_u_otp.c')) > > system_ss.add(when: 'CONFIG_SIFIVE_U_PRCI', if_true: files('sifive_u_prci.c')) > > +specific_ss.add(when: 'CONFIG_RISCV_IOPMP', if_true: files('riscv_iopmp.c')) > > > > subdir('macio') > > > > diff --git a/hw/misc/riscv_iopmp.c b/hw/misc/riscv_iopmp.c > > new file mode 100644 > > index 0000000000..db43e3c73f > > --- /dev/null > > +++ b/hw/misc/riscv_iopmp.c > > @@ -0,0 +1,1154 @@ > > +/* > > + * QEMU RISC-V IOPMP (Input Output Physical Memory Protection) > > + * > > + * Copyright (c) 2023-2024 Andes Tech. Corp. > > + * > > + * SPDX-License-Identifier: GPL-2.0-or-later > > + * > > + * This program is free software; you can redistribute it and/or modify it > > + * under the terms and conditions of the GNU General Public License, > > + * version 2 or later, as published by the Free Software Foundation. > > + * > > + * This program is distributed in the hope it will be useful, but WITHOUT > > + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or > > + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for > > + * more details. > > + * > > + * You should have received a copy of the GNU General Public License along with > > + * this program. If not, see <http://www.gnu.org/licenses/>. > > + */ > > + > > +#include "qemu/osdep.h" > > +#include "qemu/log.h" > > +#include "qapi/error.h" > > +#include "trace.h" > > +#include "exec/exec-all.h" > > +#include "exec/address-spaces.h" > > +#include "hw/qdev-properties.h" > > +#include "hw/sysbus.h" > > +#include "hw/misc/riscv_iopmp.h" > > +#include "memory.h" > > +#include "hw/irq.h" > > +#include "hw/registerfields.h" > > +#include "trace.h" > > + > > +#define TYPE_IOPMP_IOMMU_MEMORY_REGION "iopmp-iommu-memory-region" > > + > > +REG32(VERSION, 0x00) > > + FIELD(VERSION, VENDOR, 0, 24) > > + FIELD(VERSION, SPECVER , 24, 8) > > +REG32(IMP, 0x04) > > + FIELD(IMP, IMPID, 0, 32) > > +REG32(HWCFG0, 0x08) > > + FIELD(HWCFG0, MODEL, 0, 4) > > + FIELD(HWCFG0, TOR_EN, 4, 1) > > + FIELD(HWCFG0, SPS_EN, 5, 1) > > + FIELD(HWCFG0, USER_CFG_EN, 6, 1) > > + FIELD(HWCFG0, PRIENT_PROG, 7, 1) > > + FIELD(HWCFG0, RRID_TRANSL_EN, 8, 1) > > + FIELD(HWCFG0, RRID_TRANSL_PROG, 9, 1) > > + FIELD(HWCFG0, CHK_X, 10, 1) > > + FIELD(HWCFG0, NO_X, 11, 1) > > + FIELD(HWCFG0, NO_W, 12, 1) > > + FIELD(HWCFG0, STALL_EN, 13, 1) > > + FIELD(HWCFG0, PEIS, 14, 1) > > + FIELD(HWCFG0, PEES, 15, 1) > > + FIELD(HWCFG0, MFR_EN, 16, 1) > > + FIELD(HWCFG0, MD_NUM, 24, 7) > > + FIELD(HWCFG0, ENABLE, 31, 1) > > +REG32(HWCFG1, 0x0C) > > + FIELD(HWCFG1, RRID_NUM, 0, 16) > > + FIELD(HWCFG1, ENTRY_NUM, 16, 16) > > +REG32(HWCFG2, 0x10) > > + FIELD(HWCFG2, PRIO_ENTRY, 0, 16) > > + FIELD(HWCFG2, RRID_TRANSL, 16, 16) > > +REG32(ENTRYOFFSET, 0x14) > > + FIELD(ENTRYOFFSET, OFFSET, 0, 32) > > +REG32(MDSTALL, 0x30) > > + FIELD(MDSTALL, EXEMPT, 0, 1) > > + FIELD(MDSTALL, MD, 1, 31) > > +REG32(MDSTALLH, 0x34) > > + FIELD(MDSTALLH, MD, 0, 32) > > +REG32(RRIDSCP, 0x38) > > + FIELD(RRIDSCP, RRID, 0, 16) > > + FIELD(RRIDSCP, OP, 30, 2) > > +REG32(MDLCK, 0x40) > > + FIELD(MDLCK, L, 0, 1) > > + FIELD(MDLCK, MD, 1, 31) > > +REG32(MDLCKH, 0x44) > > + FIELD(MDLCKH, MDH, 0, 32) > > +REG32(MDCFGLCK, 0x48) > > + FIELD(MDCFGLCK, L, 0, 1) > > + FIELD(MDCFGLCK, F, 1, 7) > > +REG32(ENTRYLCK, 0x4C) > > + FIELD(ENTRYLCK, L, 0, 1) > > + FIELD(ENTRYLCK, F, 1, 16) > > +REG32(ERR_CFG, 0x60) > > + FIELD(ERR_CFG, L, 0, 1) > > + FIELD(ERR_CFG, IE, 1, 1) > > + FIELD(ERR_CFG, IRE, 2, 1) > > + FIELD(ERR_CFG, IWE, 3, 1) > > + FIELD(ERR_CFG, IXE, 4, 1) > > + FIELD(ERR_CFG, RRE, 5, 1) > > + FIELD(ERR_CFG, RWE, 6, 1) > > + FIELD(ERR_CFG, RXE, 7, 1) > > +REG32(ERR_REQINFO, 0x64) > > + FIELD(ERR_REQINFO, V, 0, 1) > > + FIELD(ERR_REQINFO, TTYPE, 1, 2) > > + FIELD(ERR_REQINFO, ETYPE, 4, 3) > > + FIELD(ERR_REQINFO, SVC, 7, 1) > > +REG32(ERR_REQADDR, 0x68) > > + FIELD(ERR_REQADDR, ADDR, 0, 32) > > +REG32(ERR_REQADDRH, 0x6C) > > + FIELD(ERR_REQADDRH, ADDRH, 0, 32) > > +REG32(ERR_REQID, 0x70) > > + FIELD(ERR_REQID, RRID, 0, 16) > > + FIELD(ERR_REQID, EID, 16, 16) > > +REG32(ERR_MFR, 0x74) > > + FIELD(ERR_MFR, SVW, 0, 16) > > + FIELD(ERR_MFR, SVI, 16, 12) > > + FIELD(ERR_MFR, SVS, 31, 1) > > +REG32(MDCFG0, 0x800) > > + FIELD(MDCFG0, T, 0, 16) > > +REG32(SRCMD_EN0, 0x1000) > > + FIELD(SRCMD_EN0, L, 0, 1) > > + FIELD(SRCMD_EN0, MD, 1, 31) > > +REG32(SRCMD_ENH0, 0x1004) > > + FIELD(SRCMD_ENH0, MDH, 0, 32) > > +REG32(SRCMD_R0, 0x1008) > > + FIELD(SRCMD_R0, MD, 1, 31) > > +REG32(SRCMD_RH0, 0x100C) > > + FIELD(SRCMD_RH0, MDH, 0, 32) > > +REG32(SRCMD_W0, 0x1010) > > + FIELD(SRCMD_W0, MD, 1, 31) > > +REG32(SRCMD_WH0, 0x1014) > > + FIELD(SRCMD_WH0, MDH, 0, 32) > > + > > +FIELD(ENTRY_ADDR, ADDR, 0, 32) > > +FIELD(ENTRY_ADDRH, ADDRH, 0, 32) > > + > > +FIELD(ENTRY_CFG, R, 0, 1) > > +FIELD(ENTRY_CFG, W, 1, 1) > > +FIELD(ENTRY_CFG, X, 2, 1) > > +FIELD(ENTRY_CFG, A, 3, 2) > > +FIELD(ENTRY_CFG, SIRE, 5, 1) > > +FIELD(ENTRY_CFG, SIWE, 6, 1) > > +FIELD(ENTRY_CFG, SIXE, 7, 1) > > +FIELD(ENTRY_CFG, SERE, 8, 1) > > +FIELD(ENTRY_CFG, SEWE, 9, 1) > > +FIELD(ENTRY_CFG, SEXE, 10, 1) > > + > > +FIELD(ENTRY_USER_CFG, IM, 0, 32) > > + > > +/* Offsets to SRCMD_EN(i) */ > > +#define SRCMD_EN_OFFSET 0x0 > > +#define SRCMD_ENH_OFFSET 0x4 > > +#define SRCMD_R_OFFSET 0x8 > > +#define SRCMD_RH_OFFSET 0xC > > +#define SRCMD_W_OFFSET 0x10 > > +#define SRCMD_WH_OFFSET 0x14 > > + > > +/* Offsets to ENTRY_ADDR(i) */ > > +#define ENTRY_ADDR_OFFSET 0x0 > > +#define ENTRY_ADDRH_OFFSET 0x4 > > +#define ENTRY_CFG_OFFSET 0x8 > > +#define ENTRY_USER_CFG_OFFSET 0xC > > + > > +/* Memmap for parallel IOPMPs */ > > +typedef struct iopmp_protection_memmap { > > + MemMapEntry entry; > > + IopmpState *iopmp_s; > > + QLIST_ENTRY(iopmp_protection_memmap) list; > > +} iopmp_protection_memmap; > > +QLIST_HEAD(, iopmp_protection_memmap) > > + iopmp_protection_memmaps = QLIST_HEAD_INITIALIZER(iopmp_protection_memmaps); > > + > > +static void iopmp_iommu_notify(IopmpState *s) > > +{ > > + IOMMUTLBEvent event = { > > + .entry = { > > + .iova = 0, > > + .translated_addr = 0, > > + .addr_mask = -1ULL, > > + .perm = IOMMU_NONE, > > + }, > > + .type = IOMMU_NOTIFIER_UNMAP, > > + }; > > + > > + for (int i = 0; i < s->rrid_num; i++) { > > + memory_region_notify_iommu(&s->iommu, i, event); > > + } > > +} > > + > > +static void iopmp_decode_napot(uint64_t a, uint64_t *sa, > > + uint64_t *ea) > > +{ > > + /* > > + * aaaa...aaa0 8-byte NAPOT range > > + * aaaa...aa01 16-byte NAPOT range > > + * aaaa...a011 32-byte NAPOT range > > + * ... > > + * aa01...1111 2^XLEN-byte NAPOT range > > + * a011...1111 2^(XLEN+1)-byte NAPOT range > > + * 0111...1111 2^(XLEN+2)-byte NAPOT range > > + * 1111...1111 Reserved > > + */ > > + > > + a = (a << 2) | 0x3; > > + *sa = a & (a + 1); > > + *ea = a | (a + 1); > > +} > > + > > +static void iopmp_update_rule(IopmpState *s, uint32_t entry_index) > > +{ > > + uint8_t this_cfg = s->regs.entry[entry_index].cfg_reg; > > + uint64_t this_addr = s->regs.entry[entry_index].addr_reg | > > + ((uint64_t)s->regs.entry[entry_index].addrh_reg << 32); > > + uint64_t prev_addr = 0u; > > + uint64_t sa = 0u; > > + uint64_t ea = 0u; > > + > > + if (entry_index >= 1u) { > > + prev_addr = s->regs.entry[entry_index - 1].addr_reg | > > + ((uint64_t)s->regs.entry[entry_index - 1].addrh_reg << 32); > > + } > > + > > + switch (FIELD_EX32(this_cfg, ENTRY_CFG, A)) { > > + case IOPMP_AMATCH_OFF: > > + sa = 0u; > > + ea = -1; > > + break; > > + > > + case IOPMP_AMATCH_TOR: > > + sa = (prev_addr) << 2; /* shift up from [xx:0] to [xx+2:2] */ > > + ea = ((this_addr) << 2) - 1u; > > + if (sa > ea) { > > + sa = ea = 0u; > > + } > > + break; > > + > > + case IOPMP_AMATCH_NA4: > > + sa = this_addr << 2; /* shift up from [xx:0] to [xx+2:2] */ > > + ea = (sa + 4u) - 1u; > > + break; > > + > > + case IOPMP_AMATCH_NAPOT: > > + iopmp_decode_napot(this_addr, &sa, &ea); > > + break; > > + > > + default: > > + sa = 0u; > > + ea = 0u; > > + break; > > + } > > + > > + s->entry_addr[entry_index].sa = sa; > > + s->entry_addr[entry_index].ea = ea; > > + iopmp_iommu_notify(s); > > +} > > + > > +static uint64_t iopmp_read(void *opaque, hwaddr addr, unsigned size) > > +{ > > + IopmpState *s = IOPMP(opaque); > > + uint32_t rz = 0; > > + uint32_t offset, idx; > > + > > + switch (addr) { > > + case A_VERSION: > > + rz = VENDER_VIRT << R_VERSION_VENDOR_SHIFT | > > + SPECVER_0_9_1 << R_VERSION_SPECVER_SHIFT; > > It would be better to use the FIELD_DP32() macro instead of the manual shifts It will be refined in next revision. > > > + break; > > + case A_IMP: > > + rz = IMPID_0_9_1; > > + break; > > + case A_HWCFG0: > > + rz = s->model << R_HWCFG0_MODEL_SHIFT | > > + 1 << R_HWCFG0_TOR_EN_SHIFT | > > + 0 << R_HWCFG0_SPS_EN_SHIFT | > > + 0 << R_HWCFG0_USER_CFG_EN_SHIFT | > > + s->prient_prog << R_HWCFG0_PRIENT_PROG_SHIFT | > > + 0 << R_HWCFG0_RRID_TRANSL_EN_SHIFT | > > + 0 << R_HWCFG0_RRID_TRANSL_PROG_SHIFT | > > + 1 << R_HWCFG0_CHK_X_SHIFT | > > + 0 << R_HWCFG0_NO_X_SHIFT | > > + 0 << R_HWCFG0_NO_W_SHIFT | > > + 0 << R_HWCFG0_STALL_EN_SHIFT | > > + 0 << R_HWCFG0_PEIS_SHIFT | > > + 0 << R_HWCFG0_PEES_SHIFT | > > + 0 << R_HWCFG0_MFR_EN_SHIFT | > > + s->md_num << R_HWCFG0_MD_NUM_SHIFT | > > + s->enable << R_HWCFG0_ENABLE_SHIFT ; > > + break; > > + case A_HWCFG1: > > + rz = s->rrid_num << R_HWCFG1_RRID_NUM_SHIFT | > > + s->entry_num << R_HWCFG1_ENTRY_NUM_SHIFT; > > + break; > > + case A_HWCFG2: > > + rz = s->prio_entry << R_HWCFG2_PRIO_ENTRY_SHIFT; > > + break; > > + case A_ENTRYOFFSET: > > + rz = s->entry_offset; > > + break; > > + case A_ERR_CFG: > > + rz = s->regs.err_cfg; > > + break; > > + case A_MDLCK: > > + rz = s->regs.mdlck; > > + break; > > + case A_MDLCKH: > > + rz = s->regs.mdlckh; > > + break; > > + case A_MDCFGLCK: > > + rz = s->regs.mdcfglck; > > + break; > > + case A_ENTRYLCK: > > + rz = s->regs.entrylck; > > + break; > > + case A_ERR_REQADDR: > > + rz = s->regs.err_reqaddr & UINT32_MAX; > > + break; > > + case A_ERR_REQADDRH: > > + rz = s->regs.err_reqaddr >> 32; > > + break; > > + case A_ERR_REQID: > > + rz = s->regs.err_reqid; > > + break; > > + case A_ERR_REQINFO: > > + rz = s->regs.err_reqinfo; > > + break; > > + > > + default: > > + if (addr >= A_MDCFG0 && > > + addr < A_MDCFG0 + 4 * (s->md_num - 1)) { > > + offset = addr - A_MDCFG0; > > + idx = offset >> 2; > > + if (idx == 0 && offset == 0) { > > + rz = s->regs.mdcfg[idx]; > > + } else { > > + /* Only MDCFG0 is implemented in rapid-k model */ > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + } > > + } else if (addr >= A_SRCMD_EN0 && > > + addr < A_SRCMD_WH0 + 32 * (s->rrid_num - 1)) { > > + offset = addr - A_SRCMD_EN0; > > + idx = offset >> 5; > > + offset &= 0x1f; > > + > > + switch (offset) { > > + case SRCMD_EN_OFFSET: > > + rz = s->regs.srcmd_en[idx]; > > + break; > > + case SRCMD_ENH_OFFSET: > > + rz = s->regs.srcmd_enh[idx]; > > + break; > > + default: > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + break; > > + } > > + } else if (addr >= s->entry_offset && > > + addr < s->entry_offset + ENTRY_USER_CFG_OFFSET + > > + 16 * (s->entry_num - 1)) { > > + offset = addr - s->entry_offset; > > + idx = offset >> 4; > > + offset &= 0xf; > > + > > + switch (offset) { > > + case ENTRY_ADDR_OFFSET: > > + rz = s->regs.entry[idx].addr_reg; > > + break; > > + case ENTRY_ADDRH_OFFSET: > > + rz = s->regs.entry[idx].addrh_reg; > > + break; > > + case ENTRY_CFG_OFFSET: > > + rz = s->regs.entry[idx].cfg_reg; > > + break; > > + case ENTRY_USER_CFG_OFFSET: > > + /* Does not support user customized permission */ > > + rz = 0; > > + break; > > + default: > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + break; > > + } > > + } else { > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + } > > + break; > > + } > > + trace_iopmp_read(addr, rz); > > + return rz; > > +} > > + > > +static void > > +iopmp_write(void *opaque, hwaddr addr, uint64_t value, unsigned size) > > +{ > > + IopmpState *s = IOPMP(opaque); > > + uint32_t offset, idx; > > + uint32_t value32 = value; > > + > > + trace_iopmp_write(addr, value32); > > + > > + switch (addr) { > > + case A_VERSION: /* RO */ > > + break; > > + case A_IMP: /* RO */ > > + break; > > + case A_HWCFG0: > > + if (FIELD_EX32(value32, HWCFG0, PRIENT_PROG)) { > > + /* W1C */ > > + s->prient_prog = 0; > > + } > > + if (FIELD_EX32(value32, HWCFG0, ENABLE)) { > > + /* W1S */ > > + s->enable = 1; > > + iopmp_iommu_notify(s); > > + } > > + break; > > + case A_HWCFG1: /* RO */ > > + break; > > + case A_HWCFG2: > > + if (s->prient_prog) { > > + s->prio_entry = FIELD_EX32(value32, HWCFG2, PRIO_ENTRY); > > + } > > + break; > > + case A_ERR_CFG: > > + if (!FIELD_EX32(s->regs.err_cfg, ERR_CFG, L)) { > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, L, > > + FIELD_EX32(value32, ERR_CFG, L)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IE, > > + FIELD_EX32(value32, ERR_CFG, IE)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IRE, > > + FIELD_EX32(value32, ERR_CFG, IRE)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RRE, > > + FIELD_EX32(value32, ERR_CFG, RRE)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IWE, > > + FIELD_EX32(value32, ERR_CFG, IWE)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RWE, > > + FIELD_EX32(value32, ERR_CFG, RWE)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IXE, > > + FIELD_EX32(value32, ERR_CFG, IXE)); > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RXE, > > + FIELD_EX32(value32, ERR_CFG, RXE)); > > + } > > + break; > > + case A_MDLCK: > > + if (!FIELD_EX32(s->regs.mdlck, MDLCK, L)) { > > + s->regs.mdlck = value32; > > + } > > + break; > > + case A_MDLCKH: > > + if (!FIELD_EX32(s->regs.mdlck, MDLCK, L)) { > > + s->regs.mdlckh = value32; > > + } > > + break; > > + case A_MDCFGLCK: > > + if (!FIELD_EX32(s->regs.mdcfglck, MDCFGLCK, L)) { > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, > > + FIELD_EX32(value32, MDCFGLCK, F)); > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, > > + FIELD_EX32(value32, MDCFGLCK, L)); > > + } > > + break; > > + case A_ENTRYLCK: > > + if (!(FIELD_EX32(s->regs.entrylck, ENTRYLCK, L))) { > > + s->regs.entrylck = FIELD_DP32(s->regs.entrylck, ENTRYLCK, F, > > + FIELD_EX32(value32, ENTRYLCK, F)); > > + s->regs.entrylck = FIELD_DP32(s->regs.entrylck, ENTRYLCK, L, > > + FIELD_EX32(value32, ENTRYLCK, L)); > > + } > > + case A_ERR_REQADDR: /* RO */ > > + break; > > + case A_ERR_REQADDRH: /* RO */ > > + break; > > + case A_ERR_REQID: /* RO */ > > + break; > > + case A_ERR_REQINFO: > > + if (FIELD_EX32(value32, ERR_REQINFO, V)) { > > + s->regs.err_reqinfo = FIELD_DP32(s->regs.err_reqinfo, > > + ERR_REQINFO, V, 0); > > + qemu_set_irq(s->irq, 0); > > + } > > + break; > > + > > + default: > > + if (addr >= A_MDCFG0 && > > + addr < A_MDCFG0 + 4 * (s->md_num - 1)) { > > + offset = addr - A_MDCFG0; > > + idx = offset >> 2; > > + /* RO in rapid-k model */ > > + if (idx > 0) { > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + } > > + } else if (addr >= A_SRCMD_EN0 && > > + addr < A_SRCMD_WH0 + 32 * (s->rrid_num - 1)) { > > + offset = addr - A_SRCMD_EN0; > > + idx = offset >> 5; > > + offset &= 0x1f; > > + > > + if (offset % 4) { > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + } else if (FIELD_EX32(s->regs.srcmd_en[idx], SRCMD_EN0, L) > > + == 0) { > > + switch (offset) { > > + case SRCMD_EN_OFFSET: > > + s->regs.srcmd_en[idx] = > > + FIELD_DP32(s->regs.srcmd_en[idx], SRCMD_EN0, L, > > + FIELD_EX32(value32, SRCMD_EN0, L)); > > + > > + /* MD field is protected by mdlck */ > > + value32 = (value32 & ~s->regs.mdlck) | > > + (s->regs.srcmd_en[idx] & s->regs.mdlck); > > + s->regs.srcmd_en[idx] = > > + FIELD_DP32(s->regs.srcmd_en[idx], SRCMD_EN0, MD, > > + FIELD_EX32(value32, SRCMD_EN0, MD)); > > + break; > > + case SRCMD_ENH_OFFSET: > > + value32 = (value32 & ~s->regs.mdlckh) | > > + (s->regs.srcmd_enh[idx] & s->regs.mdlckh); > > + s->regs.srcmd_enh[idx] = > > + FIELD_DP32(s->regs.srcmd_enh[idx], SRCMD_ENH0, MDH, > > + value32); > > + break; > > + default: > > + break; > > + } > > + } > > + } else if (addr >= s->entry_offset && > > + addr < s->entry_offset + ENTRY_USER_CFG_OFFSET > > + + 16 * (s->entry_num - 1)) { > > + offset = addr - s->entry_offset; > > + idx = offset >> 4; > > + offset &= 0xf; > > + > > + /* index < ENTRYLCK_F is protected */ > > + if (idx >= FIELD_EX32(s->regs.entrylck, ENTRYLCK, F)) { > > + switch (offset) { > > + case ENTRY_ADDR_OFFSET: > > + s->regs.entry[idx].addr_reg = value32; > > + break; > > + case ENTRY_ADDRH_OFFSET: > > + s->regs.entry[idx].addrh_reg = value32; > > + break; > > + case ENTRY_CFG_OFFSET: > > + s->regs.entry[idx].cfg_reg = value32; > > + break; > > + case ENTRY_USER_CFG_OFFSET: > > + /* Does not support user customized permission */ > > + break; > > + default: > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > + __func__, (int)addr); > > + break; > > + } > > + iopmp_update_rule(s, idx); > > + if (idx + 1 < s->entry_num && > > + FIELD_EX32(s->regs.entry[idx + 1].cfg_reg, ENTRY_CFG, A) == > > + IOPMP_AMATCH_TOR) { > > + iopmp_update_rule(s, idx + 1); > > + } > > + } > > + } else { > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", __func__, > > + (int)addr); > > + } > > + } > > +} > > + > > +/* Match entry in memory domain */ > > +static int match_entry_md(IopmpState *s, int md_idx, hwaddr start_addr, > > + hwaddr end_addr, int *entry_idx, > > + int *prior_entry_in_tlb) > > +{ > > + int entry_idx_s, entry_idx_e; > > + int result = ENTRY_NO_HIT; > > + int i = 0; > > + hwaddr tlb_sa = start_addr & ~(TARGET_PAGE_SIZE - 1); > > + hwaddr tlb_ea = tlb_sa + TARGET_PAGE_SIZE - 1; > > + > > + entry_idx_s = md_idx * s->regs.mdcfg[0]; > > + entry_idx_e = (md_idx + 1) * s->regs.mdcfg[0]; > > + > > + if (entry_idx_s >= s->entry_num) { > > + return result; > > + } > > + if (entry_idx_e > s->entry_num) { > > + entry_idx_e = s->entry_num; > > + } > > + i = entry_idx_s; > > + for (i = entry_idx_s; i < entry_idx_e; i++) { > > + if (FIELD_EX32(s->regs.entry[i].cfg_reg, ENTRY_CFG, A) == > > + IOPMP_AMATCH_OFF) { > > + continue; > > + } > > + if (start_addr >= s->entry_addr[i].sa && > > + start_addr <= s->entry_addr[i].ea) { > > + /* Check end address */ > > + if (end_addr >= s->entry_addr[i].sa && > > + end_addr <= s->entry_addr[i].ea) { > > + *entry_idx = i; > > + return ENTRY_HIT; > > + } else if (i >= s->prio_entry) { > > + /* Continue for non-prio_entry */ > > + continue; > > + } else { > > + *entry_idx = i; > > + return ENTRY_PAR_HIT; > > + } > > + } else if (end_addr >= s->entry_addr[i].sa && > > + end_addr <= s->entry_addr[i].ea) { > > + /* Only end address matches the entry */ > > + if (i >= s->prio_entry) { > > + continue; > > + } else { > > + *entry_idx = i; > > + return ENTRY_PAR_HIT; > > + } > > + } else if (start_addr < s->entry_addr[i].sa && > > + end_addr > s->entry_addr[i].ea) { > > + if (i >= s->prio_entry) { > > + continue; > > + } else { > > + *entry_idx = i; > > + return ENTRY_PAR_HIT; > > + } > > + } > > + if (prior_entry_in_tlb != NULL) { > > + if ((s->entry_addr[i].sa >= tlb_sa && > > + s->entry_addr[i].sa <= tlb_ea) || > > + (s->entry_addr[i].ea >= tlb_sa && > > + s->entry_addr[i].ea <= tlb_ea)) { > > + /* > > + * TLB should not use the cached result when the tlb contains > > + * higher priority entry > > + */ > > + *prior_entry_in_tlb = 1; > > + } > > + } > > + } > > + return result; > > +} > > + > > +static int match_entry(IopmpState *s, int rrid, hwaddr start_addr, > > + hwaddr end_addr, int *match_md_idx, > > + int *match_entry_idx, int *prior_entry_in_tlb) > > +{ > > + int cur_result = ENTRY_NO_HIT; > > + int result = ENTRY_NO_HIT; > > + /* Remove lock bit */ > > + uint64_t srcmd_en = ((uint64_t)s->regs.srcmd_en[rrid] | > > + ((uint64_t)s->regs.srcmd_enh[rrid] << 32)) >> 1; > > + > > + for (int md_idx = 0; md_idx < s->md_num; md_idx++) { > > + if (srcmd_en & (1ULL << md_idx)) { > > + cur_result = match_entry_md(s, md_idx, start_addr, end_addr, > > + match_entry_idx, prior_entry_in_tlb); > > + if (cur_result == ENTRY_HIT || cur_result == ENTRY_PAR_HIT) { > > + *match_md_idx = md_idx; > > + return cur_result; > > + } > > + } > > + } > > + return result; > > +} > > + > > +static void iopmp_error_reaction(IopmpState *s, uint32_t id, hwaddr start, > > + uint32_t info) > > +{ > > + if (!FIELD_EX32(s->regs.err_reqinfo, ERR_REQINFO, V)) { > > + s->regs.err_reqinfo = info; > > + s->regs.err_reqinfo = FIELD_DP32(s->regs.err_reqinfo, ERR_REQINFO, V, > > + 1); > > + s->regs.err_reqid = id; > > + /* addr[LEN+2:2] */ > > + s->regs.err_reqaddr = start >> 2; > > + > > + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_READ && > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IRE)) { > > + qemu_set_irq(s->irq, 1); > > + } > > + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_WRITE && > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IWE)) { > > + qemu_set_irq(s->irq, 1); > > + } > > + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_FETCH && > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IXE)) { > > + qemu_set_irq(s->irq, 1); > > + } > > + } > > +} > > + > > +static IOMMUTLBEntry iopmp_translate(IOMMUMemoryRegion *iommu, hwaddr addr, > > + IOMMUAccessFlags flags, int iommu_idx) > > +{ > > + int rrid = iommu_idx; > > + IopmpState *s = IOPMP(container_of(iommu, IopmpState, iommu)); > > + hwaddr start_addr, end_addr; > > + int entry_idx = -1; > > + int md_idx = -1; > > + int result; > > + uint32_t error_info = 0; > > + uint32_t error_id = 0; > > + int prior_entry_in_tlb = 0; > > + iopmp_permission iopmp_perm; > > + IOMMUTLBEntry entry = { > > + .target_as = &s->downstream_as, > > + .iova = addr, > > + .translated_addr = addr, > > + .addr_mask = 0, > > + .perm = IOMMU_NONE, > > + }; > > + > > + if (!s->enable) { > > + /* Bypass IOPMP */ > > + entry.addr_mask = -1ULL, > > + entry.perm = IOMMU_RW; > > + return entry; > > + } > > + > > + /* unknown RRID */ > > + if (rrid >= s->rrid_num) { > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + ERR_REQINFO_ETYPE_RRID); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > > + iopmp_error_reaction(s, error_id, addr, error_info); > > + entry.target_as = &s->blocked_rwx_as; > > + entry.perm = IOMMU_RW; > > + return entry; > > + } > > + > > + if (s->transaction_state[rrid].supported == true) { > > + start_addr = s->transaction_state[rrid].start_addr; > > + end_addr = s->transaction_state[rrid].end_addr; > > + } else { > > + /* No transaction information, use the same address */ > > + start_addr = addr; > > + end_addr = addr; > > + } > > + > > + result = match_entry(s, rrid, start_addr, end_addr, &md_idx, &entry_idx, > > + &prior_entry_in_tlb); > > + if (result == ENTRY_HIT) { > > + entry.addr_mask = s->entry_addr[entry_idx].ea - > > + s->entry_addr[entry_idx].sa; > > + if (prior_entry_in_tlb) { > > + /* Make TLB repeat iommu translation on every access */ > > I don't follow this, if we have a prior entry in the TLB cache we > don't cache the accesses? For the cached TLB result to be used, the highest-priority entry in the TLB must occupy the entire TLB page. If a lower-priority entry fills the entire TLB page, it is still necessary to check which entry the transaction hits on each access to the TLB page. > > > + entry.addr_mask = 0; > > + } > > + iopmp_perm = s->regs.entry[entry_idx].cfg_reg & IOPMP_RWX; > > + if (flags) { > > + if ((iopmp_perm & flags) == 0) { > > + /* Permission denied */ > > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + ERR_REQINFO_ETYPE_READ + flags - 1); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > + entry.target_as = &s->blocked_rwx_as; > > + entry.perm = IOMMU_RW; > > + } else { > > + entry.target_as = &s->downstream_as; > > + entry.perm = iopmp_perm; > > + } > > + } else { > > + /* CPU access with IOMMU_NONE flag */ > > + if (iopmp_perm & IOPMP_XO) { > > + if ((iopmp_perm & IOPMP_RW) == IOPMP_RW) { > > + entry.target_as = &s->downstream_as; > > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_RO) { > > + entry.target_as = &s->blocked_w_as; > > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_WO) { > > + entry.target_as = &s->blocked_r_as; > > + } else { > > + entry.target_as = &s->blocked_rw_as; > > + } > > + } else { > > + if ((iopmp_perm & IOPMP_RW) == IOMMU_RW) { > > + entry.target_as = &s->blocked_x_as; > > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_RO) { > > + entry.target_as = &s->blocked_wx_as; > > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_WO) { > > + entry.target_as = &s->blocked_rx_as; > > + } else { > > + entry.target_as = &s->blocked_rwx_as; > > + } > > + } > > + entry.perm = IOMMU_RW; > > + } > > + } else { > > + if (flags) { > > + if (result == ENTRY_PAR_HIT) { > > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + ERR_REQINFO_ETYPE_PARHIT); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > + } else { > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + ERR_REQINFO_ETYPE_NOHIT); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > + } > > + } > > + /* CPU access with IOMMU_NONE flag no_hit or par_hit */ > > + entry.target_as = &s->blocked_rwx_as; > > + entry.perm = IOMMU_RW; > > + } > > + return entry; > > +} > > + > > +static const MemoryRegionOps iopmp_ops = { > > + .read = iopmp_read, > > + .write = iopmp_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 4, .max_access_size = 4} > > +}; > > + > > +static MemTxResult iopmp_permssion_write(void *opaque, hwaddr addr, > > + uint64_t value, unsigned size, > > + MemTxAttrs attrs) > > +{ > > + IopmpState *s = IOPMP(opaque); > > + return address_space_write(&s->downstream_as, addr, attrs, &value, size); > > +} > > + > > +static MemTxResult iopmp_permssion_read(void *opaque, hwaddr addr, > > + uint64_t *pdata, unsigned size, > > + MemTxAttrs attrs) > > +{ > > + IopmpState *s = IOPMP(opaque); > > + return address_space_read(&s->downstream_as, addr, attrs, pdata, size); > > +} > > + > > +static MemTxResult iopmp_handle_block(void *opaque, hwaddr addr, > > + uint64_t *data, unsigned size, > > + MemTxAttrs attrs, > > + iopmp_access_type access_type) { > > + IopmpState *s = IOPMP(opaque); > > + int md_idx, entry_idx; > > + uint32_t error_info = 0; > > + uint32_t error_id = 0; > > + int rrid = attrs.requester_id; > > + int result; > > + hwaddr start_addr, end_addr; > > + start_addr = addr; > > + end_addr = addr; > > + result = match_entry(s, rrid, start_addr, end_addr, &md_idx, &entry_idx, > > + NULL); > > + > > + if (result == ENTRY_HIT) { > > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + access_type); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, access_type); > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > + } else if (result == ENTRY_PAR_HIT) { > > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + ERR_REQINFO_ETYPE_PARHIT); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, > > + access_type); > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > + } else { > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > + ERR_REQINFO_ETYPE_NOHIT); > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, access_type); > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > + } > > + > > + if (access_type == IOPMP_ACCESS_READ) { > > + > > + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RRE)) { > > + case RRE_ERROR: > > + return MEMTX_ERROR; > > + break; > > + case RRE_SUCCESS_VALUE: > > + *data = s->fabricated_v; > > + return MEMTX_OK; > > + break; > > + default: > > + break; > > + } > > + return MEMTX_OK; > > + } else if (access_type == IOPMP_ACCESS_WRITE) { > > + > > + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RWE)) { > > + case RWE_ERROR: > > + return MEMTX_ERROR; > > + break; > > + case RWE_SUCCESS: > > + return MEMTX_OK; > > + break; > > + default: > > + break; > > + } > > + return MEMTX_OK; > > + } else { > > + > > + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RXE)) { > > + case RXE_ERROR: > > + return MEMTX_ERROR; > > + break; > > + case RXE_SUCCESS_VALUE: > > + *data = s->fabricated_v; > > + return MEMTX_OK; > > + break; > > + default: > > + break; > > + } > > + return MEMTX_OK; > > + } > > + return MEMTX_OK; > > +} > > + > > +static MemTxResult iopmp_block_write(void *opaque, hwaddr addr, uint64_t value, > > + unsigned size, MemTxAttrs attrs) > > +{ > > + return iopmp_handle_block(opaque, addr, &value, size, attrs, > > + IOPMP_ACCESS_WRITE); > > +} > > + > > +static MemTxResult iopmp_block_read(void *opaque, hwaddr addr, uint64_t *pdata, > > + unsigned size, MemTxAttrs attrs) > > +{ > > + return iopmp_handle_block(opaque, addr, pdata, size, attrs, > > + IOPMP_ACCESS_READ); > > +} > > + > > +static MemTxResult iopmp_block_fetch(void *opaque, hwaddr addr, uint64_t *pdata, > > + unsigned size, MemTxAttrs attrs) > > +{ > > + return iopmp_handle_block(opaque, addr, pdata, size, attrs, > > + IOPMP_ACCESS_FETCH); > > +} > > + > > +static const MemoryRegionOps iopmp_block_rw_ops = { > > + .fetch_with_attrs = iopmp_permssion_read, > > + .read_with_attrs = iopmp_block_read, > > + .write_with_attrs = iopmp_block_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static const MemoryRegionOps iopmp_block_w_ops = { > > + .fetch_with_attrs = iopmp_permssion_read, > > + .read_with_attrs = iopmp_permssion_read, > > + .write_with_attrs = iopmp_block_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static const MemoryRegionOps iopmp_block_r_ops = { > > + .fetch_with_attrs = iopmp_permssion_read, > > + .read_with_attrs = iopmp_block_read, > > + .write_with_attrs = iopmp_permssion_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static const MemoryRegionOps iopmp_block_rwx_ops = { > > + .fetch_with_attrs = iopmp_block_fetch, > > + .read_with_attrs = iopmp_block_read, > > + .write_with_attrs = iopmp_block_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static const MemoryRegionOps iopmp_block_wx_ops = { > > + .fetch_with_attrs = iopmp_block_fetch, > > + .read_with_attrs = iopmp_permssion_read, > > + .write_with_attrs = iopmp_block_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static const MemoryRegionOps iopmp_block_rx_ops = { > > + .fetch_with_attrs = iopmp_block_fetch, > > + .read_with_attrs = iopmp_block_read, > > + .write_with_attrs = iopmp_permssion_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static const MemoryRegionOps iopmp_block_x_ops = { > > + .fetch_with_attrs = iopmp_block_fetch, > > + .read_with_attrs = iopmp_permssion_read, > > + .write_with_attrs = iopmp_permssion_write, > > + .endianness = DEVICE_NATIVE_ENDIAN, > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > +}; > > + > > +static void iopmp_realize(DeviceState *dev, Error **errp) > > +{ > > + Object *obj = OBJECT(dev); > > + SysBusDevice *sbd = SYS_BUS_DEVICE(dev); > > + IopmpState *s = IOPMP(dev); > > + uint64_t size; > > + > > + size = -1ULL; > > + s->model = IOPMP_MODEL_RAPIDK; > > Should this be a property to allow other models in the future? Sure. It will be added in next revision. > > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, s->md_num); > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, 1); > > + > > + s->prient_prog = s->default_prient_prog; > > + s->rrid_num = MIN(s->rrid_num, IOPMP_MAX_RRID_NUM); > > + s->md_num = MIN(s->md_num, IOPMP_MAX_MD_NUM); > > + s->entry_num = s->md_num * s->k; > > + s->prio_entry = MIN(s->prio_entry, s->entry_num); > > + > > + s->regs.mdcfg = g_malloc0(s->md_num * sizeof(uint32_t)); > > + s->regs.mdcfg[0] = s->k; > > + > > + s->regs.srcmd_en = g_malloc0(s->rrid_num * sizeof(uint32_t)); > > + s->regs.srcmd_enh = g_malloc0(s->rrid_num * sizeof(uint32_t)); > > + s->regs.entry = g_malloc0(s->entry_num * sizeof(iopmp_entry_t)); > > + s->entry_addr = g_malloc0(s->entry_num * sizeof(iopmp_addr_t)); > > + s->transaction_state = g_malloc0(s->rrid_num * > > + sizeof(iopmp_transaction_state)); > > + qemu_mutex_init(&s->iopmp_transaction_mutex); > > + > > + memory_region_init_iommu(&s->iommu, sizeof(s->iommu), > > + TYPE_IOPMP_IOMMU_MEMORY_REGION, > > + obj, "riscv-iopmp-sysbus-iommu", UINT64_MAX); > > + memory_region_init_io(&s->mmio, obj, &iopmp_ops, > > + s, "iopmp-regs", 0x100000); > > + sysbus_init_mmio(sbd, &s->mmio); > > + > > + memory_region_init_io(&s->blocked_rw, NULL, &iopmp_block_rw_ops, > > + s, "iopmp-blocked-rw", size); > > + memory_region_init_io(&s->blocked_w, NULL, &iopmp_block_w_ops, > > + s, "iopmp-blocked-w", size); > > + memory_region_init_io(&s->blocked_r, NULL, &iopmp_block_r_ops, > > + s, "iopmp-blocked-r", size); > > + > > + memory_region_init_io(&s->blocked_rwx, NULL, &iopmp_block_rwx_ops, > > + s, "iopmp-blocked-rwx", size); > > + memory_region_init_io(&s->blocked_wx, NULL, &iopmp_block_wx_ops, > > + s, "iopmp-blocked-wx", size); > > + memory_region_init_io(&s->blocked_rx, NULL, &iopmp_block_rx_ops, > > + s, "iopmp-blocked-rx", size); > > + memory_region_init_io(&s->blocked_x, NULL, &iopmp_block_x_ops, > > + s, "iopmp-blocked-x", size); > > + address_space_init(&s->blocked_rw_as, &s->blocked_rw, > > + "iopmp-blocked-rw-as"); > > + address_space_init(&s->blocked_w_as, &s->blocked_w, > > + "iopmp-blocked-w-as"); > > + address_space_init(&s->blocked_r_as, &s->blocked_r, > > + "iopmp-blocked-r-as"); > > + > > + address_space_init(&s->blocked_rwx_as, &s->blocked_rwx, > > + "iopmp-blocked-rwx-as"); > > + address_space_init(&s->blocked_wx_as, &s->blocked_wx, > > + "iopmp-blocked-wx-as"); > > + address_space_init(&s->blocked_rx_as, &s->blocked_rx, > > + "iopmp-blocked-rx-as"); > > + address_space_init(&s->blocked_x_as, &s->blocked_x, > > + "iopmp-blocked-x-as"); > > +} > > + > > +static void iopmp_reset(DeviceState *dev) > > +{ > > + IopmpState *s = IOPMP(dev); > > + > > + qemu_set_irq(s->irq, 0); > > + memset(s->regs.srcmd_en, 0, s->rrid_num * sizeof(uint32_t)); > > + memset(s->regs.srcmd_enh, 0, s->rrid_num * sizeof(uint32_t)); > > + memset(s->entry_addr, 0, s->entry_num * sizeof(iopmp_addr_t)); > > + > > + s->regs.mdlck = 0; > > + s->regs.mdlckh = 0; > > + s->regs.entrylck = 0; > > + s->regs.mdstall = 0; > > + s->regs.mdstallh = 0; > > + s->regs.rridscp = 0; > > + s->regs.err_cfg = 0; > > + s->regs.err_reqaddr = 0; > > + s->regs.err_reqid = 0; > > + s->regs.err_reqinfo = 0; > > + > > + s->prient_prog = s->default_prient_prog; > > + s->enable = 0; > > + > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, s->md_num); > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, 1); > > + s->regs.mdcfg[0] = s->k; > > +} > > + > > +static int iopmp_attrs_to_index(IOMMUMemoryRegion *iommu, MemTxAttrs attrs) > > +{ > > + return attrs.requester_id; > > +} > > + > > +static void iopmp_iommu_memory_region_class_init(ObjectClass *klass, void *data) > > +{ > > + IOMMUMemoryRegionClass *imrc = IOMMU_MEMORY_REGION_CLASS(klass); > > + > > + imrc->translate = iopmp_translate; > > + imrc->attrs_to_index = iopmp_attrs_to_index; > > +} > > + > > +static Property iopmp_property[] = { > > + DEFINE_PROP_BOOL("prient_prog", IopmpState, default_prient_prog, true), > > + DEFINE_PROP_UINT32("k", IopmpState, k, 6), > > + DEFINE_PROP_UINT32("prio_entry", IopmpState, prio_entry, 48), > > + DEFINE_PROP_UINT32("rrid_num", IopmpState, rrid_num, 16), > > + DEFINE_PROP_UINT32("md_num", IopmpState, md_num, 8), > > + DEFINE_PROP_UINT32("entry_offset", IopmpState, entry_offset, 0x4000), > > + DEFINE_PROP_UINT32("fabricated_v", IopmpState, fabricated_v, 0x0), > > + DEFINE_PROP_END_OF_LIST(), > > +}; > > + > > +static void iopmp_class_init(ObjectClass *klass, void *data) > > +{ > > + DeviceClass *dc = DEVICE_CLASS(klass); > > + device_class_set_props(dc, iopmp_property); > > + dc->realize = iopmp_realize; > > + dc->reset = iopmp_reset; > > +} > > + > > +static void iopmp_init(Object *obj) > > +{ > > + IopmpState *s = IOPMP(obj); > > + SysBusDevice *sbd = SYS_BUS_DEVICE(obj); > > + > > + sysbus_init_irq(sbd, &s->irq); > > +} > > + > > +static const TypeInfo iopmp_info = { > > + .name = TYPE_IOPMP, > > + .parent = TYPE_SYS_BUS_DEVICE, > > + .instance_size = sizeof(IopmpState), > > + .instance_init = iopmp_init, > > + .class_init = iopmp_class_init, > > +}; > > + > > +static const TypeInfo > > +iopmp_iommu_memory_region_info = { > > + .name = TYPE_IOPMP_IOMMU_MEMORY_REGION, > > + .parent = TYPE_IOMMU_MEMORY_REGION, > > + .class_init = iopmp_iommu_memory_region_class_init, > > +}; > > + > > +static void > > +iopmp_register_types(void) > > +{ > > + type_register_static(&iopmp_info); > > + type_register_static(&iopmp_iommu_memory_region_info); > > +} > > + > > +type_init(iopmp_register_types); > > diff --git a/hw/misc/trace-events b/hw/misc/trace-events > > index 1be0717c0c..c148166d2d 100644 > > --- a/hw/misc/trace-events > > +++ b/hw/misc/trace-events > > @@ -362,3 +362,6 @@ aspeed_sli_read(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx > > aspeed_sliio_write(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx64 " of size %u: 0x%" PRIx32 > > aspeed_sliio_read(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx64 " of size %u: 0x%" PRIx32 > > > > +# riscv_iopmp.c > > +iopmp_read(uint64_t addr, uint32_t val) "addr 0x%"PRIx64" val 0x%x" > > +iopmp_write(uint64_t addr, uint32_t val) "addr 0x%"PRIx64" val 0x%x" > > diff --git a/include/hw/misc/riscv_iopmp.h b/include/hw/misc/riscv_iopmp.h > > new file mode 100644 > > index 0000000000..b8fe479108 > > --- /dev/null > > +++ b/include/hw/misc/riscv_iopmp.h > > @@ -0,0 +1,168 @@ > > +/* > > + * QEMU RISC-V IOPMP (Input Output Physical Memory Protection) > > + * > > + * Copyright (c) 2023-2024 Andes Tech. Corp. > > + * > > + * SPDX-License-Identifier: GPL-2.0-or-later > > + * > > + * This program is free software; you can redistribute it and/or modify it > > + * under the terms and conditions of the GNU General Public License, > > + * version 2 or later, as published by the Free Software Foundation. > > + * > > + * This program is distributed in the hope it will be useful, but WITHOUT > > + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or > > + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for > > + * more details. > > + * > > + * You should have received a copy of the GNU General Public License along with > > + * this program. If not, see <http://www.gnu.org/licenses/>. > > + */ > > + > > +#ifndef RISCV_IOPMP_H > > +#define RISCV_IOPMP_H > > + > > +#include "hw/sysbus.h" > > +#include "qemu/typedefs.h" > > +#include "memory.h" > > +#include "exec/hwaddr.h" > > + > > +#define TYPE_IOPMP "iopmp" > > +#define IOPMP(obj) OBJECT_CHECK(IopmpState, (obj), TYPE_IOPMP) > > + > > +#define IOPMP_MAX_MD_NUM 63 > > +#define IOPMP_MAX_RRID_NUM 65535 > > +#define IOPMP_MAX_ENTRY_NUM 65535 > > + > > +#define VENDER_VIRT 0 > > +#define SPECVER_0_9_1 91 > > +#define IMPID_0_9_1 91 > > + > > +#define RRE_ERROR 0 > > +#define RRE_SUCCESS_VALUE 1 > > + > > +#define RWE_ERROR 0 > > +#define RWE_SUCCESS 1 > > + > > +#define RXE_ERROR 0 > > +#define RXE_SUCCESS_VALUE 1 > > + > > +#define ERR_REQINFO_TTYPE_READ 1 > > +#define ERR_REQINFO_TTYPE_WRITE 2 > > +#define ERR_REQINFO_TTYPE_FETCH 3 > > +#define ERR_REQINFO_ETYPE_NOERROR 0 > > +#define ERR_REQINFO_ETYPE_READ 1 > > +#define ERR_REQINFO_ETYPE_WRITE 2 > > +#define ERR_REQINFO_ETYPE_FETCH 3 > > +#define ERR_REQINFO_ETYPE_PARHIT 4 > > +#define ERR_REQINFO_ETYPE_NOHIT 5 > > +#define ERR_REQINFO_ETYPE_RRID 6 > > +#define ERR_REQINFO_ETYPE_USER 7 > > + > > +#define IOPMP_MODEL_FULL 0 > > +#define IOPMP_MODEL_RAPIDK 0x1 > > +#define IOPMP_MODEL_DYNAMICK 0x2 > > +#define IOPMP_MODEL_ISOLATION 0x3 > > +#define IOPMP_MODEL_COMPACTK 0x4 > > + > > +#define ENTRY_NO_HIT 0 > > +#define ENTRY_PAR_HIT 1 > > +#define ENTRY_HIT 2 > > Why not an enum? Thanks for your suggestion. It will be changed to enum in next revision. Thanks, Ethan Chen > > Alistair > > > + > > +/* The generic iopmp address space which downstream is system memory */ > > +extern AddressSpace iopmp_container_as; > > + > > +typedef enum { > > + IOPMP_AMATCH_OFF, /* Null (off) */ > > + IOPMP_AMATCH_TOR, /* Top of Range */ > > + IOPMP_AMATCH_NA4, /* Naturally aligned four-byte region */ > > + IOPMP_AMATCH_NAPOT /* Naturally aligned power-of-two region */ > > +} iopmp_am_t; > > + > > +typedef enum { > > + IOPMP_ACCESS_READ = 1, > > + IOPMP_ACCESS_WRITE = 2, > > + IOPMP_ACCESS_FETCH = 3 > > +} iopmp_access_type; > > + > > +typedef enum { > > + IOPMP_NONE = 0, > > + IOPMP_RO = 1, > > + IOPMP_WO = 2, > > + IOPMP_RW = 3, > > + IOPMP_XO = 4, > > + IOPMP_RX = 5, > > + IOPMP_WX = 6, > > + IOPMP_RWX = 7, > > +} iopmp_permission; > > + > > +typedef struct { > > + uint32_t addr_reg; > > + uint32_t addrh_reg; > > + uint32_t cfg_reg; > > +} iopmp_entry_t; > > + > > +typedef struct { > > + uint64_t sa; > > + uint64_t ea; > > +} iopmp_addr_t; > > + > > +typedef struct { > > + uint32_t *srcmd_en; > > + uint32_t *srcmd_enh; > > + uint32_t *mdcfg; > > + iopmp_entry_t *entry; > > + uint32_t mdlck; > > + uint32_t mdlckh; > > + uint32_t entrylck; > > + uint32_t mdcfglck; > > + uint32_t mdstall; > > + uint32_t mdstallh; > > + uint32_t rridscp; > > + uint32_t err_cfg; > > + uint64_t err_reqaddr; > > + uint32_t err_reqid; > > + uint32_t err_reqinfo; > > +} iopmp_regs; > > + > > + > > +/* To detect partially hit */ > > +typedef struct iopmp_transaction_state { > > + bool running; > > + bool supported; > > + hwaddr start_addr; > > + hwaddr end_addr; > > +} iopmp_transaction_state; > > + > > +typedef struct IopmpState { > > + SysBusDevice parent_obj; > > + iopmp_addr_t *entry_addr; > > + MemoryRegion mmio; > > + IOMMUMemoryRegion iommu; > > + IOMMUMemoryRegion *next_iommu; > > + iopmp_regs regs; > > + MemoryRegion *downstream; > > + MemoryRegion blocked_r, blocked_w, blocked_x, blocked_rw, blocked_rx, > > + blocked_wx, blocked_rwx; > > + MemoryRegion stall_io; > > + uint32_t model; > > + uint32_t k; > > + bool prient_prog; > > + bool default_prient_prog; > > + iopmp_transaction_state *transaction_state; > > + QemuMutex iopmp_transaction_mutex; > > + > > + AddressSpace downstream_as; > > + AddressSpace blocked_r_as, blocked_w_as, blocked_x_as, blocked_rw_as, > > + blocked_rx_as, blocked_wx_as, blocked_rwx_as; > > + qemu_irq irq; > > + bool enable; > > + > > + uint32_t prio_entry; > > + uint32_t rrid_num; > > + uint32_t md_num; > > + uint32_t entry_num; > > + uint32_t entry_offset; > > + uint32_t fabricated_v; > > +} IopmpState; > > + > > +#endif > > -- > > 2.34.1 > > > >
On Fri, Aug 9, 2024 at 7:42 PM Ethan Chen <ethan84@andestech.com> wrote: > > On Thu, Aug 08, 2024 at 01:56:35PM +1000, Alistair Francis wrote: > > [EXTERNAL MAIL] > > > > On Mon, Jul 15, 2024 at 7:58 PM Ethan Chen via <qemu-devel@nongnu.org> wrote: > > > > > > Support basic functions of IOPMP specification v0.9.1 rapid-k model. > > > The specification url: > > > https://github.com/riscv-non-isa/iopmp-spec/releases/tag/v0.9.1 > > > > > > The IOPMP checks whether memory access from a device or CPU is valid. > > > This implementation uses an IOMMU to modify the address space accessed > > > by the device. > > > > > > For device access with IOMMUAccessFlags specifying read or write > > > (IOMMU_RO or IOMMU_WO), the IOPMP checks the permission in > > > iopmp_translate. If the access is valid, the target address space is > > > downstream_as. If the access is blocked, it will be redirected to > > > blocked_rwx_as. > > > > > > For CPU access with IOMMUAccessFlags not specifying read or write > > > (IOMMU_NONE), the IOPMP translates the access to the corresponding > > > address space based on the permission. If the access has full permission > > > (rwx), the target address space is downstream_as. If the access has > > > limited permission, the target address space is blocked_ followed by > > > the lacked permissions. > > > > > > The operation of a blocked region can trigger an IOPMP interrupt, a bus > > > error, or it can respond with success and fabricated data, depending on > > > the value of the IOPMP ERR_CFG register. > > > > > > Signed-off-by: Ethan Chen <ethan84@andestech.com> > > > --- > > > hw/misc/Kconfig | 3 + > > > hw/misc/meson.build | 1 + > > > hw/misc/riscv_iopmp.c | 1154 +++++++++++++++++++++++++++++++++ > > > hw/misc/trace-events | 3 + > > > include/hw/misc/riscv_iopmp.h | 168 +++++ > > > 5 files changed, 1329 insertions(+) > > > create mode 100644 hw/misc/riscv_iopmp.c > > > create mode 100644 include/hw/misc/riscv_iopmp.h > > > > > > diff --git a/hw/misc/Kconfig b/hw/misc/Kconfig > > > index 1e08785b83..427f0c702e 100644 > > > --- a/hw/misc/Kconfig > > > +++ b/hw/misc/Kconfig > > > @@ -213,4 +213,7 @@ config IOSB > > > config XLNX_VERSAL_TRNG > > > bool > > > > > > +config RISCV_IOPMP > > > + bool > > > + > > > source macio/Kconfig > > > diff --git a/hw/misc/meson.build b/hw/misc/meson.build > > > index 2ca8717be2..d9006e1d81 100644 > > > --- a/hw/misc/meson.build > > > +++ b/hw/misc/meson.build > > > @@ -34,6 +34,7 @@ system_ss.add(when: 'CONFIG_SIFIVE_E_PRCI', if_true: files('sifive_e_prci.c')) > > > system_ss.add(when: 'CONFIG_SIFIVE_E_AON', if_true: files('sifive_e_aon.c')) > > > system_ss.add(when: 'CONFIG_SIFIVE_U_OTP', if_true: files('sifive_u_otp.c')) > > > system_ss.add(when: 'CONFIG_SIFIVE_U_PRCI', if_true: files('sifive_u_prci.c')) > > > +specific_ss.add(when: 'CONFIG_RISCV_IOPMP', if_true: files('riscv_iopmp.c')) > > > > > > subdir('macio') > > > > > > diff --git a/hw/misc/riscv_iopmp.c b/hw/misc/riscv_iopmp.c > > > new file mode 100644 > > > index 0000000000..db43e3c73f > > > --- /dev/null > > > +++ b/hw/misc/riscv_iopmp.c > > > @@ -0,0 +1,1154 @@ > > > +/* > > > + * QEMU RISC-V IOPMP (Input Output Physical Memory Protection) > > > + * > > > + * Copyright (c) 2023-2024 Andes Tech. Corp. > > > + * > > > + * SPDX-License-Identifier: GPL-2.0-or-later > > > + * > > > + * This program is free software; you can redistribute it and/or modify it > > > + * under the terms and conditions of the GNU General Public License, > > > + * version 2 or later, as published by the Free Software Foundation. > > > + * > > > + * This program is distributed in the hope it will be useful, but WITHOUT > > > + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or > > > + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for > > > + * more details. > > > + * > > > + * this program. If not, see <http://www.gnu.org/licenses/>. > > > + */ > > > + > > > +#include "qemu/osdep.h" > > > +#include "qemu/log.h" > > > +#include "qapi/error.h" > > > +#include "trace.h" > > > +#include "exec/exec-all.h" > > > +#include "exec/address-spaces.h" > > > +#include "hw/qdev-properties.h" > > > +#include "hw/sysbus.h" > > > +#include "hw/misc/riscv_iopmp.h" > > > +#include "memory.h" > > > +#include "hw/irq.h" > > > +#include "hw/registerfields.h" > > > +#include "trace.h" > > > + > > > +#define TYPE_IOPMP_IOMMU_MEMORY_REGION "iopmp-iommu-memory-region" > > > + > > > +REG32(VERSION, 0x00) > > > + FIELD(VERSION, VENDOR, 0, 24) > > > + FIELD(VERSION, SPECVER , 24, 8) > > > +REG32(IMP, 0x04) > > > + FIELD(IMP, IMPID, 0, 32) > > > +REG32(HWCFG0, 0x08) > > > + FIELD(HWCFG0, MODEL, 0, 4) > > > + FIELD(HWCFG0, TOR_EN, 4, 1) > > > + FIELD(HWCFG0, SPS_EN, 5, 1) > > > + FIELD(HWCFG0, USER_CFG_EN, 6, 1) > > > + FIELD(HWCFG0, PRIENT_PROG, 7, 1) > > > + FIELD(HWCFG0, RRID_TRANSL_EN, 8, 1) > > > + FIELD(HWCFG0, RRID_TRANSL_PROG, 9, 1) > > > + FIELD(HWCFG0, CHK_X, 10, 1) > > > + FIELD(HWCFG0, NO_X, 11, 1) > > > + FIELD(HWCFG0, NO_W, 12, 1) > > > + FIELD(HWCFG0, STALL_EN, 13, 1) > > > + FIELD(HWCFG0, PEIS, 14, 1) > > > + FIELD(HWCFG0, PEES, 15, 1) > > > + FIELD(HWCFG0, MFR_EN, 16, 1) > > > + FIELD(HWCFG0, MD_NUM, 24, 7) > > > + FIELD(HWCFG0, ENABLE, 31, 1) > > > +REG32(HWCFG1, 0x0C) > > > + FIELD(HWCFG1, RRID_NUM, 0, 16) > > > + FIELD(HWCFG1, ENTRY_NUM, 16, 16) > > > +REG32(HWCFG2, 0x10) > > > + FIELD(HWCFG2, PRIO_ENTRY, 0, 16) > > > + FIELD(HWCFG2, RRID_TRANSL, 16, 16) > > > +REG32(ENTRYOFFSET, 0x14) > > > + FIELD(ENTRYOFFSET, OFFSET, 0, 32) > > > +REG32(MDSTALL, 0x30) > > > + FIELD(MDSTALL, EXEMPT, 0, 1) > > > + FIELD(MDSTALL, MD, 1, 31) > > > +REG32(MDSTALLH, 0x34) > > > + FIELD(MDSTALLH, MD, 0, 32) > > > +REG32(RRIDSCP, 0x38) > > > + FIELD(RRIDSCP, RRID, 0, 16) > > > + FIELD(RRIDSCP, OP, 30, 2) > > > +REG32(MDLCK, 0x40) > > > + FIELD(MDLCK, L, 0, 1) > > > + FIELD(MDLCK, MD, 1, 31) > > > +REG32(MDLCKH, 0x44) > > > + FIELD(MDLCKH, MDH, 0, 32) > > > +REG32(MDCFGLCK, 0x48) > > > + FIELD(MDCFGLCK, L, 0, 1) > > > + FIELD(MDCFGLCK, F, 1, 7) > > > +REG32(ENTRYLCK, 0x4C) > > > + FIELD(ENTRYLCK, L, 0, 1) > > > + FIELD(ENTRYLCK, F, 1, 16) > > > +REG32(ERR_CFG, 0x60) > > > + FIELD(ERR_CFG, L, 0, 1) > > > + FIELD(ERR_CFG, IE, 1, 1) > > > + FIELD(ERR_CFG, IRE, 2, 1) > > > + FIELD(ERR_CFG, IWE, 3, 1) > > > + FIELD(ERR_CFG, IXE, 4, 1) > > > + FIELD(ERR_CFG, RRE, 5, 1) > > > + FIELD(ERR_CFG, RWE, 6, 1) > > > + FIELD(ERR_CFG, RXE, 7, 1) > > > +REG32(ERR_REQINFO, 0x64) > > > + FIELD(ERR_REQINFO, V, 0, 1) > > > + FIELD(ERR_REQINFO, TTYPE, 1, 2) > > > + FIELD(ERR_REQINFO, ETYPE, 4, 3) > > > + FIELD(ERR_REQINFO, SVC, 7, 1) > > > +REG32(ERR_REQADDR, 0x68) > > > + FIELD(ERR_REQADDR, ADDR, 0, 32) > > > +REG32(ERR_REQADDRH, 0x6C) > > > + FIELD(ERR_REQADDRH, ADDRH, 0, 32) > > > +REG32(ERR_REQID, 0x70) > > > + FIELD(ERR_REQID, RRID, 0, 16) > > > + FIELD(ERR_REQID, EID, 16, 16) > > > +REG32(ERR_MFR, 0x74) > > > + FIELD(ERR_MFR, SVW, 0, 16) > > > + FIELD(ERR_MFR, SVI, 16, 12) > > > + FIELD(ERR_MFR, SVS, 31, 1) > > > +REG32(MDCFG0, 0x800) > > > + FIELD(MDCFG0, T, 0, 16) > > > +REG32(SRCMD_EN0, 0x1000) > > > + FIELD(SRCMD_EN0, L, 0, 1) > > > + FIELD(SRCMD_EN0, MD, 1, 31) > > > +REG32(SRCMD_ENH0, 0x1004) > > > + FIELD(SRCMD_ENH0, MDH, 0, 32) > > > +REG32(SRCMD_R0, 0x1008) > > > + FIELD(SRCMD_R0, MD, 1, 31) > > > +REG32(SRCMD_RH0, 0x100C) > > > + FIELD(SRCMD_RH0, MDH, 0, 32) > > > +REG32(SRCMD_W0, 0x1010) > > > + FIELD(SRCMD_W0, MD, 1, 31) > > > +REG32(SRCMD_WH0, 0x1014) > > > + FIELD(SRCMD_WH0, MDH, 0, 32) > > > + > > > +FIELD(ENTRY_ADDR, ADDR, 0, 32) > > > +FIELD(ENTRY_ADDRH, ADDRH, 0, 32) > > > + > > > +FIELD(ENTRY_CFG, R, 0, 1) > > > +FIELD(ENTRY_CFG, W, 1, 1) > > > +FIELD(ENTRY_CFG, X, 2, 1) > > > +FIELD(ENTRY_CFG, A, 3, 2) > > > +FIELD(ENTRY_CFG, SIRE, 5, 1) > > > +FIELD(ENTRY_CFG, SIWE, 6, 1) > > > +FIELD(ENTRY_CFG, SIXE, 7, 1) > > > +FIELD(ENTRY_CFG, SERE, 8, 1) > > > +FIELD(ENTRY_CFG, SEWE, 9, 1) > > > +FIELD(ENTRY_CFG, SEXE, 10, 1) > > > + > > > +FIELD(ENTRY_USER_CFG, IM, 0, 32) > > > + > > > +/* Offsets to SRCMD_EN(i) */ > > > +#define SRCMD_EN_OFFSET 0x0 > > > +#define SRCMD_ENH_OFFSET 0x4 > > > +#define SRCMD_R_OFFSET 0x8 > > > +#define SRCMD_RH_OFFSET 0xC > > > +#define SRCMD_W_OFFSET 0x10 > > > +#define SRCMD_WH_OFFSET 0x14 > > > + > > > +/* Offsets to ENTRY_ADDR(i) */ > > > +#define ENTRY_ADDR_OFFSET 0x0 > > > +#define ENTRY_ADDRH_OFFSET 0x4 > > > +#define ENTRY_CFG_OFFSET 0x8 > > > +#define ENTRY_USER_CFG_OFFSET 0xC > > > + > > > +/* Memmap for parallel IOPMPs */ > > > +typedef struct iopmp_protection_memmap { > > > + MemMapEntry entry; > > > + IopmpState *iopmp_s; > > > + QLIST_ENTRY(iopmp_protection_memmap) list; > > > +} iopmp_protection_memmap; > > > +QLIST_HEAD(, iopmp_protection_memmap) > > > + iopmp_protection_memmaps = QLIST_HEAD_INITIALIZER(iopmp_protection_memmaps); > > > + > > > +static void iopmp_iommu_notify(IopmpState *s) > > > +{ > > > + IOMMUTLBEvent event = { > > > + .entry = { > > > + .iova = 0, > > > + .translated_addr = 0, > > > + .addr_mask = -1ULL, > > > + .perm = IOMMU_NONE, > > > + }, > > > + .type = IOMMU_NOTIFIER_UNMAP, > > > + }; > > > + > > > + for (int i = 0; i < s->rrid_num; i++) { > > > + memory_region_notify_iommu(&s->iommu, i, event); > > > + } > > > +} > > > + > > > +static void iopmp_decode_napot(uint64_t a, uint64_t *sa, > > > + uint64_t *ea) > > > +{ > > > + /* > > > + * aaaa...aaa0 8-byte NAPOT range > > > + * aaaa...aa01 16-byte NAPOT range > > > + * aaaa...a011 32-byte NAPOT range > > > + * ... > > > + * aa01...1111 2^XLEN-byte NAPOT range > > > + * a011...1111 2^(XLEN+1)-byte NAPOT range > > > + * 0111...1111 2^(XLEN+2)-byte NAPOT range > > > + * 1111...1111 Reserved > > > + */ > > > + > > > + a = (a << 2) | 0x3; > > > + *sa = a & (a + 1); > > > + *ea = a | (a + 1); > > > +} > > > + > > > +static void iopmp_update_rule(IopmpState *s, uint32_t entry_index) > > > +{ > > > + uint8_t this_cfg = s->regs.entry[entry_index].cfg_reg; > > > + uint64_t this_addr = s->regs.entry[entry_index].addr_reg | > > > + ((uint64_t)s->regs.entry[entry_index].addrh_reg << 32); > > > + uint64_t prev_addr = 0u; > > > + uint64_t sa = 0u; > > > + uint64_t ea = 0u; > > > + > > > + if (entry_index >= 1u) { > > > + prev_addr = s->regs.entry[entry_index - 1].addr_reg | > > > + ((uint64_t)s->regs.entry[entry_index - 1].addrh_reg << 32); > > > + } > > > + > > > + switch (FIELD_EX32(this_cfg, ENTRY_CFG, A)) { > > > + case IOPMP_AMATCH_OFF: > > > + sa = 0u; > > > + ea = -1; > > > + break; > > > + > > > + case IOPMP_AMATCH_TOR: > > > + sa = (prev_addr) << 2; /* shift up from [xx:0] to [xx+2:2] */ > > > + ea = ((this_addr) << 2) - 1u; > > > + if (sa > ea) { > > > + sa = ea = 0u; > > > + } > > > + break; > > > + > > > + case IOPMP_AMATCH_NA4: > > > + sa = this_addr << 2; /* shift up from [xx:0] to [xx+2:2] */ > > > + ea = (sa + 4u) - 1u; > > > + break; > > > + > > > + case IOPMP_AMATCH_NAPOT: > > > + iopmp_decode_napot(this_addr, &sa, &ea); > > > + break; > > > + > > > + default: > > > + sa = 0u; > > > + ea = 0u; > > > + break; > > > + } > > > + > > > + s->entry_addr[entry_index].sa = sa; > > > + s->entry_addr[entry_index].ea = ea; > > > + iopmp_iommu_notify(s); > > > +} > > > + > > > +static uint64_t iopmp_read(void *opaque, hwaddr addr, unsigned size) > > > +{ > > > + IopmpState *s = IOPMP(opaque); > > > + uint32_t rz = 0; > > > + uint32_t offset, idx; > > > + > > > + switch (addr) { > > > + case A_VERSION: > > > + rz = VENDER_VIRT << R_VERSION_VENDOR_SHIFT | > > > + SPECVER_0_9_1 << R_VERSION_SPECVER_SHIFT; > > > > It would be better to use the FIELD_DP32() macro instead of the manual shifts > > It will be refined in next revision. > > > > > > + break; > > > + case A_IMP: > > > + rz = IMPID_0_9_1; > > > + break; > > > + case A_HWCFG0: > > > + rz = s->model << R_HWCFG0_MODEL_SHIFT | > > > + 1 << R_HWCFG0_TOR_EN_SHIFT | > > > + 0 << R_HWCFG0_SPS_EN_SHIFT | > > > + 0 << R_HWCFG0_USER_CFG_EN_SHIFT | > > > + s->prient_prog << R_HWCFG0_PRIENT_PROG_SHIFT | > > > + 0 << R_HWCFG0_RRID_TRANSL_EN_SHIFT | > > > + 0 << R_HWCFG0_RRID_TRANSL_PROG_SHIFT | > > > + 1 << R_HWCFG0_CHK_X_SHIFT | > > > + 0 << R_HWCFG0_NO_X_SHIFT | > > > + 0 << R_HWCFG0_NO_W_SHIFT | > > > + 0 << R_HWCFG0_STALL_EN_SHIFT | > > > + 0 << R_HWCFG0_PEIS_SHIFT | > > > + 0 << R_HWCFG0_PEES_SHIFT | > > > + 0 << R_HWCFG0_MFR_EN_SHIFT | > > > + s->md_num << R_HWCFG0_MD_NUM_SHIFT | > > > + s->enable << R_HWCFG0_ENABLE_SHIFT ; > > > + break; > > > + case A_HWCFG1: > > > + rz = s->rrid_num << R_HWCFG1_RRID_NUM_SHIFT | > > > + s->entry_num << R_HWCFG1_ENTRY_NUM_SHIFT; > > > + break; > > > + case A_HWCFG2: > > > + rz = s->prio_entry << R_HWCFG2_PRIO_ENTRY_SHIFT; > > > + break; > > > + case A_ENTRYOFFSET: > > > + rz = s->entry_offset; > > > + break; > > > + case A_ERR_CFG: > > > + rz = s->regs.err_cfg; > > > + break; > > > + case A_MDLCK: > > > + rz = s->regs.mdlck; > > > + break; > > > + case A_MDLCKH: > > > + rz = s->regs.mdlckh; > > > + break; > > > + case A_MDCFGLCK: > > > + rz = s->regs.mdcfglck; > > > + break; > > > + case A_ENTRYLCK: > > > + rz = s->regs.entrylck; > > > + break; > > > + case A_ERR_REQADDR: > > > + rz = s->regs.err_reqaddr & UINT32_MAX; > > > + break; > > > + case A_ERR_REQADDRH: > > > + rz = s->regs.err_reqaddr >> 32; > > > + break; > > > + case A_ERR_REQID: > > > + rz = s->regs.err_reqid; > > > + break; > > > + case A_ERR_REQINFO: > > > + rz = s->regs.err_reqinfo; > > > + break; > > > + > > > + default: > > > + if (addr >= A_MDCFG0 && > > > + addr < A_MDCFG0 + 4 * (s->md_num - 1)) { > > > + offset = addr - A_MDCFG0; > > > + idx = offset >> 2; > > > + if (idx == 0 && offset == 0) { > > > + rz = s->regs.mdcfg[idx]; > > > + } else { > > > + /* Only MDCFG0 is implemented in rapid-k model */ > > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > > + __func__, (int)addr); > > > + } > > > + } else if (addr >= A_SRCMD_EN0 && > > > + addr < A_SRCMD_WH0 + 32 * (s->rrid_num - 1)) { > > > + offset = addr - A_SRCMD_EN0; > > > + idx = offset >> 5; > > > + offset &= 0x1f; > > > + > > > + switch (offset) { > > > + case SRCMD_EN_OFFSET: > > > + rz = s->regs.srcmd_en[idx]; > > > + break; > > > + case SRCMD_ENH_OFFSET: > > > + rz = s->regs.srcmd_enh[idx]; > > > + break; > > > + default: > > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > > + __func__, (int)addr); > > > + break; > > > + } > > > + } else if (addr >= s->entry_offset && > > > + addr < s->entry_offset + ENTRY_USER_CFG_OFFSET + > > > + 16 * (s->entry_num - 1)) { > > > + offset = addr - s->entry_offset; > > > + idx = offset >> 4; > > > + offset &= 0xf; > > > + > > > + switch (offset) { > > > + case ENTRY_ADDR_OFFSET: > > > + rz = s->regs.entry[idx].addr_reg; > > > + break; > > > + case ENTRY_ADDRH_OFFSET: > > > + rz = s->regs.entry[idx].addrh_reg; > > > + break; > > > + case ENTRY_CFG_OFFSET: > > > + rz = s->regs.entry[idx].cfg_reg; > > > + break; > > > + case ENTRY_USER_CFG_OFFSET: > > > + /* Does not support user customized permission */ > > > + rz = 0; > > > + break; > > > + default: > > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > > + __func__, (int)addr); > > > + break; > > > + } > > > + } else { > > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > > + __func__, (int)addr); > > > + } > > > + break; > > > + } > > > + trace_iopmp_read(addr, rz); > > > + return rz; > > > +} > > > + > > > +static void > > > +iopmp_write(void *opaque, hwaddr addr, uint64_t value, unsigned size) > > > +{ > > > + IopmpState *s = IOPMP(opaque); > > > + uint32_t offset, idx; > > > + uint32_t value32 = value; > > > + > > > + trace_iopmp_write(addr, value32); > > > + > > > + switch (addr) { > > > + case A_VERSION: /* RO */ > > > + break; > > > + case A_IMP: /* RO */ > > > + break; > > > + case A_HWCFG0: > > > + if (FIELD_EX32(value32, HWCFG0, PRIENT_PROG)) { > > > + /* W1C */ > > > + s->prient_prog = 0; > > > + } > > > + if (FIELD_EX32(value32, HWCFG0, ENABLE)) { > > > + /* W1S */ > > > + s->enable = 1; > > > + iopmp_iommu_notify(s); > > > + } > > > + break; > > > + case A_HWCFG1: /* RO */ > > > + break; > > > + case A_HWCFG2: > > > + if (s->prient_prog) { > > > + s->prio_entry = FIELD_EX32(value32, HWCFG2, PRIO_ENTRY); > > > + } > > > + break; > > > + case A_ERR_CFG: > > > + if (!FIELD_EX32(s->regs.err_cfg, ERR_CFG, L)) { > > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, L, > > > + FIELD_EX32(value32, ERR_CFG, L)); > > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IE, > > > + FIELD_EX32(value32, ERR_CFG, IE)); > > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IRE, > > > + FIELD_EX32(value32, ERR_CFG, IRE)); > > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RRE, > > > + FIELD_EX32(value32, ERR_CFG, RRE)); > > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IWE, > > > + FIELD_EX32(value32, ERR_CFG, IWE)); > > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RWE, > > > + FIELD_EX32(value32, ERR_CFG, RWE)); > > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IXE, > > > + FIELD_EX32(value32, ERR_CFG, IXE)); > > > + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RXE, > > > + FIELD_EX32(value32, ERR_CFG, RXE)); > > > + } > > > + break; > > > + case A_MDLCK: > > > + if (!FIELD_EX32(s->regs.mdlck, MDLCK, L)) { > > > + s->regs.mdlck = value32; > > > + } > > > + break; > > > + case A_MDLCKH: > > > + if (!FIELD_EX32(s->regs.mdlck, MDLCK, L)) { > > > + s->regs.mdlckh = value32; > > > + } > > > + break; > > > + case A_MDCFGLCK: > > > + if (!FIELD_EX32(s->regs.mdcfglck, MDCFGLCK, L)) { > > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, > > > + FIELD_EX32(value32, MDCFGLCK, F)); > > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, > > > + FIELD_EX32(value32, MDCFGLCK, L)); > > > + } > > > + break; > > > + case A_ENTRYLCK: > > > + if (!(FIELD_EX32(s->regs.entrylck, ENTRYLCK, L))) { > > > + s->regs.entrylck = FIELD_DP32(s->regs.entrylck, ENTRYLCK, F, > > > + FIELD_EX32(value32, ENTRYLCK, F)); > > > + s->regs.entrylck = FIELD_DP32(s->regs.entrylck, ENTRYLCK, L, > > > + FIELD_EX32(value32, ENTRYLCK, L)); > > > + } > > > + case A_ERR_REQADDR: /* RO */ > > > + break; > > > + case A_ERR_REQADDRH: /* RO */ > > > + break; > > > + case A_ERR_REQID: /* RO */ > > > + break; > > > + case A_ERR_REQINFO: > > > + if (FIELD_EX32(value32, ERR_REQINFO, V)) { > > > + s->regs.err_reqinfo = FIELD_DP32(s->regs.err_reqinfo, > > > + ERR_REQINFO, V, 0); > > > + qemu_set_irq(s->irq, 0); > > > + } > > > + break; > > > + > > > + default: > > > + if (addr >= A_MDCFG0 && > > > + addr < A_MDCFG0 + 4 * (s->md_num - 1)) { > > > + offset = addr - A_MDCFG0; > > > + idx = offset >> 2; > > > + /* RO in rapid-k model */ > > > + if (idx > 0) { > > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > > + __func__, (int)addr); > > > + } > > > + } else if (addr >= A_SRCMD_EN0 && > > > + addr < A_SRCMD_WH0 + 32 * (s->rrid_num - 1)) { > > > + offset = addr - A_SRCMD_EN0; > > > + idx = offset >> 5; > > > + offset &= 0x1f; > > > + > > > + if (offset % 4) { > > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > > + __func__, (int)addr); > > > + } else if (FIELD_EX32(s->regs.srcmd_en[idx], SRCMD_EN0, L) > > > + == 0) { > > > + switch (offset) { > > > + case SRCMD_EN_OFFSET: > > > + s->regs.srcmd_en[idx] = > > > + FIELD_DP32(s->regs.srcmd_en[idx], SRCMD_EN0, L, > > > + FIELD_EX32(value32, SRCMD_EN0, L)); > > > + > > > + /* MD field is protected by mdlck */ > > > + value32 = (value32 & ~s->regs.mdlck) | > > > + (s->regs.srcmd_en[idx] & s->regs.mdlck); > > > + s->regs.srcmd_en[idx] = > > > + FIELD_DP32(s->regs.srcmd_en[idx], SRCMD_EN0, MD, > > > + FIELD_EX32(value32, SRCMD_EN0, MD)); > > > + break; > > > + case SRCMD_ENH_OFFSET: > > > + value32 = (value32 & ~s->regs.mdlckh) | > > > + (s->regs.srcmd_enh[idx] & s->regs.mdlckh); > > > + s->regs.srcmd_enh[idx] = > > > + FIELD_DP32(s->regs.srcmd_enh[idx], SRCMD_ENH0, MDH, > > > + value32); > > > + break; > > > + default: > > > + break; > > > + } > > > + } > > > + } else if (addr >= s->entry_offset && > > > + addr < s->entry_offset + ENTRY_USER_CFG_OFFSET > > > + + 16 * (s->entry_num - 1)) { > > > + offset = addr - s->entry_offset; > > > + idx = offset >> 4; > > > + offset &= 0xf; > > > + > > > + /* index < ENTRYLCK_F is protected */ > > > + if (idx >= FIELD_EX32(s->regs.entrylck, ENTRYLCK, F)) { > > > + switch (offset) { > > > + case ENTRY_ADDR_OFFSET: > > > + s->regs.entry[idx].addr_reg = value32; > > > + break; > > > + case ENTRY_ADDRH_OFFSET: > > > + s->regs.entry[idx].addrh_reg = value32; > > > + break; > > > + case ENTRY_CFG_OFFSET: > > > + s->regs.entry[idx].cfg_reg = value32; > > > + break; > > > + case ENTRY_USER_CFG_OFFSET: > > > + /* Does not support user customized permission */ > > > + break; > > > + default: > > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", > > > + __func__, (int)addr); > > > + break; > > > + } > > > + iopmp_update_rule(s, idx); > > > + if (idx + 1 < s->entry_num && > > > + FIELD_EX32(s->regs.entry[idx + 1].cfg_reg, ENTRY_CFG, A) == > > > + IOPMP_AMATCH_TOR) { > > > + iopmp_update_rule(s, idx + 1); > > > + } > > > + } > > > + } else { > > > + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", __func__, > > > + (int)addr); > > > + } > > > + } > > > +} > > > + > > > +/* Match entry in memory domain */ > > > +static int match_entry_md(IopmpState *s, int md_idx, hwaddr start_addr, > > > + hwaddr end_addr, int *entry_idx, > > > + int *prior_entry_in_tlb) > > > +{ > > > + int entry_idx_s, entry_idx_e; > > > + int result = ENTRY_NO_HIT; > > > + int i = 0; > > > + hwaddr tlb_sa = start_addr & ~(TARGET_PAGE_SIZE - 1); > > > + hwaddr tlb_ea = tlb_sa + TARGET_PAGE_SIZE - 1; > > > + > > > + entry_idx_s = md_idx * s->regs.mdcfg[0]; > > > + entry_idx_e = (md_idx + 1) * s->regs.mdcfg[0]; > > > + > > > + if (entry_idx_s >= s->entry_num) { > > > + return result; > > > + } > > > + if (entry_idx_e > s->entry_num) { > > > + entry_idx_e = s->entry_num; > > > + } > > > + i = entry_idx_s; > > > + for (i = entry_idx_s; i < entry_idx_e; i++) { > > > + if (FIELD_EX32(s->regs.entry[i].cfg_reg, ENTRY_CFG, A) == > > > + IOPMP_AMATCH_OFF) { > > > + continue; > > > + } > > > + if (start_addr >= s->entry_addr[i].sa && > > > + start_addr <= s->entry_addr[i].ea) { > > > + /* Check end address */ > > > + if (end_addr >= s->entry_addr[i].sa && > > > + end_addr <= s->entry_addr[i].ea) { > > > + *entry_idx = i; > > > + return ENTRY_HIT; > > > + } else if (i >= s->prio_entry) { > > > + /* Continue for non-prio_entry */ > > > + continue; > > > + } else { > > > + *entry_idx = i; > > > + return ENTRY_PAR_HIT; > > > + } > > > + } else if (end_addr >= s->entry_addr[i].sa && > > > + end_addr <= s->entry_addr[i].ea) { > > > + /* Only end address matches the entry */ > > > + if (i >= s->prio_entry) { > > > + continue; > > > + } else { > > > + *entry_idx = i; > > > + return ENTRY_PAR_HIT; > > > + } > > > + } else if (start_addr < s->entry_addr[i].sa && > > > + end_addr > s->entry_addr[i].ea) { > > > + if (i >= s->prio_entry) { > > > + continue; > > > + } else { > > > + *entry_idx = i; > > > + return ENTRY_PAR_HIT; > > > + } > > > + } > > > + if (prior_entry_in_tlb != NULL) { > > > + if ((s->entry_addr[i].sa >= tlb_sa && > > > + s->entry_addr[i].sa <= tlb_ea) || > > > + (s->entry_addr[i].ea >= tlb_sa && > > > + s->entry_addr[i].ea <= tlb_ea)) { > > > + /* > > > + * TLB should not use the cached result when the tlb contains > > > + * higher priority entry > > > + */ > > > + *prior_entry_in_tlb = 1; > > > + } > > > + } > > > + } > > > + return result; > > > +} > > > + > > > +static int match_entry(IopmpState *s, int rrid, hwaddr start_addr, > > > + hwaddr end_addr, int *match_md_idx, > > > + int *match_entry_idx, int *prior_entry_in_tlb) > > > +{ > > > + int cur_result = ENTRY_NO_HIT; > > > + int result = ENTRY_NO_HIT; > > > + /* Remove lock bit */ > > > + uint64_t srcmd_en = ((uint64_t)s->regs.srcmd_en[rrid] | > > > + ((uint64_t)s->regs.srcmd_enh[rrid] << 32)) >> 1; > > > + > > > + for (int md_idx = 0; md_idx < s->md_num; md_idx++) { > > > + if (srcmd_en & (1ULL << md_idx)) { > > > + cur_result = match_entry_md(s, md_idx, start_addr, end_addr, > > > + match_entry_idx, prior_entry_in_tlb); > > > + if (cur_result == ENTRY_HIT || cur_result == ENTRY_PAR_HIT) { > > > + *match_md_idx = md_idx; > > > + return cur_result; > > > + } > > > + } > > > + } > > > + return result; > > > +} > > > + > > > +static void iopmp_error_reaction(IopmpState *s, uint32_t id, hwaddr start, > > > + uint32_t info) > > > +{ > > > + if (!FIELD_EX32(s->regs.err_reqinfo, ERR_REQINFO, V)) { > > > + s->regs.err_reqinfo = info; > > > + s->regs.err_reqinfo = FIELD_DP32(s->regs.err_reqinfo, ERR_REQINFO, V, > > > + 1); > > > + s->regs.err_reqid = id; > > > + /* addr[LEN+2:2] */ > > > + s->regs.err_reqaddr = start >> 2; > > > + > > > + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_READ && > > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && > > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IRE)) { > > > + qemu_set_irq(s->irq, 1); > > > + } > > > + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_WRITE && > > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && > > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IWE)) { > > > + qemu_set_irq(s->irq, 1); > > > + } > > > + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_FETCH && > > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && > > > + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IXE)) { > > > + qemu_set_irq(s->irq, 1); > > > + } > > > + } > > > +} > > > + > > > +static IOMMUTLBEntry iopmp_translate(IOMMUMemoryRegion *iommu, hwaddr addr, > > > + IOMMUAccessFlags flags, int iommu_idx) > > > +{ > > > + int rrid = iommu_idx; > > > + IopmpState *s = IOPMP(container_of(iommu, IopmpState, iommu)); > > > + hwaddr start_addr, end_addr; > > > + int entry_idx = -1; > > > + int md_idx = -1; > > > + int result; > > > + uint32_t error_info = 0; > > > + uint32_t error_id = 0; > > > + int prior_entry_in_tlb = 0; > > > + iopmp_permission iopmp_perm; > > > + IOMMUTLBEntry entry = { > > > + .target_as = &s->downstream_as, > > > + .iova = addr, > > > + .translated_addr = addr, > > > + .addr_mask = 0, > > > + .perm = IOMMU_NONE, > > > + }; > > > + > > > + if (!s->enable) { > > > + /* Bypass IOPMP */ > > > + entry.addr_mask = -1ULL, > > > + entry.perm = IOMMU_RW; > > > + return entry; > > > + } > > > + > > > + /* unknown RRID */ > > > + if (rrid >= s->rrid_num) { > > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > > + ERR_REQINFO_ETYPE_RRID); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > > > + iopmp_error_reaction(s, error_id, addr, error_info); > > > + entry.target_as = &s->blocked_rwx_as; > > > + entry.perm = IOMMU_RW; > > > + return entry; > > > + } > > > + > > > + if (s->transaction_state[rrid].supported == true) { > > > + start_addr = s->transaction_state[rrid].start_addr; > > > + end_addr = s->transaction_state[rrid].end_addr; > > > + } else { > > > + /* No transaction information, use the same address */ > > > + start_addr = addr; > > > + end_addr = addr; > > > + } > > > + > > > + result = match_entry(s, rrid, start_addr, end_addr, &md_idx, &entry_idx, > > > + &prior_entry_in_tlb); > > > + if (result == ENTRY_HIT) { > > > + entry.addr_mask = s->entry_addr[entry_idx].ea - > > > + s->entry_addr[entry_idx].sa; > > > + if (prior_entry_in_tlb) { > > > + /* Make TLB repeat iommu translation on every access */ > > > > I don't follow this, if we have a prior entry in the TLB cache we > > don't cache the accesses? > > For the cached TLB result to be used, the highest-priority entry in the TLB must > occupy the entire TLB page. If a lower-priority entry fills the entire TLB page, > it is still necessary to check which entry the transaction hits on each access > to the TLB page. Oh! When you say "prior" you mean priority, not prior. That is a little confusing. Maybe just write `priority_entry_in_tlb` to be clear. Also I think it's worth including you entire response in the code comment Alistair > > > > > > + entry.addr_mask = 0; > > > + } > > > + iopmp_perm = s->regs.entry[entry_idx].cfg_reg & IOPMP_RWX; > > > + if (flags) { > > > + if ((iopmp_perm & flags) == 0) { > > > + /* Permission denied */ > > > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > > + ERR_REQINFO_ETYPE_READ + flags - 1); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > > + entry.target_as = &s->blocked_rwx_as; > > > + entry.perm = IOMMU_RW; > > > + } else { > > > + entry.target_as = &s->downstream_as; > > > + entry.perm = iopmp_perm; > > > + } > > > + } else { > > > + /* CPU access with IOMMU_NONE flag */ > > > + if (iopmp_perm & IOPMP_XO) { > > > + if ((iopmp_perm & IOPMP_RW) == IOPMP_RW) { > > > + entry.target_as = &s->downstream_as; > > > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_RO) { > > > + entry.target_as = &s->blocked_w_as; > > > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_WO) { > > > + entry.target_as = &s->blocked_r_as; > > > + } else { > > > + entry.target_as = &s->blocked_rw_as; > > > + } > > > + } else { > > > + if ((iopmp_perm & IOPMP_RW) == IOMMU_RW) { > > > + entry.target_as = &s->blocked_x_as; > > > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_RO) { > > > + entry.target_as = &s->blocked_wx_as; > > > + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_WO) { > > > + entry.target_as = &s->blocked_rx_as; > > > + } else { > > > + entry.target_as = &s->blocked_rwx_as; > > > + } > > > + } > > > + entry.perm = IOMMU_RW; > > > + } > > > + } else { > > > + if (flags) { > > > + if (result == ENTRY_PAR_HIT) { > > > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > > + ERR_REQINFO_ETYPE_PARHIT); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > > + } else { > > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > > + ERR_REQINFO_ETYPE_NOHIT); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); > > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > > + } > > > + } > > > + /* CPU access with IOMMU_NONE flag no_hit or par_hit */ > > > + entry.target_as = &s->blocked_rwx_as; > > > + entry.perm = IOMMU_RW; > > > + } > > > + return entry; > > > +} > > > + > > > +static const MemoryRegionOps iopmp_ops = { > > > + .read = iopmp_read, > > > + .write = iopmp_write, > > > + .endianness = DEVICE_NATIVE_ENDIAN, > > > + .valid = {.min_access_size = 4, .max_access_size = 4} > > > +}; > > > + > > > +static MemTxResult iopmp_permssion_write(void *opaque, hwaddr addr, > > > + uint64_t value, unsigned size, > > > + MemTxAttrs attrs) > > > +{ > > > + IopmpState *s = IOPMP(opaque); > > > + return address_space_write(&s->downstream_as, addr, attrs, &value, size); > > > +} > > > + > > > +static MemTxResult iopmp_permssion_read(void *opaque, hwaddr addr, > > > + uint64_t *pdata, unsigned size, > > > + MemTxAttrs attrs) > > > +{ > > > + IopmpState *s = IOPMP(opaque); > > > + return address_space_read(&s->downstream_as, addr, attrs, pdata, size); > > > +} > > > + > > > +static MemTxResult iopmp_handle_block(void *opaque, hwaddr addr, > > > + uint64_t *data, unsigned size, > > > + MemTxAttrs attrs, > > > + iopmp_access_type access_type) { > > > + IopmpState *s = IOPMP(opaque); > > > + int md_idx, entry_idx; > > > + uint32_t error_info = 0; > > > + uint32_t error_id = 0; > > > + int rrid = attrs.requester_id; > > > + int result; > > > + hwaddr start_addr, end_addr; > > > + start_addr = addr; > > > + end_addr = addr; > > > + result = match_entry(s, rrid, start_addr, end_addr, &md_idx, &entry_idx, > > > + NULL); > > > + > > > + if (result == ENTRY_HIT) { > > > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > > + access_type); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, access_type); > > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > > + } else if (result == ENTRY_PAR_HIT) { > > > + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); > > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > > + ERR_REQINFO_ETYPE_PARHIT); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, > > > + access_type); > > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > > + } else { > > > + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, > > > + ERR_REQINFO_ETYPE_NOHIT); > > > + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, access_type); > > > + iopmp_error_reaction(s, error_id, start_addr, error_info); > > > + } > > > + > > > + if (access_type == IOPMP_ACCESS_READ) { > > > + > > > + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RRE)) { > > > + case RRE_ERROR: > > > + return MEMTX_ERROR; > > > + break; > > > + case RRE_SUCCESS_VALUE: > > > + *data = s->fabricated_v; > > > + return MEMTX_OK; > > > + break; > > > + default: > > > + break; > > > + } > > > + return MEMTX_OK; > > > + } else if (access_type == IOPMP_ACCESS_WRITE) { > > > + > > > + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RWE)) { > > > + case RWE_ERROR: > > > + return MEMTX_ERROR; > > > + break; > > > + case RWE_SUCCESS: > > > + return MEMTX_OK; > > > + break; > > > + default: > > > + break; > > > + } > > > + return MEMTX_OK; > > > + } else { > > > + > > > + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RXE)) { > > > + case RXE_ERROR: > > > + return MEMTX_ERROR; > > > + break; > > > + case RXE_SUCCESS_VALUE: > > > + *data = s->fabricated_v; > > > + return MEMTX_OK; > > > + break; > > > + default: > > > + break; > > > + } > > > + return MEMTX_OK; > > > + } > > > + return MEMTX_OK; > > > +} > > > + > > > +static MemTxResult iopmp_block_write(void *opaque, hwaddr addr, uint64_t value, > > > + unsigned size, MemTxAttrs attrs) > > > +{ > > > + return iopmp_handle_block(opaque, addr, &value, size, attrs, > > > + IOPMP_ACCESS_WRITE); > > > +} > > > + > > > +static MemTxResult iopmp_block_read(void *opaque, hwaddr addr, uint64_t *pdata, > > > + unsigned size, MemTxAttrs attrs) > > > +{ > > > + return iopmp_handle_block(opaque, addr, pdata, size, attrs, > > > + IOPMP_ACCESS_READ); > > > +} > > > + > > > +static MemTxResult iopmp_block_fetch(void *opaque, hwaddr addr, uint64_t *pdata, > > > + unsigned size, MemTxAttrs attrs) > > > +{ > > > + return iopmp_handle_block(opaque, addr, pdata, size, attrs, > > > + IOPMP_ACCESS_FETCH); > > > +} > > > + > > > +static const MemoryRegionOps iopmp_block_rw_ops = { > > > + .fetch_with_attrs = iopmp_permssion_read, > > > + .read_with_attrs = iopmp_block_read, > > > + .write_with_attrs = iopmp_block_write, > > > + .endianness = DEVICE_NATIVE_ENDIAN, > > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > > +}; > > > + > > > +static const MemoryRegionOps iopmp_block_w_ops = { > > > + .fetch_with_attrs = iopmp_permssion_read, > > > + .read_with_attrs = iopmp_permssion_read, > > > + .write_with_attrs = iopmp_block_write, > > > + .endianness = DEVICE_NATIVE_ENDIAN, > > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > > +}; > > > + > > > +static const MemoryRegionOps iopmp_block_r_ops = { > > > + .fetch_with_attrs = iopmp_permssion_read, > > > + .read_with_attrs = iopmp_block_read, > > > + .write_with_attrs = iopmp_permssion_write, > > > + .endianness = DEVICE_NATIVE_ENDIAN, > > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > > +}; > > > + > > > +static const MemoryRegionOps iopmp_block_rwx_ops = { > > > + .fetch_with_attrs = iopmp_block_fetch, > > > + .read_with_attrs = iopmp_block_read, > > > + .write_with_attrs = iopmp_block_write, > > > + .endianness = DEVICE_NATIVE_ENDIAN, > > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > > +}; > > > + > > > +static const MemoryRegionOps iopmp_block_wx_ops = { > > > + .fetch_with_attrs = iopmp_block_fetch, > > > + .read_with_attrs = iopmp_permssion_read, > > > + .write_with_attrs = iopmp_block_write, > > > + .endianness = DEVICE_NATIVE_ENDIAN, > > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > > +}; > > > + > > > +static const MemoryRegionOps iopmp_block_rx_ops = { > > > + .fetch_with_attrs = iopmp_block_fetch, > > > + .read_with_attrs = iopmp_block_read, > > > + .write_with_attrs = iopmp_permssion_write, > > > + .endianness = DEVICE_NATIVE_ENDIAN, > > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > > +}; > > > + > > > +static const MemoryRegionOps iopmp_block_x_ops = { > > > + .fetch_with_attrs = iopmp_block_fetch, > > > + .read_with_attrs = iopmp_permssion_read, > > > + .write_with_attrs = iopmp_permssion_write, > > > + .endianness = DEVICE_NATIVE_ENDIAN, > > > + .valid = {.min_access_size = 1, .max_access_size = 8}, > > > +}; > > > + > > > +static void iopmp_realize(DeviceState *dev, Error **errp) > > > +{ > > > + Object *obj = OBJECT(dev); > > > + SysBusDevice *sbd = SYS_BUS_DEVICE(dev); > > > + IopmpState *s = IOPMP(dev); > > > + uint64_t size; > > > + > > > + size = -1ULL; > > > + s->model = IOPMP_MODEL_RAPIDK; > > > > Should this be a property to allow other models in the future? > > Yes, it will be refined in next revision. > > > > > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, s->md_num); > > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, 1); > > > + > > > + s->prient_prog = s->default_prient_prog; > > > + s->rrid_num = MIN(s->rrid_num, IOPMP_MAX_RRID_NUM); > > > + s->md_num = MIN(s->md_num, IOPMP_MAX_MD_NUM); > > > + s->entry_num = s->md_num * s->k; > > > + s->prio_entry = MIN(s->prio_entry, s->entry_num); > > > + > > > + s->regs.mdcfg = g_malloc0(s->md_num * sizeof(uint32_t)); > > > + s->regs.mdcfg[0] = s->k; > > > + > > > + s->regs.srcmd_en = g_malloc0(s->rrid_num * sizeof(uint32_t)); > > > + s->regs.srcmd_enh = g_malloc0(s->rrid_num * sizeof(uint32_t)); > > > + s->regs.entry = g_malloc0(s->entry_num * sizeof(iopmp_entry_t)); > > > + s->entry_addr = g_malloc0(s->entry_num * sizeof(iopmp_addr_t)); > > > + s->transaction_state = g_malloc0(s->rrid_num * > > > + sizeof(iopmp_transaction_state)); > > > + qemu_mutex_init(&s->iopmp_transaction_mutex); > > > + > > > + memory_region_init_iommu(&s->iommu, sizeof(s->iommu), > > > + TYPE_IOPMP_IOMMU_MEMORY_REGION, > > > + obj, "riscv-iopmp-sysbus-iommu", UINT64_MAX); > > > + memory_region_init_io(&s->mmio, obj, &iopmp_ops, > > > + s, "iopmp-regs", 0x100000); > > > + sysbus_init_mmio(sbd, &s->mmio); > > > + > > > + memory_region_init_io(&s->blocked_rw, NULL, &iopmp_block_rw_ops, > > > + s, "iopmp-blocked-rw", size); > > > + memory_region_init_io(&s->blocked_w, NULL, &iopmp_block_w_ops, > > > + s, "iopmp-blocked-w", size); > > > + memory_region_init_io(&s->blocked_r, NULL, &iopmp_block_r_ops, > > > + s, "iopmp-blocked-r", size); > > > + > > > + memory_region_init_io(&s->blocked_rwx, NULL, &iopmp_block_rwx_ops, > > > + s, "iopmp-blocked-rwx", size); > > > + memory_region_init_io(&s->blocked_wx, NULL, &iopmp_block_wx_ops, > > > + s, "iopmp-blocked-wx", size); > > > + memory_region_init_io(&s->blocked_rx, NULL, &iopmp_block_rx_ops, > > > + s, "iopmp-blocked-rx", size); > > > + memory_region_init_io(&s->blocked_x, NULL, &iopmp_block_x_ops, > > > + s, "iopmp-blocked-x", size); > > > + address_space_init(&s->blocked_rw_as, &s->blocked_rw, > > > + "iopmp-blocked-rw-as"); > > > + address_space_init(&s->blocked_w_as, &s->blocked_w, > > > + "iopmp-blocked-w-as"); > > > + address_space_init(&s->blocked_r_as, &s->blocked_r, > > > + "iopmp-blocked-r-as"); > > > + > > > + address_space_init(&s->blocked_rwx_as, &s->blocked_rwx, > > > + "iopmp-blocked-rwx-as"); > > > + address_space_init(&s->blocked_wx_as, &s->blocked_wx, > > > + "iopmp-blocked-wx-as"); > > > + address_space_init(&s->blocked_rx_as, &s->blocked_rx, > > > + "iopmp-blocked-rx-as"); > > > + address_space_init(&s->blocked_x_as, &s->blocked_x, > > > + "iopmp-blocked-x-as"); > > > +} > > > + > > > +static void iopmp_reset(DeviceState *dev) > > > +{ > > > + IopmpState *s = IOPMP(dev); > > > + > > > + qemu_set_irq(s->irq, 0); > > > + memset(s->regs.srcmd_en, 0, s->rrid_num * sizeof(uint32_t)); > > > + memset(s->regs.srcmd_enh, 0, s->rrid_num * sizeof(uint32_t)); > > > + memset(s->entry_addr, 0, s->entry_num * sizeof(iopmp_addr_t)); > > > + > > > + s->regs.mdlck = 0; > > > + s->regs.mdlckh = 0; > > > + s->regs.entrylck = 0; > > > + s->regs.mdstall = 0; > > > + s->regs.mdstallh = 0; > > > + s->regs.rridscp = 0; > > > + s->regs.err_cfg = 0; > > > + s->regs.err_reqaddr = 0; > > > + s->regs.err_reqid = 0; > > > + s->regs.err_reqinfo = 0; > > > + > > > + s->prient_prog = s->default_prient_prog; > > > + s->enable = 0; > > > + > > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, s->md_num); > > > + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, 1); > > > + s->regs.mdcfg[0] = s->k; > > > +} > > > + > > > +static int iopmp_attrs_to_index(IOMMUMemoryRegion *iommu, MemTxAttrs attrs) > > > +{ > > > + return attrs.requester_id; > > > +} > > > + > > > +static void iopmp_iommu_memory_region_class_init(ObjectClass *klass, void *data) > > > +{ > > > + IOMMUMemoryRegionClass *imrc = IOMMU_MEMORY_REGION_CLASS(klass); > > > + > > > + imrc->translate = iopmp_translate; > > > + imrc->attrs_to_index = iopmp_attrs_to_index; > > > +} > > > + > > > +static Property iopmp_property[] = { > > > + DEFINE_PROP_BOOL("prient_prog", IopmpState, default_prient_prog, true), > > > + DEFINE_PROP_UINT32("k", IopmpState, k, 6), > > > + DEFINE_PROP_UINT32("prio_entry", IopmpState, prio_entry, 48), > > > + DEFINE_PROP_UINT32("rrid_num", IopmpState, rrid_num, 16), > > > + DEFINE_PROP_UINT32("md_num", IopmpState, md_num, 8), > > > + DEFINE_PROP_UINT32("entry_offset", IopmpState, entry_offset, 0x4000), > > > + DEFINE_PROP_UINT32("fabricated_v", IopmpState, fabricated_v, 0x0), > > > + DEFINE_PROP_END_OF_LIST(), > > > +}; > > > + > > > +static void iopmp_class_init(ObjectClass *klass, void *data) > > > +{ > > > + DeviceClass *dc = DEVICE_CLASS(klass); > > > + device_class_set_props(dc, iopmp_property); > > > + dc->realize = iopmp_realize; > > > + dc->reset = iopmp_reset; > > > +} > > > + > > > +static void iopmp_init(Object *obj) > > > +{ > > > + IopmpState *s = IOPMP(obj); > > > + SysBusDevice *sbd = SYS_BUS_DEVICE(obj); > > > + > > > + sysbus_init_irq(sbd, &s->irq); > > > +} > > > + > > > +static const TypeInfo iopmp_info = { > > > + .name = TYPE_IOPMP, > > > + .parent = TYPE_SYS_BUS_DEVICE, > > > + .instance_size = sizeof(IopmpState), > > > + .instance_init = iopmp_init, > > > + .class_init = iopmp_class_init, > > > +}; > > > + > > > +static const TypeInfo > > > +iopmp_iommu_memory_region_info = { > > > + .name = TYPE_IOPMP_IOMMU_MEMORY_REGION, > > > + .parent = TYPE_IOMMU_MEMORY_REGION, > > > + .class_init = iopmp_iommu_memory_region_class_init, > > > +}; > > > + > > > +static void > > > +iopmp_register_types(void) > > > +{ > > > + type_register_static(&iopmp_info); > > > + type_register_static(&iopmp_iommu_memory_region_info); > > > +} > > > + > > > +type_init(iopmp_register_types); > > > diff --git a/hw/misc/trace-events b/hw/misc/trace-events > > > index 1be0717c0c..c148166d2d 100644 > > > --- a/hw/misc/trace-events > > > +++ b/hw/misc/trace-events > > > @@ -362,3 +362,6 @@ aspeed_sli_read(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx > > > aspeed_sliio_write(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx64 " of size %u: 0x%" PRIx32 > > > aspeed_sliio_read(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx64 " of size %u: 0x%" PRIx32 > > > > > > +# riscv_iopmp.c > > > +iopmp_read(uint64_t addr, uint32_t val) "addr 0x%"PRIx64" val 0x%x" > > > +iopmp_write(uint64_t addr, uint32_t val) "addr 0x%"PRIx64" val 0x%x" > > > diff --git a/include/hw/misc/riscv_iopmp.h b/include/hw/misc/riscv_iopmp.h > > > new file mode 100644 > > > index 0000000000..b8fe479108 > > > --- /dev/null > > > +++ b/include/hw/misc/riscv_iopmp.h > > > @@ -0,0 +1,168 @@ > > > +/* > > > + * QEMU RISC-V IOPMP (Input Output Physical Memory Protection) > > > + * > > > + * Copyright (c) 2023-2024 Andes Tech. Corp. > > > + * > > > + * SPDX-License-Identifier: GPL-2.0-or-later > > > + * > > > + * This program is free software; you can redistribute it and/or modify it > > > + * under the terms and conditions of the GNU General Public License, > > > + * version 2 or later, as published by the Free Software Foundation. > > > + * > > > + * This program is distributed in the hope it will be useful, but WITHOUT > > > + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or > > > + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for > > > + * more details. > > > + * > > > + * You should have received a copy of the GNU General Public License along with > > > + * this program. If not, see <http://www.gnu.org/licenses/>. > > > + */ > > > + > > > +#ifndef RISCV_IOPMP_H > > > +#define RISCV_IOPMP_H > > > + > > > +#include "hw/sysbus.h" > > > +#include "qemu/typedefs.h" > > > +#include "memory.h" > > > +#include "exec/hwaddr.h" > > > + > > > +#define TYPE_IOPMP "iopmp" > > > +#define IOPMP(obj) OBJECT_CHECK(IopmpState, (obj), TYPE_IOPMP) > > > + > > > +#define IOPMP_MAX_MD_NUM 63 > > > +#define IOPMP_MAX_RRID_NUM 65535 > > > +#define IOPMP_MAX_ENTRY_NUM 65535 > > > + > > > +#define VENDER_VIRT 0 > > > +#define SPECVER_0_9_1 91 > > > +#define IMPID_0_9_1 91 > > > + > > > +#define RRE_ERROR 0 > > > +#define RRE_SUCCESS_VALUE 1 > > > + > > > +#define RWE_ERROR 0 > > > +#define RWE_SUCCESS 1 > > > + > > > +#define RXE_ERROR 0 > > > +#define RXE_SUCCESS_VALUE 1 > > > + > > > +#define ERR_REQINFO_TTYPE_READ 1 > > > +#define ERR_REQINFO_TTYPE_WRITE 2 > > > +#define ERR_REQINFO_TTYPE_FETCH 3 > > > +#define ERR_REQINFO_ETYPE_NOERROR 0 > > > +#define ERR_REQINFO_ETYPE_READ 1 > > > +#define ERR_REQINFO_ETYPE_WRITE 2 > > > +#define ERR_REQINFO_ETYPE_FETCH 3 > > > +#define ERR_REQINFO_ETYPE_PARHIT 4 > > > +#define ERR_REQINFO_ETYPE_NOHIT 5 > > > +#define ERR_REQINFO_ETYPE_RRID 6 > > > +#define ERR_REQINFO_ETYPE_USER 7 > > > + > > > +#define IOPMP_MODEL_FULL 0 > > > +#define IOPMP_MODEL_RAPIDK 0x1 > > > +#define IOPMP_MODEL_DYNAMICK 0x2 > > > +#define IOPMP_MODEL_ISOLATION 0x3 > > > +#define IOPMP_MODEL_COMPACTK 0x4 > > > + > > > +#define ENTRY_NO_HIT 0 > > > +#define ENTRY_PAR_HIT 1 > > > +#define ENTRY_HIT 2 > > > > Why not an enum? > > > > Thank you for your suggestion. There will be enums in next version. > > Thanks, > Ethan Chen > > > Alistair > > > > > + > > > +/* The generic iopmp address space which downstream is system memory */ > > > +extern AddressSpace iopmp_container_as; > > > + > > > +typedef enum { > > > + IOPMP_AMATCH_OFF, /* Null (off) */ > > > + IOPMP_AMATCH_TOR, /* Top of Range */ > > > + IOPMP_AMATCH_NA4, /* Naturally aligned four-byte region */ > > > + IOPMP_AMATCH_NAPOT /* Naturally aligned power-of-two region */ > > > +} iopmp_am_t; > > > + > > > +typedef enum { > > > + IOPMP_ACCESS_READ = 1, > > > + IOPMP_ACCESS_WRITE = 2, > > > + IOPMP_ACCESS_FETCH = 3 > > > +} iopmp_access_type; > > > + > > > +typedef enum { > > > + IOPMP_NONE = 0, > > > + IOPMP_RO = 1, > > > + IOPMP_WO = 2, > > > + IOPMP_RW = 3, > > > + IOPMP_XO = 4, > > > + IOPMP_RX = 5, > > > + IOPMP_WX = 6, > > > + IOPMP_RWX = 7, > > > +} iopmp_permission; > > > + > > > +typedef struct { > > > + uint32_t addr_reg; > > > + uint32_t addrh_reg; > > > + uint32_t cfg_reg; > > > +} iopmp_entry_t; > > > + > > > +typedef struct { > > > + uint64_t sa; > > > + uint64_t ea; > > > +} iopmp_addr_t; > > > + > > > +typedef struct { > > > + uint32_t *srcmd_en; > > > + uint32_t *srcmd_enh; > > > + uint32_t *mdcfg; > > > + iopmp_entry_t *entry; > > > + uint32_t mdlck; > > > + uint32_t mdlckh; > > > + uint32_t entrylck; > > > + uint32_t mdcfglck; > > > + uint32_t mdstall; > > > + uint32_t mdstallh; > > > + uint32_t rridscp; > > > + uint32_t err_cfg; > > > + uint64_t err_reqaddr; > > > + uint32_t err_reqid; > > > + uint32_t err_reqinfo; > > > +} iopmp_regs; > > > + > > > + > > > +/* To detect partially hit */ > > > +typedef struct iopmp_transaction_state { > > > + bool running; > > > + bool supported; > > > + hwaddr start_addr; > > > + hwaddr end_addr; > > > +} iopmp_transaction_state; > > > + > > > +typedef struct IopmpState { > > > + SysBusDevice parent_obj; > > > + iopmp_addr_t *entry_addr; > > > + MemoryRegion mmio; > > > + IOMMUMemoryRegion iommu; > > > + IOMMUMemoryRegion *next_iommu; > > > + iopmp_regs regs; > > > + MemoryRegion *downstream; > > > + MemoryRegion blocked_r, blocked_w, blocked_x, blocked_rw, blocked_rx, > > > + blocked_wx, blocked_rwx; > > > + MemoryRegion stall_io; > > > + uint32_t model; > > > + uint32_t k; > > > + bool prient_prog; > > > + bool default_prient_prog; > > > + iopmp_transaction_state *transaction_state; > > > + QemuMutex iopmp_transaction_mutex; > > > + > > > + AddressSpace downstream_as; > > > + AddressSpace blocked_r_as, blocked_w_as, blocked_x_as, blocked_rw_as, > > > + blocked_rx_as, blocked_wx_as, blocked_rwx_as; > > > + qemu_irq irq; > > > + bool enable; > > > + > > > + uint32_t prio_entry; > > > + uint32_t rrid_num; > > > + uint32_t md_num; > > > + uint32_t entry_num; > > > + uint32_t entry_offset; > > > + uint32_t fabricated_v; > > > +} IopmpState; > > > + > > > +#endif > > > -- > > > 2.34.1 > > > > > >
diff --git a/hw/misc/Kconfig b/hw/misc/Kconfig index 1e08785b83..427f0c702e 100644 --- a/hw/misc/Kconfig +++ b/hw/misc/Kconfig @@ -213,4 +213,7 @@ config IOSB config XLNX_VERSAL_TRNG bool +config RISCV_IOPMP + bool + source macio/Kconfig diff --git a/hw/misc/meson.build b/hw/misc/meson.build index 2ca8717be2..d9006e1d81 100644 --- a/hw/misc/meson.build +++ b/hw/misc/meson.build @@ -34,6 +34,7 @@ system_ss.add(when: 'CONFIG_SIFIVE_E_PRCI', if_true: files('sifive_e_prci.c')) system_ss.add(when: 'CONFIG_SIFIVE_E_AON', if_true: files('sifive_e_aon.c')) system_ss.add(when: 'CONFIG_SIFIVE_U_OTP', if_true: files('sifive_u_otp.c')) system_ss.add(when: 'CONFIG_SIFIVE_U_PRCI', if_true: files('sifive_u_prci.c')) +specific_ss.add(when: 'CONFIG_RISCV_IOPMP', if_true: files('riscv_iopmp.c')) subdir('macio') diff --git a/hw/misc/riscv_iopmp.c b/hw/misc/riscv_iopmp.c new file mode 100644 index 0000000000..db43e3c73f --- /dev/null +++ b/hw/misc/riscv_iopmp.c @@ -0,0 +1,1154 @@ +/* + * QEMU RISC-V IOPMP (Input Output Physical Memory Protection) + * + * Copyright (c) 2023-2024 Andes Tech. Corp. + * + * SPDX-License-Identifier: GPL-2.0-or-later + * + * This program is free software; you can redistribute it and/or modify it + * under the terms and conditions of the GNU General Public License, + * version 2 or later, as published by the Free Software Foundation. + * + * This program is distributed in the hope it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for + * more details. + * + * You should have received a copy of the GNU General Public License along with + * this program. If not, see <http://www.gnu.org/licenses/>. + */ + +#include "qemu/osdep.h" +#include "qemu/log.h" +#include "qapi/error.h" +#include "trace.h" +#include "exec/exec-all.h" +#include "exec/address-spaces.h" +#include "hw/qdev-properties.h" +#include "hw/sysbus.h" +#include "hw/misc/riscv_iopmp.h" +#include "memory.h" +#include "hw/irq.h" +#include "hw/registerfields.h" +#include "trace.h" + +#define TYPE_IOPMP_IOMMU_MEMORY_REGION "iopmp-iommu-memory-region" + +REG32(VERSION, 0x00) + FIELD(VERSION, VENDOR, 0, 24) + FIELD(VERSION, SPECVER , 24, 8) +REG32(IMP, 0x04) + FIELD(IMP, IMPID, 0, 32) +REG32(HWCFG0, 0x08) + FIELD(HWCFG0, MODEL, 0, 4) + FIELD(HWCFG0, TOR_EN, 4, 1) + FIELD(HWCFG0, SPS_EN, 5, 1) + FIELD(HWCFG0, USER_CFG_EN, 6, 1) + FIELD(HWCFG0, PRIENT_PROG, 7, 1) + FIELD(HWCFG0, RRID_TRANSL_EN, 8, 1) + FIELD(HWCFG0, RRID_TRANSL_PROG, 9, 1) + FIELD(HWCFG0, CHK_X, 10, 1) + FIELD(HWCFG0, NO_X, 11, 1) + FIELD(HWCFG0, NO_W, 12, 1) + FIELD(HWCFG0, STALL_EN, 13, 1) + FIELD(HWCFG0, PEIS, 14, 1) + FIELD(HWCFG0, PEES, 15, 1) + FIELD(HWCFG0, MFR_EN, 16, 1) + FIELD(HWCFG0, MD_NUM, 24, 7) + FIELD(HWCFG0, ENABLE, 31, 1) +REG32(HWCFG1, 0x0C) + FIELD(HWCFG1, RRID_NUM, 0, 16) + FIELD(HWCFG1, ENTRY_NUM, 16, 16) +REG32(HWCFG2, 0x10) + FIELD(HWCFG2, PRIO_ENTRY, 0, 16) + FIELD(HWCFG2, RRID_TRANSL, 16, 16) +REG32(ENTRYOFFSET, 0x14) + FIELD(ENTRYOFFSET, OFFSET, 0, 32) +REG32(MDSTALL, 0x30) + FIELD(MDSTALL, EXEMPT, 0, 1) + FIELD(MDSTALL, MD, 1, 31) +REG32(MDSTALLH, 0x34) + FIELD(MDSTALLH, MD, 0, 32) +REG32(RRIDSCP, 0x38) + FIELD(RRIDSCP, RRID, 0, 16) + FIELD(RRIDSCP, OP, 30, 2) +REG32(MDLCK, 0x40) + FIELD(MDLCK, L, 0, 1) + FIELD(MDLCK, MD, 1, 31) +REG32(MDLCKH, 0x44) + FIELD(MDLCKH, MDH, 0, 32) +REG32(MDCFGLCK, 0x48) + FIELD(MDCFGLCK, L, 0, 1) + FIELD(MDCFGLCK, F, 1, 7) +REG32(ENTRYLCK, 0x4C) + FIELD(ENTRYLCK, L, 0, 1) + FIELD(ENTRYLCK, F, 1, 16) +REG32(ERR_CFG, 0x60) + FIELD(ERR_CFG, L, 0, 1) + FIELD(ERR_CFG, IE, 1, 1) + FIELD(ERR_CFG, IRE, 2, 1) + FIELD(ERR_CFG, IWE, 3, 1) + FIELD(ERR_CFG, IXE, 4, 1) + FIELD(ERR_CFG, RRE, 5, 1) + FIELD(ERR_CFG, RWE, 6, 1) + FIELD(ERR_CFG, RXE, 7, 1) +REG32(ERR_REQINFO, 0x64) + FIELD(ERR_REQINFO, V, 0, 1) + FIELD(ERR_REQINFO, TTYPE, 1, 2) + FIELD(ERR_REQINFO, ETYPE, 4, 3) + FIELD(ERR_REQINFO, SVC, 7, 1) +REG32(ERR_REQADDR, 0x68) + FIELD(ERR_REQADDR, ADDR, 0, 32) +REG32(ERR_REQADDRH, 0x6C) + FIELD(ERR_REQADDRH, ADDRH, 0, 32) +REG32(ERR_REQID, 0x70) + FIELD(ERR_REQID, RRID, 0, 16) + FIELD(ERR_REQID, EID, 16, 16) +REG32(ERR_MFR, 0x74) + FIELD(ERR_MFR, SVW, 0, 16) + FIELD(ERR_MFR, SVI, 16, 12) + FIELD(ERR_MFR, SVS, 31, 1) +REG32(MDCFG0, 0x800) + FIELD(MDCFG0, T, 0, 16) +REG32(SRCMD_EN0, 0x1000) + FIELD(SRCMD_EN0, L, 0, 1) + FIELD(SRCMD_EN0, MD, 1, 31) +REG32(SRCMD_ENH0, 0x1004) + FIELD(SRCMD_ENH0, MDH, 0, 32) +REG32(SRCMD_R0, 0x1008) + FIELD(SRCMD_R0, MD, 1, 31) +REG32(SRCMD_RH0, 0x100C) + FIELD(SRCMD_RH0, MDH, 0, 32) +REG32(SRCMD_W0, 0x1010) + FIELD(SRCMD_W0, MD, 1, 31) +REG32(SRCMD_WH0, 0x1014) + FIELD(SRCMD_WH0, MDH, 0, 32) + +FIELD(ENTRY_ADDR, ADDR, 0, 32) +FIELD(ENTRY_ADDRH, ADDRH, 0, 32) + +FIELD(ENTRY_CFG, R, 0, 1) +FIELD(ENTRY_CFG, W, 1, 1) +FIELD(ENTRY_CFG, X, 2, 1) +FIELD(ENTRY_CFG, A, 3, 2) +FIELD(ENTRY_CFG, SIRE, 5, 1) +FIELD(ENTRY_CFG, SIWE, 6, 1) +FIELD(ENTRY_CFG, SIXE, 7, 1) +FIELD(ENTRY_CFG, SERE, 8, 1) +FIELD(ENTRY_CFG, SEWE, 9, 1) +FIELD(ENTRY_CFG, SEXE, 10, 1) + +FIELD(ENTRY_USER_CFG, IM, 0, 32) + +/* Offsets to SRCMD_EN(i) */ +#define SRCMD_EN_OFFSET 0x0 +#define SRCMD_ENH_OFFSET 0x4 +#define SRCMD_R_OFFSET 0x8 +#define SRCMD_RH_OFFSET 0xC +#define SRCMD_W_OFFSET 0x10 +#define SRCMD_WH_OFFSET 0x14 + +/* Offsets to ENTRY_ADDR(i) */ +#define ENTRY_ADDR_OFFSET 0x0 +#define ENTRY_ADDRH_OFFSET 0x4 +#define ENTRY_CFG_OFFSET 0x8 +#define ENTRY_USER_CFG_OFFSET 0xC + +/* Memmap for parallel IOPMPs */ +typedef struct iopmp_protection_memmap { + MemMapEntry entry; + IopmpState *iopmp_s; + QLIST_ENTRY(iopmp_protection_memmap) list; +} iopmp_protection_memmap; +QLIST_HEAD(, iopmp_protection_memmap) + iopmp_protection_memmaps = QLIST_HEAD_INITIALIZER(iopmp_protection_memmaps); + +static void iopmp_iommu_notify(IopmpState *s) +{ + IOMMUTLBEvent event = { + .entry = { + .iova = 0, + .translated_addr = 0, + .addr_mask = -1ULL, + .perm = IOMMU_NONE, + }, + .type = IOMMU_NOTIFIER_UNMAP, + }; + + for (int i = 0; i < s->rrid_num; i++) { + memory_region_notify_iommu(&s->iommu, i, event); + } +} + +static void iopmp_decode_napot(uint64_t a, uint64_t *sa, + uint64_t *ea) +{ + /* + * aaaa...aaa0 8-byte NAPOT range + * aaaa...aa01 16-byte NAPOT range + * aaaa...a011 32-byte NAPOT range + * ... + * aa01...1111 2^XLEN-byte NAPOT range + * a011...1111 2^(XLEN+1)-byte NAPOT range + * 0111...1111 2^(XLEN+2)-byte NAPOT range + * 1111...1111 Reserved + */ + + a = (a << 2) | 0x3; + *sa = a & (a + 1); + *ea = a | (a + 1); +} + +static void iopmp_update_rule(IopmpState *s, uint32_t entry_index) +{ + uint8_t this_cfg = s->regs.entry[entry_index].cfg_reg; + uint64_t this_addr = s->regs.entry[entry_index].addr_reg | + ((uint64_t)s->regs.entry[entry_index].addrh_reg << 32); + uint64_t prev_addr = 0u; + uint64_t sa = 0u; + uint64_t ea = 0u; + + if (entry_index >= 1u) { + prev_addr = s->regs.entry[entry_index - 1].addr_reg | + ((uint64_t)s->regs.entry[entry_index - 1].addrh_reg << 32); + } + + switch (FIELD_EX32(this_cfg, ENTRY_CFG, A)) { + case IOPMP_AMATCH_OFF: + sa = 0u; + ea = -1; + break; + + case IOPMP_AMATCH_TOR: + sa = (prev_addr) << 2; /* shift up from [xx:0] to [xx+2:2] */ + ea = ((this_addr) << 2) - 1u; + if (sa > ea) { + sa = ea = 0u; + } + break; + + case IOPMP_AMATCH_NA4: + sa = this_addr << 2; /* shift up from [xx:0] to [xx+2:2] */ + ea = (sa + 4u) - 1u; + break; + + case IOPMP_AMATCH_NAPOT: + iopmp_decode_napot(this_addr, &sa, &ea); + break; + + default: + sa = 0u; + ea = 0u; + break; + } + + s->entry_addr[entry_index].sa = sa; + s->entry_addr[entry_index].ea = ea; + iopmp_iommu_notify(s); +} + +static uint64_t iopmp_read(void *opaque, hwaddr addr, unsigned size) +{ + IopmpState *s = IOPMP(opaque); + uint32_t rz = 0; + uint32_t offset, idx; + + switch (addr) { + case A_VERSION: + rz = VENDER_VIRT << R_VERSION_VENDOR_SHIFT | + SPECVER_0_9_1 << R_VERSION_SPECVER_SHIFT; + break; + case A_IMP: + rz = IMPID_0_9_1; + break; + case A_HWCFG0: + rz = s->model << R_HWCFG0_MODEL_SHIFT | + 1 << R_HWCFG0_TOR_EN_SHIFT | + 0 << R_HWCFG0_SPS_EN_SHIFT | + 0 << R_HWCFG0_USER_CFG_EN_SHIFT | + s->prient_prog << R_HWCFG0_PRIENT_PROG_SHIFT | + 0 << R_HWCFG0_RRID_TRANSL_EN_SHIFT | + 0 << R_HWCFG0_RRID_TRANSL_PROG_SHIFT | + 1 << R_HWCFG0_CHK_X_SHIFT | + 0 << R_HWCFG0_NO_X_SHIFT | + 0 << R_HWCFG0_NO_W_SHIFT | + 0 << R_HWCFG0_STALL_EN_SHIFT | + 0 << R_HWCFG0_PEIS_SHIFT | + 0 << R_HWCFG0_PEES_SHIFT | + 0 << R_HWCFG0_MFR_EN_SHIFT | + s->md_num << R_HWCFG0_MD_NUM_SHIFT | + s->enable << R_HWCFG0_ENABLE_SHIFT ; + break; + case A_HWCFG1: + rz = s->rrid_num << R_HWCFG1_RRID_NUM_SHIFT | + s->entry_num << R_HWCFG1_ENTRY_NUM_SHIFT; + break; + case A_HWCFG2: + rz = s->prio_entry << R_HWCFG2_PRIO_ENTRY_SHIFT; + break; + case A_ENTRYOFFSET: + rz = s->entry_offset; + break; + case A_ERR_CFG: + rz = s->regs.err_cfg; + break; + case A_MDLCK: + rz = s->regs.mdlck; + break; + case A_MDLCKH: + rz = s->regs.mdlckh; + break; + case A_MDCFGLCK: + rz = s->regs.mdcfglck; + break; + case A_ENTRYLCK: + rz = s->regs.entrylck; + break; + case A_ERR_REQADDR: + rz = s->regs.err_reqaddr & UINT32_MAX; + break; + case A_ERR_REQADDRH: + rz = s->regs.err_reqaddr >> 32; + break; + case A_ERR_REQID: + rz = s->regs.err_reqid; + break; + case A_ERR_REQINFO: + rz = s->regs.err_reqinfo; + break; + + default: + if (addr >= A_MDCFG0 && + addr < A_MDCFG0 + 4 * (s->md_num - 1)) { + offset = addr - A_MDCFG0; + idx = offset >> 2; + if (idx == 0 && offset == 0) { + rz = s->regs.mdcfg[idx]; + } else { + /* Only MDCFG0 is implemented in rapid-k model */ + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", + __func__, (int)addr); + } + } else if (addr >= A_SRCMD_EN0 && + addr < A_SRCMD_WH0 + 32 * (s->rrid_num - 1)) { + offset = addr - A_SRCMD_EN0; + idx = offset >> 5; + offset &= 0x1f; + + switch (offset) { + case SRCMD_EN_OFFSET: + rz = s->regs.srcmd_en[idx]; + break; + case SRCMD_ENH_OFFSET: + rz = s->regs.srcmd_enh[idx]; + break; + default: + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", + __func__, (int)addr); + break; + } + } else if (addr >= s->entry_offset && + addr < s->entry_offset + ENTRY_USER_CFG_OFFSET + + 16 * (s->entry_num - 1)) { + offset = addr - s->entry_offset; + idx = offset >> 4; + offset &= 0xf; + + switch (offset) { + case ENTRY_ADDR_OFFSET: + rz = s->regs.entry[idx].addr_reg; + break; + case ENTRY_ADDRH_OFFSET: + rz = s->regs.entry[idx].addrh_reg; + break; + case ENTRY_CFG_OFFSET: + rz = s->regs.entry[idx].cfg_reg; + break; + case ENTRY_USER_CFG_OFFSET: + /* Does not support user customized permission */ + rz = 0; + break; + default: + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", + __func__, (int)addr); + break; + } + } else { + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", + __func__, (int)addr); + } + break; + } + trace_iopmp_read(addr, rz); + return rz; +} + +static void +iopmp_write(void *opaque, hwaddr addr, uint64_t value, unsigned size) +{ + IopmpState *s = IOPMP(opaque); + uint32_t offset, idx; + uint32_t value32 = value; + + trace_iopmp_write(addr, value32); + + switch (addr) { + case A_VERSION: /* RO */ + break; + case A_IMP: /* RO */ + break; + case A_HWCFG0: + if (FIELD_EX32(value32, HWCFG0, PRIENT_PROG)) { + /* W1C */ + s->prient_prog = 0; + } + if (FIELD_EX32(value32, HWCFG0, ENABLE)) { + /* W1S */ + s->enable = 1; + iopmp_iommu_notify(s); + } + break; + case A_HWCFG1: /* RO */ + break; + case A_HWCFG2: + if (s->prient_prog) { + s->prio_entry = FIELD_EX32(value32, HWCFG2, PRIO_ENTRY); + } + break; + case A_ERR_CFG: + if (!FIELD_EX32(s->regs.err_cfg, ERR_CFG, L)) { + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, L, + FIELD_EX32(value32, ERR_CFG, L)); + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IE, + FIELD_EX32(value32, ERR_CFG, IE)); + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IRE, + FIELD_EX32(value32, ERR_CFG, IRE)); + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RRE, + FIELD_EX32(value32, ERR_CFG, RRE)); + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IWE, + FIELD_EX32(value32, ERR_CFG, IWE)); + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RWE, + FIELD_EX32(value32, ERR_CFG, RWE)); + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, IXE, + FIELD_EX32(value32, ERR_CFG, IXE)); + s->regs.err_cfg = FIELD_DP32(s->regs.err_cfg, ERR_CFG, RXE, + FIELD_EX32(value32, ERR_CFG, RXE)); + } + break; + case A_MDLCK: + if (!FIELD_EX32(s->regs.mdlck, MDLCK, L)) { + s->regs.mdlck = value32; + } + break; + case A_MDLCKH: + if (!FIELD_EX32(s->regs.mdlck, MDLCK, L)) { + s->regs.mdlckh = value32; + } + break; + case A_MDCFGLCK: + if (!FIELD_EX32(s->regs.mdcfglck, MDCFGLCK, L)) { + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, + FIELD_EX32(value32, MDCFGLCK, F)); + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, + FIELD_EX32(value32, MDCFGLCK, L)); + } + break; + case A_ENTRYLCK: + if (!(FIELD_EX32(s->regs.entrylck, ENTRYLCK, L))) { + s->regs.entrylck = FIELD_DP32(s->regs.entrylck, ENTRYLCK, F, + FIELD_EX32(value32, ENTRYLCK, F)); + s->regs.entrylck = FIELD_DP32(s->regs.entrylck, ENTRYLCK, L, + FIELD_EX32(value32, ENTRYLCK, L)); + } + case A_ERR_REQADDR: /* RO */ + break; + case A_ERR_REQADDRH: /* RO */ + break; + case A_ERR_REQID: /* RO */ + break; + case A_ERR_REQINFO: + if (FIELD_EX32(value32, ERR_REQINFO, V)) { + s->regs.err_reqinfo = FIELD_DP32(s->regs.err_reqinfo, + ERR_REQINFO, V, 0); + qemu_set_irq(s->irq, 0); + } + break; + + default: + if (addr >= A_MDCFG0 && + addr < A_MDCFG0 + 4 * (s->md_num - 1)) { + offset = addr - A_MDCFG0; + idx = offset >> 2; + /* RO in rapid-k model */ + if (idx > 0) { + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", + __func__, (int)addr); + } + } else if (addr >= A_SRCMD_EN0 && + addr < A_SRCMD_WH0 + 32 * (s->rrid_num - 1)) { + offset = addr - A_SRCMD_EN0; + idx = offset >> 5; + offset &= 0x1f; + + if (offset % 4) { + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", + __func__, (int)addr); + } else if (FIELD_EX32(s->regs.srcmd_en[idx], SRCMD_EN0, L) + == 0) { + switch (offset) { + case SRCMD_EN_OFFSET: + s->regs.srcmd_en[idx] = + FIELD_DP32(s->regs.srcmd_en[idx], SRCMD_EN0, L, + FIELD_EX32(value32, SRCMD_EN0, L)); + + /* MD field is protected by mdlck */ + value32 = (value32 & ~s->regs.mdlck) | + (s->regs.srcmd_en[idx] & s->regs.mdlck); + s->regs.srcmd_en[idx] = + FIELD_DP32(s->regs.srcmd_en[idx], SRCMD_EN0, MD, + FIELD_EX32(value32, SRCMD_EN0, MD)); + break; + case SRCMD_ENH_OFFSET: + value32 = (value32 & ~s->regs.mdlckh) | + (s->regs.srcmd_enh[idx] & s->regs.mdlckh); + s->regs.srcmd_enh[idx] = + FIELD_DP32(s->regs.srcmd_enh[idx], SRCMD_ENH0, MDH, + value32); + break; + default: + break; + } + } + } else if (addr >= s->entry_offset && + addr < s->entry_offset + ENTRY_USER_CFG_OFFSET + + 16 * (s->entry_num - 1)) { + offset = addr - s->entry_offset; + idx = offset >> 4; + offset &= 0xf; + + /* index < ENTRYLCK_F is protected */ + if (idx >= FIELD_EX32(s->regs.entrylck, ENTRYLCK, F)) { + switch (offset) { + case ENTRY_ADDR_OFFSET: + s->regs.entry[idx].addr_reg = value32; + break; + case ENTRY_ADDRH_OFFSET: + s->regs.entry[idx].addrh_reg = value32; + break; + case ENTRY_CFG_OFFSET: + s->regs.entry[idx].cfg_reg = value32; + break; + case ENTRY_USER_CFG_OFFSET: + /* Does not support user customized permission */ + break; + default: + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", + __func__, (int)addr); + break; + } + iopmp_update_rule(s, idx); + if (idx + 1 < s->entry_num && + FIELD_EX32(s->regs.entry[idx + 1].cfg_reg, ENTRY_CFG, A) == + IOPMP_AMATCH_TOR) { + iopmp_update_rule(s, idx + 1); + } + } + } else { + qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad addr %x\n", __func__, + (int)addr); + } + } +} + +/* Match entry in memory domain */ +static int match_entry_md(IopmpState *s, int md_idx, hwaddr start_addr, + hwaddr end_addr, int *entry_idx, + int *prior_entry_in_tlb) +{ + int entry_idx_s, entry_idx_e; + int result = ENTRY_NO_HIT; + int i = 0; + hwaddr tlb_sa = start_addr & ~(TARGET_PAGE_SIZE - 1); + hwaddr tlb_ea = tlb_sa + TARGET_PAGE_SIZE - 1; + + entry_idx_s = md_idx * s->regs.mdcfg[0]; + entry_idx_e = (md_idx + 1) * s->regs.mdcfg[0]; + + if (entry_idx_s >= s->entry_num) { + return result; + } + if (entry_idx_e > s->entry_num) { + entry_idx_e = s->entry_num; + } + i = entry_idx_s; + for (i = entry_idx_s; i < entry_idx_e; i++) { + if (FIELD_EX32(s->regs.entry[i].cfg_reg, ENTRY_CFG, A) == + IOPMP_AMATCH_OFF) { + continue; + } + if (start_addr >= s->entry_addr[i].sa && + start_addr <= s->entry_addr[i].ea) { + /* Check end address */ + if (end_addr >= s->entry_addr[i].sa && + end_addr <= s->entry_addr[i].ea) { + *entry_idx = i; + return ENTRY_HIT; + } else if (i >= s->prio_entry) { + /* Continue for non-prio_entry */ + continue; + } else { + *entry_idx = i; + return ENTRY_PAR_HIT; + } + } else if (end_addr >= s->entry_addr[i].sa && + end_addr <= s->entry_addr[i].ea) { + /* Only end address matches the entry */ + if (i >= s->prio_entry) { + continue; + } else { + *entry_idx = i; + return ENTRY_PAR_HIT; + } + } else if (start_addr < s->entry_addr[i].sa && + end_addr > s->entry_addr[i].ea) { + if (i >= s->prio_entry) { + continue; + } else { + *entry_idx = i; + return ENTRY_PAR_HIT; + } + } + if (prior_entry_in_tlb != NULL) { + if ((s->entry_addr[i].sa >= tlb_sa && + s->entry_addr[i].sa <= tlb_ea) || + (s->entry_addr[i].ea >= tlb_sa && + s->entry_addr[i].ea <= tlb_ea)) { + /* + * TLB should not use the cached result when the tlb contains + * higher priority entry + */ + *prior_entry_in_tlb = 1; + } + } + } + return result; +} + +static int match_entry(IopmpState *s, int rrid, hwaddr start_addr, + hwaddr end_addr, int *match_md_idx, + int *match_entry_idx, int *prior_entry_in_tlb) +{ + int cur_result = ENTRY_NO_HIT; + int result = ENTRY_NO_HIT; + /* Remove lock bit */ + uint64_t srcmd_en = ((uint64_t)s->regs.srcmd_en[rrid] | + ((uint64_t)s->regs.srcmd_enh[rrid] << 32)) >> 1; + + for (int md_idx = 0; md_idx < s->md_num; md_idx++) { + if (srcmd_en & (1ULL << md_idx)) { + cur_result = match_entry_md(s, md_idx, start_addr, end_addr, + match_entry_idx, prior_entry_in_tlb); + if (cur_result == ENTRY_HIT || cur_result == ENTRY_PAR_HIT) { + *match_md_idx = md_idx; + return cur_result; + } + } + } + return result; +} + +static void iopmp_error_reaction(IopmpState *s, uint32_t id, hwaddr start, + uint32_t info) +{ + if (!FIELD_EX32(s->regs.err_reqinfo, ERR_REQINFO, V)) { + s->regs.err_reqinfo = info; + s->regs.err_reqinfo = FIELD_DP32(s->regs.err_reqinfo, ERR_REQINFO, V, + 1); + s->regs.err_reqid = id; + /* addr[LEN+2:2] */ + s->regs.err_reqaddr = start >> 2; + + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_READ && + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IRE)) { + qemu_set_irq(s->irq, 1); + } + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_WRITE && + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IWE)) { + qemu_set_irq(s->irq, 1); + } + if (FIELD_EX32(info, ERR_REQINFO, TTYPE) == ERR_REQINFO_TTYPE_FETCH && + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IE) && + FIELD_EX32(s->regs.err_cfg, ERR_CFG, IXE)) { + qemu_set_irq(s->irq, 1); + } + } +} + +static IOMMUTLBEntry iopmp_translate(IOMMUMemoryRegion *iommu, hwaddr addr, + IOMMUAccessFlags flags, int iommu_idx) +{ + int rrid = iommu_idx; + IopmpState *s = IOPMP(container_of(iommu, IopmpState, iommu)); + hwaddr start_addr, end_addr; + int entry_idx = -1; + int md_idx = -1; + int result; + uint32_t error_info = 0; + uint32_t error_id = 0; + int prior_entry_in_tlb = 0; + iopmp_permission iopmp_perm; + IOMMUTLBEntry entry = { + .target_as = &s->downstream_as, + .iova = addr, + .translated_addr = addr, + .addr_mask = 0, + .perm = IOMMU_NONE, + }; + + if (!s->enable) { + /* Bypass IOPMP */ + entry.addr_mask = -1ULL, + entry.perm = IOMMU_RW; + return entry; + } + + /* unknown RRID */ + if (rrid >= s->rrid_num) { + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, + ERR_REQINFO_ETYPE_RRID); + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); + iopmp_error_reaction(s, error_id, addr, error_info); + entry.target_as = &s->blocked_rwx_as; + entry.perm = IOMMU_RW; + return entry; + } + + if (s->transaction_state[rrid].supported == true) { + start_addr = s->transaction_state[rrid].start_addr; + end_addr = s->transaction_state[rrid].end_addr; + } else { + /* No transaction information, use the same address */ + start_addr = addr; + end_addr = addr; + } + + result = match_entry(s, rrid, start_addr, end_addr, &md_idx, &entry_idx, + &prior_entry_in_tlb); + if (result == ENTRY_HIT) { + entry.addr_mask = s->entry_addr[entry_idx].ea - + s->entry_addr[entry_idx].sa; + if (prior_entry_in_tlb) { + /* Make TLB repeat iommu translation on every access */ + entry.addr_mask = 0; + } + iopmp_perm = s->regs.entry[entry_idx].cfg_reg & IOPMP_RWX; + if (flags) { + if ((iopmp_perm & flags) == 0) { + /* Permission denied */ + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, + ERR_REQINFO_ETYPE_READ + flags - 1); + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); + iopmp_error_reaction(s, error_id, start_addr, error_info); + entry.target_as = &s->blocked_rwx_as; + entry.perm = IOMMU_RW; + } else { + entry.target_as = &s->downstream_as; + entry.perm = iopmp_perm; + } + } else { + /* CPU access with IOMMU_NONE flag */ + if (iopmp_perm & IOPMP_XO) { + if ((iopmp_perm & IOPMP_RW) == IOPMP_RW) { + entry.target_as = &s->downstream_as; + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_RO) { + entry.target_as = &s->blocked_w_as; + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_WO) { + entry.target_as = &s->blocked_r_as; + } else { + entry.target_as = &s->blocked_rw_as; + } + } else { + if ((iopmp_perm & IOPMP_RW) == IOMMU_RW) { + entry.target_as = &s->blocked_x_as; + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_RO) { + entry.target_as = &s->blocked_wx_as; + } else if ((iopmp_perm & IOPMP_RW) == IOPMP_WO) { + entry.target_as = &s->blocked_rx_as; + } else { + entry.target_as = &s->blocked_rwx_as; + } + } + entry.perm = IOMMU_RW; + } + } else { + if (flags) { + if (result == ENTRY_PAR_HIT) { + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, + ERR_REQINFO_ETYPE_PARHIT); + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); + iopmp_error_reaction(s, error_id, start_addr, error_info); + } else { + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, + ERR_REQINFO_ETYPE_NOHIT); + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, flags); + iopmp_error_reaction(s, error_id, start_addr, error_info); + } + } + /* CPU access with IOMMU_NONE flag no_hit or par_hit */ + entry.target_as = &s->blocked_rwx_as; + entry.perm = IOMMU_RW; + } + return entry; +} + +static const MemoryRegionOps iopmp_ops = { + .read = iopmp_read, + .write = iopmp_write, + .endianness = DEVICE_NATIVE_ENDIAN, + .valid = {.min_access_size = 4, .max_access_size = 4} +}; + +static MemTxResult iopmp_permssion_write(void *opaque, hwaddr addr, + uint64_t value, unsigned size, + MemTxAttrs attrs) +{ + IopmpState *s = IOPMP(opaque); + return address_space_write(&s->downstream_as, addr, attrs, &value, size); +} + +static MemTxResult iopmp_permssion_read(void *opaque, hwaddr addr, + uint64_t *pdata, unsigned size, + MemTxAttrs attrs) +{ + IopmpState *s = IOPMP(opaque); + return address_space_read(&s->downstream_as, addr, attrs, pdata, size); +} + +static MemTxResult iopmp_handle_block(void *opaque, hwaddr addr, + uint64_t *data, unsigned size, + MemTxAttrs attrs, + iopmp_access_type access_type) { + IopmpState *s = IOPMP(opaque); + int md_idx, entry_idx; + uint32_t error_info = 0; + uint32_t error_id = 0; + int rrid = attrs.requester_id; + int result; + hwaddr start_addr, end_addr; + start_addr = addr; + end_addr = addr; + result = match_entry(s, rrid, start_addr, end_addr, &md_idx, &entry_idx, + NULL); + + if (result == ENTRY_HIT) { + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, + access_type); + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, access_type); + iopmp_error_reaction(s, error_id, start_addr, error_info); + } else if (result == ENTRY_PAR_HIT) { + error_id = FIELD_DP32(error_id, ERR_REQID, EID, entry_idx); + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, + ERR_REQINFO_ETYPE_PARHIT); + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, + access_type); + iopmp_error_reaction(s, error_id, start_addr, error_info); + } else { + error_id = FIELD_DP32(error_id, ERR_REQID, RRID, rrid); + error_info = FIELD_DP32(error_info, ERR_REQINFO, ETYPE, + ERR_REQINFO_ETYPE_NOHIT); + error_info = FIELD_DP32(error_info, ERR_REQINFO, TTYPE, access_type); + iopmp_error_reaction(s, error_id, start_addr, error_info); + } + + if (access_type == IOPMP_ACCESS_READ) { + + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RRE)) { + case RRE_ERROR: + return MEMTX_ERROR; + break; + case RRE_SUCCESS_VALUE: + *data = s->fabricated_v; + return MEMTX_OK; + break; + default: + break; + } + return MEMTX_OK; + } else if (access_type == IOPMP_ACCESS_WRITE) { + + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RWE)) { + case RWE_ERROR: + return MEMTX_ERROR; + break; + case RWE_SUCCESS: + return MEMTX_OK; + break; + default: + break; + } + return MEMTX_OK; + } else { + + switch (FIELD_EX32(s->regs.err_cfg, ERR_CFG, RXE)) { + case RXE_ERROR: + return MEMTX_ERROR; + break; + case RXE_SUCCESS_VALUE: + *data = s->fabricated_v; + return MEMTX_OK; + break; + default: + break; + } + return MEMTX_OK; + } + return MEMTX_OK; +} + +static MemTxResult iopmp_block_write(void *opaque, hwaddr addr, uint64_t value, + unsigned size, MemTxAttrs attrs) +{ + return iopmp_handle_block(opaque, addr, &value, size, attrs, + IOPMP_ACCESS_WRITE); +} + +static MemTxResult iopmp_block_read(void *opaque, hwaddr addr, uint64_t *pdata, + unsigned size, MemTxAttrs attrs) +{ + return iopmp_handle_block(opaque, addr, pdata, size, attrs, + IOPMP_ACCESS_READ); +} + +static MemTxResult iopmp_block_fetch(void *opaque, hwaddr addr, uint64_t *pdata, + unsigned size, MemTxAttrs attrs) +{ + return iopmp_handle_block(opaque, addr, pdata, size, attrs, + IOPMP_ACCESS_FETCH); +} + +static const MemoryRegionOps iopmp_block_rw_ops = { + .fetch_with_attrs = iopmp_permssion_read, + .read_with_attrs = iopmp_block_read, + .write_with_attrs = iopmp_block_write, + .endianness = DEVICE_NATIVE_ENDIAN, + .valid = {.min_access_size = 1, .max_access_size = 8}, +}; + +static const MemoryRegionOps iopmp_block_w_ops = { + .fetch_with_attrs = iopmp_permssion_read, + .read_with_attrs = iopmp_permssion_read, + .write_with_attrs = iopmp_block_write, + .endianness = DEVICE_NATIVE_ENDIAN, + .valid = {.min_access_size = 1, .max_access_size = 8}, +}; + +static const MemoryRegionOps iopmp_block_r_ops = { + .fetch_with_attrs = iopmp_permssion_read, + .read_with_attrs = iopmp_block_read, + .write_with_attrs = iopmp_permssion_write, + .endianness = DEVICE_NATIVE_ENDIAN, + .valid = {.min_access_size = 1, .max_access_size = 8}, +}; + +static const MemoryRegionOps iopmp_block_rwx_ops = { + .fetch_with_attrs = iopmp_block_fetch, + .read_with_attrs = iopmp_block_read, + .write_with_attrs = iopmp_block_write, + .endianness = DEVICE_NATIVE_ENDIAN, + .valid = {.min_access_size = 1, .max_access_size = 8}, +}; + +static const MemoryRegionOps iopmp_block_wx_ops = { + .fetch_with_attrs = iopmp_block_fetch, + .read_with_attrs = iopmp_permssion_read, + .write_with_attrs = iopmp_block_write, + .endianness = DEVICE_NATIVE_ENDIAN, + .valid = {.min_access_size = 1, .max_access_size = 8}, +}; + +static const MemoryRegionOps iopmp_block_rx_ops = { + .fetch_with_attrs = iopmp_block_fetch, + .read_with_attrs = iopmp_block_read, + .write_with_attrs = iopmp_permssion_write, + .endianness = DEVICE_NATIVE_ENDIAN, + .valid = {.min_access_size = 1, .max_access_size = 8}, +}; + +static const MemoryRegionOps iopmp_block_x_ops = { + .fetch_with_attrs = iopmp_block_fetch, + .read_with_attrs = iopmp_permssion_read, + .write_with_attrs = iopmp_permssion_write, + .endianness = DEVICE_NATIVE_ENDIAN, + .valid = {.min_access_size = 1, .max_access_size = 8}, +}; + +static void iopmp_realize(DeviceState *dev, Error **errp) +{ + Object *obj = OBJECT(dev); + SysBusDevice *sbd = SYS_BUS_DEVICE(dev); + IopmpState *s = IOPMP(dev); + uint64_t size; + + size = -1ULL; + s->model = IOPMP_MODEL_RAPIDK; + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, s->md_num); + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, 1); + + s->prient_prog = s->default_prient_prog; + s->rrid_num = MIN(s->rrid_num, IOPMP_MAX_RRID_NUM); + s->md_num = MIN(s->md_num, IOPMP_MAX_MD_NUM); + s->entry_num = s->md_num * s->k; + s->prio_entry = MIN(s->prio_entry, s->entry_num); + + s->regs.mdcfg = g_malloc0(s->md_num * sizeof(uint32_t)); + s->regs.mdcfg[0] = s->k; + + s->regs.srcmd_en = g_malloc0(s->rrid_num * sizeof(uint32_t)); + s->regs.srcmd_enh = g_malloc0(s->rrid_num * sizeof(uint32_t)); + s->regs.entry = g_malloc0(s->entry_num * sizeof(iopmp_entry_t)); + s->entry_addr = g_malloc0(s->entry_num * sizeof(iopmp_addr_t)); + s->transaction_state = g_malloc0(s->rrid_num * + sizeof(iopmp_transaction_state)); + qemu_mutex_init(&s->iopmp_transaction_mutex); + + memory_region_init_iommu(&s->iommu, sizeof(s->iommu), + TYPE_IOPMP_IOMMU_MEMORY_REGION, + obj, "riscv-iopmp-sysbus-iommu", UINT64_MAX); + memory_region_init_io(&s->mmio, obj, &iopmp_ops, + s, "iopmp-regs", 0x100000); + sysbus_init_mmio(sbd, &s->mmio); + + memory_region_init_io(&s->blocked_rw, NULL, &iopmp_block_rw_ops, + s, "iopmp-blocked-rw", size); + memory_region_init_io(&s->blocked_w, NULL, &iopmp_block_w_ops, + s, "iopmp-blocked-w", size); + memory_region_init_io(&s->blocked_r, NULL, &iopmp_block_r_ops, + s, "iopmp-blocked-r", size); + + memory_region_init_io(&s->blocked_rwx, NULL, &iopmp_block_rwx_ops, + s, "iopmp-blocked-rwx", size); + memory_region_init_io(&s->blocked_wx, NULL, &iopmp_block_wx_ops, + s, "iopmp-blocked-wx", size); + memory_region_init_io(&s->blocked_rx, NULL, &iopmp_block_rx_ops, + s, "iopmp-blocked-rx", size); + memory_region_init_io(&s->blocked_x, NULL, &iopmp_block_x_ops, + s, "iopmp-blocked-x", size); + address_space_init(&s->blocked_rw_as, &s->blocked_rw, + "iopmp-blocked-rw-as"); + address_space_init(&s->blocked_w_as, &s->blocked_w, + "iopmp-blocked-w-as"); + address_space_init(&s->blocked_r_as, &s->blocked_r, + "iopmp-blocked-r-as"); + + address_space_init(&s->blocked_rwx_as, &s->blocked_rwx, + "iopmp-blocked-rwx-as"); + address_space_init(&s->blocked_wx_as, &s->blocked_wx, + "iopmp-blocked-wx-as"); + address_space_init(&s->blocked_rx_as, &s->blocked_rx, + "iopmp-blocked-rx-as"); + address_space_init(&s->blocked_x_as, &s->blocked_x, + "iopmp-blocked-x-as"); +} + +static void iopmp_reset(DeviceState *dev) +{ + IopmpState *s = IOPMP(dev); + + qemu_set_irq(s->irq, 0); + memset(s->regs.srcmd_en, 0, s->rrid_num * sizeof(uint32_t)); + memset(s->regs.srcmd_enh, 0, s->rrid_num * sizeof(uint32_t)); + memset(s->entry_addr, 0, s->entry_num * sizeof(iopmp_addr_t)); + + s->regs.mdlck = 0; + s->regs.mdlckh = 0; + s->regs.entrylck = 0; + s->regs.mdstall = 0; + s->regs.mdstallh = 0; + s->regs.rridscp = 0; + s->regs.err_cfg = 0; + s->regs.err_reqaddr = 0; + s->regs.err_reqid = 0; + s->regs.err_reqinfo = 0; + + s->prient_prog = s->default_prient_prog; + s->enable = 0; + + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, F, s->md_num); + s->regs.mdcfglck = FIELD_DP32(s->regs.mdcfglck, MDCFGLCK, L, 1); + s->regs.mdcfg[0] = s->k; +} + +static int iopmp_attrs_to_index(IOMMUMemoryRegion *iommu, MemTxAttrs attrs) +{ + return attrs.requester_id; +} + +static void iopmp_iommu_memory_region_class_init(ObjectClass *klass, void *data) +{ + IOMMUMemoryRegionClass *imrc = IOMMU_MEMORY_REGION_CLASS(klass); + + imrc->translate = iopmp_translate; + imrc->attrs_to_index = iopmp_attrs_to_index; +} + +static Property iopmp_property[] = { + DEFINE_PROP_BOOL("prient_prog", IopmpState, default_prient_prog, true), + DEFINE_PROP_UINT32("k", IopmpState, k, 6), + DEFINE_PROP_UINT32("prio_entry", IopmpState, prio_entry, 48), + DEFINE_PROP_UINT32("rrid_num", IopmpState, rrid_num, 16), + DEFINE_PROP_UINT32("md_num", IopmpState, md_num, 8), + DEFINE_PROP_UINT32("entry_offset", IopmpState, entry_offset, 0x4000), + DEFINE_PROP_UINT32("fabricated_v", IopmpState, fabricated_v, 0x0), + DEFINE_PROP_END_OF_LIST(), +}; + +static void iopmp_class_init(ObjectClass *klass, void *data) +{ + DeviceClass *dc = DEVICE_CLASS(klass); + device_class_set_props(dc, iopmp_property); + dc->realize = iopmp_realize; + dc->reset = iopmp_reset; +} + +static void iopmp_init(Object *obj) +{ + IopmpState *s = IOPMP(obj); + SysBusDevice *sbd = SYS_BUS_DEVICE(obj); + + sysbus_init_irq(sbd, &s->irq); +} + +static const TypeInfo iopmp_info = { + .name = TYPE_IOPMP, + .parent = TYPE_SYS_BUS_DEVICE, + .instance_size = sizeof(IopmpState), + .instance_init = iopmp_init, + .class_init = iopmp_class_init, +}; + +static const TypeInfo +iopmp_iommu_memory_region_info = { + .name = TYPE_IOPMP_IOMMU_MEMORY_REGION, + .parent = TYPE_IOMMU_MEMORY_REGION, + .class_init = iopmp_iommu_memory_region_class_init, +}; + +static void +iopmp_register_types(void) +{ + type_register_static(&iopmp_info); + type_register_static(&iopmp_iommu_memory_region_info); +} + +type_init(iopmp_register_types); diff --git a/hw/misc/trace-events b/hw/misc/trace-events index 1be0717c0c..c148166d2d 100644 --- a/hw/misc/trace-events +++ b/hw/misc/trace-events @@ -362,3 +362,6 @@ aspeed_sli_read(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx aspeed_sliio_write(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx64 " of size %u: 0x%" PRIx32 aspeed_sliio_read(uint64_t offset, unsigned int size, uint32_t data) "To 0x%" PRIx64 " of size %u: 0x%" PRIx32 +# riscv_iopmp.c +iopmp_read(uint64_t addr, uint32_t val) "addr 0x%"PRIx64" val 0x%x" +iopmp_write(uint64_t addr, uint32_t val) "addr 0x%"PRIx64" val 0x%x" diff --git a/include/hw/misc/riscv_iopmp.h b/include/hw/misc/riscv_iopmp.h new file mode 100644 index 0000000000..b8fe479108 --- /dev/null +++ b/include/hw/misc/riscv_iopmp.h @@ -0,0 +1,168 @@ +/* + * QEMU RISC-V IOPMP (Input Output Physical Memory Protection) + * + * Copyright (c) 2023-2024 Andes Tech. Corp. + * + * SPDX-License-Identifier: GPL-2.0-or-later + * + * This program is free software; you can redistribute it and/or modify it + * under the terms and conditions of the GNU General Public License, + * version 2 or later, as published by the Free Software Foundation. + * + * This program is distributed in the hope it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for + * more details. + * + * You should have received a copy of the GNU General Public License along with + * this program. If not, see <http://www.gnu.org/licenses/>. + */ + +#ifndef RISCV_IOPMP_H +#define RISCV_IOPMP_H + +#include "hw/sysbus.h" +#include "qemu/typedefs.h" +#include "memory.h" +#include "exec/hwaddr.h" + +#define TYPE_IOPMP "iopmp" +#define IOPMP(obj) OBJECT_CHECK(IopmpState, (obj), TYPE_IOPMP) + +#define IOPMP_MAX_MD_NUM 63 +#define IOPMP_MAX_RRID_NUM 65535 +#define IOPMP_MAX_ENTRY_NUM 65535 + +#define VENDER_VIRT 0 +#define SPECVER_0_9_1 91 +#define IMPID_0_9_1 91 + +#define RRE_ERROR 0 +#define RRE_SUCCESS_VALUE 1 + +#define RWE_ERROR 0 +#define RWE_SUCCESS 1 + +#define RXE_ERROR 0 +#define RXE_SUCCESS_VALUE 1 + +#define ERR_REQINFO_TTYPE_READ 1 +#define ERR_REQINFO_TTYPE_WRITE 2 +#define ERR_REQINFO_TTYPE_FETCH 3 +#define ERR_REQINFO_ETYPE_NOERROR 0 +#define ERR_REQINFO_ETYPE_READ 1 +#define ERR_REQINFO_ETYPE_WRITE 2 +#define ERR_REQINFO_ETYPE_FETCH 3 +#define ERR_REQINFO_ETYPE_PARHIT 4 +#define ERR_REQINFO_ETYPE_NOHIT 5 +#define ERR_REQINFO_ETYPE_RRID 6 +#define ERR_REQINFO_ETYPE_USER 7 + +#define IOPMP_MODEL_FULL 0 +#define IOPMP_MODEL_RAPIDK 0x1 +#define IOPMP_MODEL_DYNAMICK 0x2 +#define IOPMP_MODEL_ISOLATION 0x3 +#define IOPMP_MODEL_COMPACTK 0x4 + +#define ENTRY_NO_HIT 0 +#define ENTRY_PAR_HIT 1 +#define ENTRY_HIT 2 + +/* The generic iopmp address space which downstream is system memory */ +extern AddressSpace iopmp_container_as; + +typedef enum { + IOPMP_AMATCH_OFF, /* Null (off) */ + IOPMP_AMATCH_TOR, /* Top of Range */ + IOPMP_AMATCH_NA4, /* Naturally aligned four-byte region */ + IOPMP_AMATCH_NAPOT /* Naturally aligned power-of-two region */ +} iopmp_am_t; + +typedef enum { + IOPMP_ACCESS_READ = 1, + IOPMP_ACCESS_WRITE = 2, + IOPMP_ACCESS_FETCH = 3 +} iopmp_access_type; + +typedef enum { + IOPMP_NONE = 0, + IOPMP_RO = 1, + IOPMP_WO = 2, + IOPMP_RW = 3, + IOPMP_XO = 4, + IOPMP_RX = 5, + IOPMP_WX = 6, + IOPMP_RWX = 7, +} iopmp_permission; + +typedef struct { + uint32_t addr_reg; + uint32_t addrh_reg; + uint32_t cfg_reg; +} iopmp_entry_t; + +typedef struct { + uint64_t sa; + uint64_t ea; +} iopmp_addr_t; + +typedef struct { + uint32_t *srcmd_en; + uint32_t *srcmd_enh; + uint32_t *mdcfg; + iopmp_entry_t *entry; + uint32_t mdlck; + uint32_t mdlckh; + uint32_t entrylck; + uint32_t mdcfglck; + uint32_t mdstall; + uint32_t mdstallh; + uint32_t rridscp; + uint32_t err_cfg; + uint64_t err_reqaddr; + uint32_t err_reqid; + uint32_t err_reqinfo; +} iopmp_regs; + + +/* To detect partially hit */ +typedef struct iopmp_transaction_state { + bool running; + bool supported; + hwaddr start_addr; + hwaddr end_addr; +} iopmp_transaction_state; + +typedef struct IopmpState { + SysBusDevice parent_obj; + iopmp_addr_t *entry_addr; + MemoryRegion mmio; + IOMMUMemoryRegion iommu; + IOMMUMemoryRegion *next_iommu; + iopmp_regs regs; + MemoryRegion *downstream; + MemoryRegion blocked_r, blocked_w, blocked_x, blocked_rw, blocked_rx, + blocked_wx, blocked_rwx; + MemoryRegion stall_io; + uint32_t model; + uint32_t k; + bool prient_prog; + bool default_prient_prog; + iopmp_transaction_state *transaction_state; + QemuMutex iopmp_transaction_mutex; + + AddressSpace downstream_as; + AddressSpace blocked_r_as, blocked_w_as, blocked_x_as, blocked_rw_as, + blocked_rx_as, blocked_wx_as, blocked_rwx_as; + qemu_irq irq; + bool enable; + + uint32_t prio_entry; + uint32_t rrid_num; + uint32_t md_num; + uint32_t entry_num; + uint32_t entry_offset; + uint32_t fabricated_v; +} IopmpState; + +#endif
Support basic functions of IOPMP specification v0.9.1 rapid-k model. The specification url: https://github.com/riscv-non-isa/iopmp-spec/releases/tag/v0.9.1 The IOPMP checks whether memory access from a device or CPU is valid. This implementation uses an IOMMU to modify the address space accessed by the device. For device access with IOMMUAccessFlags specifying read or write (IOMMU_RO or IOMMU_WO), the IOPMP checks the permission in iopmp_translate. If the access is valid, the target address space is downstream_as. If the access is blocked, it will be redirected to blocked_rwx_as. For CPU access with IOMMUAccessFlags not specifying read or write (IOMMU_NONE), the IOPMP translates the access to the corresponding address space based on the permission. If the access has full permission (rwx), the target address space is downstream_as. If the access has limited permission, the target address space is blocked_ followed by the lacked permissions. The operation of a blocked region can trigger an IOPMP interrupt, a bus error, or it can respond with success and fabricated data, depending on the value of the IOPMP ERR_CFG register. Signed-off-by: Ethan Chen <ethan84@andestech.com> --- hw/misc/Kconfig | 3 + hw/misc/meson.build | 1 + hw/misc/riscv_iopmp.c | 1154 +++++++++++++++++++++++++++++++++ hw/misc/trace-events | 3 + include/hw/misc/riscv_iopmp.h | 168 +++++ 5 files changed, 1329 insertions(+) create mode 100644 hw/misc/riscv_iopmp.c create mode 100644 include/hw/misc/riscv_iopmp.h