Message ID | 20190619094907.10131-1-pagupta@redhat.com (mailing list archive) |
---|---|
Headers | show |
Series | Qemu virtio pmem device | expand |
Ping. > > This patch series has implementation for "virtio pmem" > device. "virtio pmem" is persistent memory(nvdimm) device in > guest which allows to bypass the guest page cache. This > also implements a VIRTIO based asynchronous flush mechanism. > Details of project idea for 'virtio pmem' flushing interface > is shared [2] & [3]. > > Sharing Qemu device emulation in this patchset. Tested with > guest kernel driver [1]. This series is based on David's > memory device refactoring [5] work with modified version of > my initial virtio pmem [4] series. > > Usage: > ./qemu -name test -machine pc -m 8G,slots=240,maxmem=20G > -object memory-backend-file,id=mem1,share,mem-path=test.img, > size=4G,share > -device virtio-pmem-pci,memdev=mem1,id=nv1 > > (qemu) info memory-devices > Memory device [virtio-pmem]: "nv1" > memaddr: 0x240000000 > size: 4294967296 > memdev: /objects/mem1 > > Implementation is divided into two parts: > New virtio pmem guest driver and qemu code changes for new > virtio pmem paravirtualized device. In this series we are > sharing Qemu device emulation. > > 1. Guest virtio-pmem kernel driver > --------------------------------- > - Reads persistent memory range from paravirt device and > registers with 'nvdimm_bus'. > - 'nvdimm/pmem' driver uses this information to allocate > persistent memory region and setup filesystem operations > to the allocated memory. > - virtio pmem driver implements asynchronous flushing > interface to flush from guest to host. > > 2. Qemu virtio-pmem device > --------------------------------- > - Creates virtio pmem device and exposes a memory range to > KVM guest. > - At host side this is file backed memory which acts as > persistent memory. > - Qemu side flush uses aio thread pool API's and virtio > for asynchronous guest multi request handling. > > Virtio-pmem security implications and suggested countermeasures: > --------------------------------------------------------------- > > In previous posting of kernel driver, there was discussion [7] > on possible implications of page cache side channel attacks with > virtio pmem. After thorough analysis of details of known side > channel attacks, below are the suggestions: > > - Depends entirely on how host backing image file is mapped > into guest address space. > > - virtio-pmem device emulation, by default shared mapping is used > to map host backing file. It is recommended to use separate > backing file at host side for every guest. This will prevent > any possibility of executing common code from multiple guests > and any chance of inferring guest local data based based on > execution time. > > - If backing file is required to be shared among multiple guests > it is recommended to don't support host page cache eviction > commands from the guest driver. This will avoid any possibility > of inferring guest local data or host data from another guest. > > - Proposed device specification [6] for virtio-pmem device with > details of possible security implications and suggested > countermeasures for device emulation. > > Changes from PATCH v1: > - Change proposed version from qemu 4.0 to 4.1 - Eric > - Remove virtio queue_add from unrealize function - Cornelia > > [1] https://lkml.org/lkml/2019/6/12/624 > [2] https://www.spinics.net/lists/kvm/msg149761.html > [3] https://www.spinics.net/lists/kvm/msg153095.html > [4] https://marc.info/?l=linux-kernel&m=153572228719237&w=2 > [5] https://marc.info/?l=qemu-devel&m=153555721901824&w=2 > [6] https://lists.oasis-open.org/archives/virtio-dev/201903/msg00083.html > [7] https://lkml.org/lkml/2019/1/9/1191 > > Pankaj Gupta (3): > virtio-pmem: add virtio device > virtio-pmem: sync linux headers > virtio-pci: proxy for virtio-pmem > > David Hildenbrand (4): > virtio-pci: Allow to specify additional interfaces for the base type > hmp: Handle virtio-pmem when printing memory device infos > numa: Handle virtio-pmem in NUMA stats > pc: Support for virtio-pmem-pci > > hmp.c | 27 ++- > hw/i386/Kconfig | 1 > hw/i386/pc.c | 72 ++++++++++ > hw/virtio/Kconfig | 10 + > hw/virtio/Makefile.objs | 2 > hw/virtio/virtio-pci.c | 1 > hw/virtio/virtio-pci.h | 1 > hw/virtio/virtio-pmem-pci.c | 131 ++++++++++++++++++ > hw/virtio/virtio-pmem-pci.h | 34 ++++ > hw/virtio/virtio-pmem.c | 189 > +++++++++++++++++++++++++++ > include/hw/pci/pci.h | 1 > include/hw/virtio/virtio-pmem.h | 49 +++++++ > include/standard-headers/linux/virtio_ids.h | 1 > include/standard-headers/linux/virtio_pmem.h | 35 +++++ > numa.c | 24 +-- > qapi/misc.json | 28 +++- > 16 files changed, 580 insertions(+), 26 deletions(-) > ---- > >
On Wed, Jun 19, 2019 at 03:19:00PM +0530, Pankaj Gupta wrote: > This patch series has implementation for "virtio pmem" > device. "virtio pmem" is persistent memory(nvdimm) device in > guest which allows to bypass the guest page cache. This > also implements a VIRTIO based asynchronous flush mechanism. > Details of project idea for 'virtio pmem' flushing interface > is shared [2] & [3]. > > Sharing Qemu device emulation in this patchset. Tested with > guest kernel driver [1]. This series is based on David's > memory device refactoring [5] work with modified version of > my initial virtio pmem [4] series. > > Usage: > ./qemu -name test -machine pc -m 8G,slots=240,maxmem=20G > -object memory-backend-file,id=mem1,share,mem-path=test.img, > size=4G,share > -device virtio-pmem-pci,memdev=mem1,id=nv1 > Hi, Pankaj I tried this series with v14 kernel driver, while getting some error on using this. Not sure this is my error configuration. The qemu command line is: -object memory-backend-file,id=mem1,share=on,mem-path=/dev/dax0.0,size=1G,align=2M -device virtio-pmem-pci,memdev=mem1,id=nv1 The guest boots up and I can see /dev/pmem0 device. But when I want to partition this device, I got the error: # parted /dev/pmem0 mklabel gpt Warning: Error fsyncing/closing /dev/pmem0: Input/output error Also I see an error when running "ndctl list": libndctl: __sysfs_device_parse: ndctl0: add_dev() failed Would you mind letting me know which part I am wrong? > (qemu) info memory-devices > Memory device [virtio-pmem]: "nv1" > memaddr: 0x240000000 > size: 4294967296 > memdev: /objects/mem1 > > Implementation is divided into two parts: > New virtio pmem guest driver and qemu code changes for new > virtio pmem paravirtualized device. In this series we are > sharing Qemu device emulation. > >1. Guest virtio-pmem kernel driver >--------------------------------- > - Reads persistent memory range from paravirt device and > registers with 'nvdimm_bus'. > - 'nvdimm/pmem' driver uses this information to allocate > persistent memory region and setup filesystem operations > to the allocated memory. > - virtio pmem driver implements asynchronous flushing > interface to flush from guest to host. > >2. Qemu virtio-pmem device >--------------------------------- > - Creates virtio pmem device and exposes a memory range to > KVM guest. > - At host side this is file backed memory which acts as > persistent memory. > - Qemu side flush uses aio thread pool API's and virtio > for asynchronous guest multi request handling. > > Virtio-pmem security implications and suggested countermeasures: > --------------------------------------------------------------- > > In previous posting of kernel driver, there was discussion [7] > on possible implications of page cache side channel attacks with > virtio pmem. After thorough analysis of details of known side > channel attacks, below are the suggestions: > > - Depends entirely on how host backing image file is mapped > into guest address space. > > - virtio-pmem device emulation, by default shared mapping is used > to map host backing file. It is recommended to use separate > backing file at host side for every guest. This will prevent > any possibility of executing common code from multiple guests > and any chance of inferring guest local data based based on > execution time. > > - If backing file is required to be shared among multiple guests > it is recommended to don't support host page cache eviction > commands from the guest driver. This will avoid any possibility > of inferring guest local data or host data from another guest. > > - Proposed device specification [6] for virtio-pmem device with > details of possible security implications and suggested > countermeasures for device emulation. > >Changes from PATCH v1: > - Change proposed version from qemu 4.0 to 4.1 - Eric > - Remove virtio queue_add from unrealize function - Cornelia > >[1] https://lkml.org/lkml/2019/6/12/624 >[2] https://www.spinics.net/lists/kvm/msg149761.html >[3] https://www.spinics.net/lists/kvm/msg153095.html >[4] https://marc.info/?l=linux-kernel&m=153572228719237&w=2 >[5] https://marc.info/?l=qemu-devel&m=153555721901824&w=2 >[6] https://lists.oasis-open.org/archives/virtio-dev/201903/msg00083.html >[7] https://lkml.org/lkml/2019/1/9/1191 > > Pankaj Gupta (3): > virtio-pmem: add virtio device > virtio-pmem: sync linux headers > virtio-pci: proxy for virtio-pmem > > David Hildenbrand (4): > virtio-pci: Allow to specify additional interfaces for the base type > hmp: Handle virtio-pmem when printing memory device infos > numa: Handle virtio-pmem in NUMA stats > pc: Support for virtio-pmem-pci > > hmp.c | 27 ++- > hw/i386/Kconfig | 1 > hw/i386/pc.c | 72 ++++++++++ > hw/virtio/Kconfig | 10 + > hw/virtio/Makefile.objs | 2 > hw/virtio/virtio-pci.c | 1 > hw/virtio/virtio-pci.h | 1 > hw/virtio/virtio-pmem-pci.c | 131 ++++++++++++++++++ > hw/virtio/virtio-pmem-pci.h | 34 ++++ > hw/virtio/virtio-pmem.c | 189 +++++++++++++++++++++++++++ > include/hw/pci/pci.h | 1 > include/hw/virtio/virtio-pmem.h | 49 +++++++ > include/standard-headers/linux/virtio_ids.h | 1 > include/standard-headers/linux/virtio_pmem.h | 35 +++++ > numa.c | 24 +-- > qapi/misc.json | 28 +++- > 16 files changed, 580 insertions(+), 26 deletions(-) >----
> > On Wed, Jun 19, 2019 at 03:19:00PM +0530, Pankaj Gupta wrote: > > This patch series has implementation for "virtio pmem" > > device. "virtio pmem" is persistent memory(nvdimm) device in > > guest which allows to bypass the guest page cache. This > > also implements a VIRTIO based asynchronous flush mechanism. > > Details of project idea for 'virtio pmem' flushing interface > > is shared [2] & [3]. > > > > Sharing Qemu device emulation in this patchset. Tested with > > guest kernel driver [1]. This series is based on David's > > memory device refactoring [5] work with modified version of > > my initial virtio pmem [4] series. > > > > Usage: > > ./qemu -name test -machine pc -m 8G,slots=240,maxmem=20G > > -object memory-backend-file,id=mem1,share,mem-path=test.img, > > size=4G,share > > -device virtio-pmem-pci,memdev=mem1,id=nv1 > > > > Hi, Pankaj Hi Wei, > > I tried this series with v14 kernel driver, while getting some error on using > this. Not sure this is my error configuration. > > The qemu command line is: > > -object > memory-backend-file,id=mem1,share=on,mem-path=/dev/dax0.0,size=1G,align=2M > -device virtio-pmem-pci,memdev=mem1,id=nv1 Are you using host backing on real NVDIMM? Currently, we only support backing image file on regular SSD. We have plans to support backing file on real NVDIMM device as well but this is part of future work. Can you please try by allocating a raw image file on regular SSD. It is working fine for me. > > The guest boots up and I can see /dev/pmem0 device. But when I want to > partition this device, I got the error: > > # parted /dev/pmem0 mklabel gpt > Warning: Error fsyncing/closing /dev/pmem0: Input/output error > > Also I see an error when running "ndctl list": > > libndctl: __sysfs_device_parse: ndctl0: add_dev() failed Will look at this if it is related. Thanks, Pankaj > > Would you mind letting me know which part I am wrong? > > > (qemu) info memory-devices > > Memory device [virtio-pmem]: "nv1" > > memaddr: 0x240000000 > > size: 4294967296 > > memdev: /objects/mem1 > > > > Implementation is divided into two parts: > > New virtio pmem guest driver and qemu code changes for new > > virtio pmem paravirtualized device. In this series we are > > sharing Qemu device emulation. > > > >1. Guest virtio-pmem kernel driver > >--------------------------------- > > - Reads persistent memory range from paravirt device and > > registers with 'nvdimm_bus'. > > - 'nvdimm/pmem' driver uses this information to allocate > > persistent memory region and setup filesystem operations > > to the allocated memory. > > - virtio pmem driver implements asynchronous flushing > > interface to flush from guest to host. > > > >2. Qemu virtio-pmem device > >--------------------------------- > > - Creates virtio pmem device and exposes a memory range to > > KVM guest. > > - At host side this is file backed memory which acts as > > persistent memory. > > - Qemu side flush uses aio thread pool API's and virtio > > for asynchronous guest multi request handling. > > > > Virtio-pmem security implications and suggested countermeasures: > > --------------------------------------------------------------- > > > > In previous posting of kernel driver, there was discussion [7] > > on possible implications of page cache side channel attacks with > > virtio pmem. After thorough analysis of details of known side > > channel attacks, below are the suggestions: > > > > - Depends entirely on how host backing image file is mapped > > into guest address space. > > > > - virtio-pmem device emulation, by default shared mapping is used > > to map host backing file. It is recommended to use separate > > backing file at host side for every guest. This will prevent > > any possibility of executing common code from multiple guests > > and any chance of inferring guest local data based based on > > execution time. > > > > - If backing file is required to be shared among multiple guests > > it is recommended to don't support host page cache eviction > > commands from the guest driver. This will avoid any possibility > > of inferring guest local data or host data from another guest. > > > > - Proposed device specification [6] for virtio-pmem device with > > details of possible security implications and suggested > > countermeasures for device emulation. > > > >Changes from PATCH v1: > > - Change proposed version from qemu 4.0 to 4.1 - Eric > > - Remove virtio queue_add from unrealize function - Cornelia > > > >[1] https://lkml.org/lkml/2019/6/12/624 > >[2] https://www.spinics.net/lists/kvm/msg149761.html > >[3] https://www.spinics.net/lists/kvm/msg153095.html > >[4] https://marc.info/?l=linux-kernel&m=153572228719237&w=2 > >[5] https://marc.info/?l=qemu-devel&m=153555721901824&w=2 > >[6] https://lists.oasis-open.org/archives/virtio-dev/201903/msg00083.html > >[7] https://lkml.org/lkml/2019/1/9/1191 > > > > Pankaj Gupta (3): > > virtio-pmem: add virtio device > > virtio-pmem: sync linux headers > > virtio-pci: proxy for virtio-pmem > > > > David Hildenbrand (4): > > virtio-pci: Allow to specify additional interfaces for the base type > > hmp: Handle virtio-pmem when printing memory device infos > > numa: Handle virtio-pmem in NUMA stats > > pc: Support for virtio-pmem-pci > > > > hmp.c | 27 ++- > > hw/i386/Kconfig | 1 > > hw/i386/pc.c | 72 ++++++++++ > > hw/virtio/Kconfig | 10 + > > hw/virtio/Makefile.objs | 2 > > hw/virtio/virtio-pci.c | 1 > > hw/virtio/virtio-pci.h | 1 > > hw/virtio/virtio-pmem-pci.c | 131 ++++++++++++++++++ > > hw/virtio/virtio-pmem-pci.h | 34 ++++ > > hw/virtio/virtio-pmem.c | 189 > > +++++++++++++++++++++++++++ > > include/hw/pci/pci.h | 1 > > include/hw/virtio/virtio-pmem.h | 49 +++++++ > > include/standard-headers/linux/virtio_ids.h | 1 > > include/standard-headers/linux/virtio_pmem.h | 35 +++++ > > numa.c | 24 +-- > > qapi/misc.json | 28 +++- > > 16 files changed, 580 insertions(+), 26 deletions(-) > >---- > > -- > Wei Yang > Help you, Help me > >
On Tue, Jul 02, 2019 at 06:07:44AM -0400, Pankaj Gupta wrote: > >> >> On Wed, Jun 19, 2019 at 03:19:00PM +0530, Pankaj Gupta wrote: >> > This patch series has implementation for "virtio pmem" >> > device. "virtio pmem" is persistent memory(nvdimm) device in >> > guest which allows to bypass the guest page cache. This >> > also implements a VIRTIO based asynchronous flush mechanism. >> > Details of project idea for 'virtio pmem' flushing interface >> > is shared [2] & [3]. >> > >> > Sharing Qemu device emulation in this patchset. Tested with >> > guest kernel driver [1]. This series is based on David's >> > memory device refactoring [5] work with modified version of >> > my initial virtio pmem [4] series. >> > >> > Usage: >> > ./qemu -name test -machine pc -m 8G,slots=240,maxmem=20G >> > -object memory-backend-file,id=mem1,share,mem-path=test.img, >> > size=4G,share >> > -device virtio-pmem-pci,memdev=mem1,id=nv1 >> > >> >> Hi, Pankaj > >Hi Wei, > >> >> I tried this series with v14 kernel driver, while getting some error on using >> this. Not sure this is my error configuration. >> >> The qemu command line is: >> >> -object >> memory-backend-file,id=mem1,share=on,mem-path=/dev/dax0.0,size=1G,align=2M >> -device virtio-pmem-pci,memdev=mem1,id=nv1 > >Are you using host backing on real NVDIMM? Currently, we only support backing image >file on regular SSD. We have plans to support backing file on real NVDIMM device >as well but this is part of future work. > >Can you please try by allocating a raw image file on regular SSD. It is working fine >for me. > I created a file with 2G on my disk. #ll -h 2G-file -rw-r--r-- 1 richard richard 2.0G 6月 26 09:26 2G-file The command line is changed to: -object memory-backend-file,id=mem1,share,mem-path=/home/richard/guest/2G-file,size=2G -device virtio-pmem-pci,memdev=mem1,id=nv1 The behavior in guest is the same. I took a look into the directory /sys/bus/nd/device. These files are listed. Compared with normal system, one device file is missed. btt0.0 dax0.0 namespace0.0 ndbus0 pfn0.0 region0 But the sysfs shows pmem0 block device is created. /sys/devices/pci0000:00/0000:00:04.0/virtio0/ndbus0/region0/namespace0.0/block/pmem0 Then I took a look into the pci device: # lspci -vs 00:04.0 00:04.0 Unclassified device [00ff]: Red Hat, Inc. Device 1013 Subsystem: Red Hat, Inc. Device 001b Physical Slot: 4 Flags: bus master, fast devsel, latency 0, IRQ 11 I/O ports at c040 [size=64] Memory at fe000000 (64-bit, prefetchable) [size=16K] Capabilities: [84] Vendor Specific Information: VirtIO: <unknown> Capabilities: [70] Vendor Specific Information: VirtIO: Notify Capabilities: [60] Vendor Specific Information: VirtIO: DeviceCfg Capabilities: [50] Vendor Specific Information: VirtIO: ISR Capabilities: [40] Vendor Specific Information: VirtIO: CommonCfg Kernel driver in use: virtio-pci This looks good. >> >> The guest boots up and I can see /dev/pmem0 device. But when I want to >> partition this device, I got the error: >> >> # parted /dev/pmem0 mklabel gpt >> Warning: Error fsyncing/closing /dev/pmem0: Input/output error >> >> Also I see an error when running "ndctl list": >> >> libndctl: __sysfs_device_parse: ndctl0: add_dev() failed > >Will look at this if it is related. This log still there. > >Thanks, >Pankaj >> >> Would you mind letting me know which part I am wrong? >> >> > (qemu) info memory-devices >> > Memory device [virtio-pmem]: "nv1" >> > memaddr: 0x240000000 >> > size: 4294967296 >> > memdev: /objects/mem1 >> > >> > Implementation is divided into two parts: >> > New virtio pmem guest driver and qemu code changes for new >> > virtio pmem paravirtualized device. In this series we are >> > sharing Qemu device emulation. >> > >> >1. Guest virtio-pmem kernel driver >> >--------------------------------- >> > - Reads persistent memory range from paravirt device and >> > registers with 'nvdimm_bus'. >> > - 'nvdimm/pmem' driver uses this information to allocate >> > persistent memory region and setup filesystem operations >> > to the allocated memory. >> > - virtio pmem driver implements asynchronous flushing >> > interface to flush from guest to host. >> > >> >2. Qemu virtio-pmem device >> >--------------------------------- >> > - Creates virtio pmem device and exposes a memory range to >> > KVM guest. >> > - At host side this is file backed memory which acts as >> > persistent memory. >> > - Qemu side flush uses aio thread pool API's and virtio >> > for asynchronous guest multi request handling. >> > >> > Virtio-pmem security implications and suggested countermeasures: >> > --------------------------------------------------------------- >> > >> > In previous posting of kernel driver, there was discussion [7] >> > on possible implications of page cache side channel attacks with >> > virtio pmem. After thorough analysis of details of known side >> > channel attacks, below are the suggestions: >> > >> > - Depends entirely on how host backing image file is mapped >> > into guest address space. >> > >> > - virtio-pmem device emulation, by default shared mapping is used >> > to map host backing file. It is recommended to use separate >> > backing file at host side for every guest. This will prevent >> > any possibility of executing common code from multiple guests >> > and any chance of inferring guest local data based based on >> > execution time. >> > >> > - If backing file is required to be shared among multiple guests >> > it is recommended to don't support host page cache eviction >> > commands from the guest driver. This will avoid any possibility >> > of inferring guest local data or host data from another guest. >> > >> > - Proposed device specification [6] for virtio-pmem device with >> > details of possible security implications and suggested >> > countermeasures for device emulation. >> > >> >Changes from PATCH v1: >> > - Change proposed version from qemu 4.0 to 4.1 - Eric >> > - Remove virtio queue_add from unrealize function - Cornelia >> > >> >[1] https://lkml.org/lkml/2019/6/12/624 >> >[2] https://www.spinics.net/lists/kvm/msg149761.html >> >[3] https://www.spinics.net/lists/kvm/msg153095.html >> >[4] https://marc.info/?l=linux-kernel&m=153572228719237&w=2 >> >[5] https://marc.info/?l=qemu-devel&m=153555721901824&w=2 >> >[6] https://lists.oasis-open.org/archives/virtio-dev/201903/msg00083.html >> >[7] https://lkml.org/lkml/2019/1/9/1191 >> > >> > Pankaj Gupta (3): >> > virtio-pmem: add virtio device >> > virtio-pmem: sync linux headers >> > virtio-pci: proxy for virtio-pmem >> > >> > David Hildenbrand (4): >> > virtio-pci: Allow to specify additional interfaces for the base type >> > hmp: Handle virtio-pmem when printing memory device infos >> > numa: Handle virtio-pmem in NUMA stats >> > pc: Support for virtio-pmem-pci >> > >> > hmp.c | 27 ++- >> > hw/i386/Kconfig | 1 >> > hw/i386/pc.c | 72 ++++++++++ >> > hw/virtio/Kconfig | 10 + >> > hw/virtio/Makefile.objs | 2 >> > hw/virtio/virtio-pci.c | 1 >> > hw/virtio/virtio-pci.h | 1 >> > hw/virtio/virtio-pmem-pci.c | 131 ++++++++++++++++++ >> > hw/virtio/virtio-pmem-pci.h | 34 ++++ >> > hw/virtio/virtio-pmem.c | 189 >> > +++++++++++++++++++++++++++ >> > include/hw/pci/pci.h | 1 >> > include/hw/virtio/virtio-pmem.h | 49 +++++++ >> > include/standard-headers/linux/virtio_ids.h | 1 >> > include/standard-headers/linux/virtio_pmem.h | 35 +++++ >> > numa.c | 24 +-- >> > qapi/misc.json | 28 +++- >> > 16 files changed, 580 insertions(+), 26 deletions(-) >> >---- >> >> -- >> Wei Yang >> Help you, Help me >> >>
> > > >> > >> On Wed, Jun 19, 2019 at 03:19:00PM +0530, Pankaj Gupta wrote: > >> > This patch series has implementation for "virtio pmem" > >> > device. "virtio pmem" is persistent memory(nvdimm) device in > >> > guest which allows to bypass the guest page cache. This > >> > also implements a VIRTIO based asynchronous flush mechanism. > >> > Details of project idea for 'virtio pmem' flushing interface > >> > is shared [2] & [3]. > >> > > >> > Sharing Qemu device emulation in this patchset. Tested with > >> > guest kernel driver [1]. This series is based on David's > >> > memory device refactoring [5] work with modified version of > >> > my initial virtio pmem [4] series. > >> > > >> > Usage: > >> > ./qemu -name test -machine pc -m 8G,slots=240,maxmem=20G > >> > -object memory-backend-file,id=mem1,share,mem-path=test.img, > >> > size=4G,share > >> > -device virtio-pmem-pci,memdev=mem1,id=nv1 > >> > > >> > >> Hi, Pankaj > > > >Hi Wei, > > > >> > >> I tried this series with v14 kernel driver, while getting some error on > >> using > >> this. Not sure this is my error configuration. > >> > >> The qemu command line is: > >> > >> -object > >> memory-backend-file,id=mem1,share=on,mem-path=/dev/dax0.0,size=1G,align=2M > >> -device virtio-pmem-pci,memdev=mem1,id=nv1 > > > >Are you using host backing on real NVDIMM? Currently, we only support > >backing image > >file on regular SSD. We have plans to support backing file on real NVDIMM > >device > >as well but this is part of future work. > > > >Can you please try by allocating a raw image file on regular SSD. It is > >working fine > >for me. > > > > I created a file with 2G on my disk. > > #ll -h 2G-file > -rw-r--r-- 1 richard richard 2.0G 6月 26 09:26 2G-file > > The command line is changed to: > > -object > memory-backend-file,id=mem1,share,mem-path=/home/richard/guest/2G-file,size=2G > -device virtio-pmem-pci,memdev=mem1,id=nv1 > > The behavior in guest is the same. Are you still facing an error with this? or its working fine for you? > > I took a look into the directory /sys/bus/nd/device. These files are listed. > Compared with normal system, one device file is missed. virtio pmem does not support namespace/region mappings which ACPI NFIT supports. > > btt0.0 dax0.0 namespace0.0 ndbus0 pfn0.0 region0 > > But the sysfs shows pmem0 block device is created. > > /sys/devices/pci0000:00/0000:00:04.0/virtio0/ndbus0/region0/namespace0.0/block/pmem0 > > Then I took a look into the pci device: > > # lspci -vs 00:04.0 > 00:04.0 Unclassified device [00ff]: Red Hat, Inc. Device 1013 > Subsystem: Red Hat, Inc. Device 001b > Physical Slot: 4 > Flags: bus master, fast devsel, latency 0, IRQ 11 > I/O ports at c040 [size=64] > Memory at fe000000 (64-bit, prefetchable) [size=16K] > Capabilities: [84] Vendor Specific Information: VirtIO: > Capabilities: [70] Vendor Specific Information: VirtIO: Notify > Capabilities: [60] Vendor Specific Information: VirtIO: DeviceCfg > Capabilities: [50] Vendor Specific Information: VirtIO: ISR > Capabilities: [40] Vendor Specific Information: VirtIO: CommonCfg > Kernel driver in use: virtio-pci > > This looks good. Good to know. > > >> > >> The guest boots up and I can see /dev/pmem0 device. But when I want to > >> partition this device, I got the error: > >> > >> # parted /dev/pmem0 mklabel gpt > >> Warning: Error fsyncing/closing /dev/pmem0: Input/output error > >> > >> Also I see an error when running "ndctl list": > >> > >> libndctl: __sysfs_device_parse: ndctl0: add_dev() failed > > > >Will look at this if it is related. > > This log still there. It looks to me libndctl needs to be taught about about virtio pmem data parsing. But this is unrelated to kernel and qemu patch series. > > > > >Thanks, > >Pankaj > >> > >> Would you mind letting me know which part I am wrong? > >> > >> > (qemu) info memory-devices > >> > Memory device [virtio-pmem]: "nv1" > >> > memaddr: 0x240000000 > >> > size: 4294967296 > >> > memdev: /objects/mem1 > >> > > >> > Implementation is divided into two parts: > >> > New virtio pmem guest driver and qemu code changes for new > >> > virtio pmem paravirtualized device. In this series we are > >> > sharing Qemu device emulation. > >> > > >> >1. Guest virtio-pmem kernel driver > >> >--------------------------------- > >> > - Reads persistent memory range from paravirt device and > >> > registers with 'nvdimm_bus'. > >> > - 'nvdimm/pmem' driver uses this information to allocate > >> > persistent memory region and setup filesystem operations > >> > to the allocated memory. > >> > - virtio pmem driver implements asynchronous flushing > >> > interface to flush from guest to host. > >> > > >> >2. Qemu virtio-pmem device > >> >--------------------------------- > >> > - Creates virtio pmem device and exposes a memory range to > >> > KVM guest. > >> > - At host side this is file backed memory which acts as > >> > persistent memory. > >> > - Qemu side flush uses aio thread pool API's and virtio > >> > for asynchronous guest multi request handling. > >> > > >> > Virtio-pmem security implications and suggested countermeasures: > >> > --------------------------------------------------------------- > >> > > >> > In previous posting of kernel driver, there was discussion [7] > >> > on possible implications of page cache side channel attacks with > >> > virtio pmem. After thorough analysis of details of known side > >> > channel attacks, below are the suggestions: > >> > > >> > - Depends entirely on how host backing image file is mapped > >> > into guest address space. > >> > > >> > - virtio-pmem device emulation, by default shared mapping is used > >> > to map host backing file. It is recommended to use separate > >> > backing file at host side for every guest. This will prevent > >> > any possibility of executing common code from multiple guests > >> > and any chance of inferring guest local data based based on > >> > execution time. > >> > > >> > - If backing file is required to be shared among multiple guests > >> > it is recommended to don't support host page cache eviction > >> > commands from the guest driver. This will avoid any possibility > >> > of inferring guest local data or host data from another guest. > >> > > >> > - Proposed device specification [6] for virtio-pmem device with > >> > details of possible security implications and suggested > >> > countermeasures for device emulation. > >> > > >> >Changes from PATCH v1: > >> > - Change proposed version from qemu 4.0 to 4.1 - Eric > >> > - Remove virtio queue_add from unrealize function - Cornelia > >> > > >> >[1] https://lkml.org/lkml/2019/6/12/624 > >> >[2] https://www.spinics.net/lists/kvm/msg149761.html > >> >[3] https://www.spinics.net/lists/kvm/msg153095.html > >> >[4] https://marc.info/?l=linux-kernel&m=153572228719237&w=2 > >> >[5] https://marc.info/?l=qemu-devel&m=153555721901824&w=2 > >> >[6] https://lists.oasis-open.org/archives/virtio-dev/201903/msg00083.html > >> >[7] https://lkml.org/lkml/2019/1/9/1191 > >> > > >> > Pankaj Gupta (3): > >> > virtio-pmem: add virtio device > >> > virtio-pmem: sync linux headers > >> > virtio-pci: proxy for virtio-pmem > >> > > >> > David Hildenbrand (4): > >> > virtio-pci: Allow to specify additional interfaces for the base type > >> > hmp: Handle virtio-pmem when printing memory device infos > >> > numa: Handle virtio-pmem in NUMA stats > >> > pc: Support for virtio-pmem-pci > >> > > >> > hmp.c | 27 ++- > >> > hw/i386/Kconfig | 1 > >> > hw/i386/pc.c | 72 ++++++++++ > >> > hw/virtio/Kconfig | 10 + > >> > hw/virtio/Makefile.objs | 2 > >> > hw/virtio/virtio-pci.c | 1 > >> > hw/virtio/virtio-pci.h | 1 > >> > hw/virtio/virtio-pmem-pci.c | 131 ++++++++++++++++++ > >> > hw/virtio/virtio-pmem-pci.h | 34 ++++ > >> > hw/virtio/virtio-pmem.c | 189 > >> > +++++++++++++++++++++++++++ > >> > include/hw/pci/pci.h | 1 > >> > include/hw/virtio/virtio-pmem.h | 49 +++++++ > >> > include/standard-headers/linux/virtio_ids.h | 1 > >> > include/standard-headers/linux/virtio_pmem.h | 35 +++++ > >> > numa.c | 24 +-- > >> > qapi/misc.json | 28 +++- > >> > 16 files changed, 580 insertions(+), 26 deletions(-) > >> >---- > >> > >> -- > >> Wei Yang > >> Help you, Help me > >> > >> > > -- > Wei Yang > Help you, Help me > >
On Tue, Jul 02, 2019 at 09:31:43PM -0400, Pankaj Gupta wrote: > >> > >> >> >> >> On Wed, Jun 19, 2019 at 03:19:00PM +0530, Pankaj Gupta wrote: >> >> > This patch series has implementation for "virtio pmem" >> >> > device. "virtio pmem" is persistent memory(nvdimm) device in >> >> > guest which allows to bypass the guest page cache. This >> >> > also implements a VIRTIO based asynchronous flush mechanism. >> >> > Details of project idea for 'virtio pmem' flushing interface >> >> > is shared [2] & [3]. >> >> > >> >> > Sharing Qemu device emulation in this patchset. Tested with >> >> > guest kernel driver [1]. This series is based on David's >> >> > memory device refactoring [5] work with modified version of >> >> > my initial virtio pmem [4] series. >> >> > >> >> > Usage: >> >> > ./qemu -name test -machine pc -m 8G,slots=240,maxmem=20G >> >> > -object memory-backend-file,id=mem1,share,mem-path=test.img, >> >> > size=4G,share >> >> > -device virtio-pmem-pci,memdev=mem1,id=nv1 >> >> > >> >> >> >> Hi, Pankaj >> > >> >Hi Wei, >> > >> >> >> >> I tried this series with v14 kernel driver, while getting some error on >> >> using >> >> this. Not sure this is my error configuration. >> >> >> >> The qemu command line is: >> >> >> >> -object >> >> memory-backend-file,id=mem1,share=on,mem-path=/dev/dax0.0,size=1G,align=2M >> >> -device virtio-pmem-pci,memdev=mem1,id=nv1 >> > >> >Are you using host backing on real NVDIMM? Currently, we only support >> >backing image >> >file on regular SSD. We have plans to support backing file on real NVDIMM >> >device >> >as well but this is part of future work. >> > >> >Can you please try by allocating a raw image file on regular SSD. It is >> >working fine >> >for me. >> > >> >> I created a file with 2G on my disk. >> >> #ll -h 2G-file >> -rw-r--r-- 1 richard richard 2.0G 6月 26 09:26 2G-file >> >> The command line is changed to: >> >> -object >> memory-backend-file,id=mem1,share,mem-path=/home/richard/guest/2G-file,size=2G >> -device virtio-pmem-pci,memdev=mem1,id=nv1 >> >> The behavior in guest is the same. > >Are you still facing an error with this? or its working fine for you? > I can partition and mkfs with /dev/pmem0 after I use the SSD backend. >> >> I took a look into the directory /sys/bus/nd/device. These files are listed. >> Compared with normal system, one device file is missed. > >virtio pmem does not support namespace/region mappings which ACPI NFIT supports. > Ok, thanks. >> >> btt0.0 dax0.0 namespace0.0 ndbus0 pfn0.0 region0 >> >> But the sysfs shows pmem0 block device is created. >> >> /sys/devices/pci0000:00/0000:00:04.0/virtio0/ndbus0/region0/namespace0.0/block/pmem0 >> >> Then I took a look into the pci device: >> >> # lspci -vs 00:04.0 >> 00:04.0 Unclassified device [00ff]: Red Hat, Inc. Device 1013 >> Subsystem: Red Hat, Inc. Device 001b >> Physical Slot: 4 >> Flags: bus master, fast devsel, latency 0, IRQ 11 >> I/O ports at c040 [size=64] >> Memory at fe000000 (64-bit, prefetchable) [size=16K] >> Capabilities: [84] Vendor Specific Information: VirtIO: >> Capabilities: [70] Vendor Specific Information: VirtIO: Notify >> Capabilities: [60] Vendor Specific Information: VirtIO: DeviceCfg >> Capabilities: [50] Vendor Specific Information: VirtIO: ISR >> Capabilities: [40] Vendor Specific Information: VirtIO: CommonCfg >> Kernel driver in use: virtio-pci >> >> This looks good. > >Good to know. > >> >> >> >> >> The guest boots up and I can see /dev/pmem0 device. But when I want to >> >> partition this device, I got the error: >> >> >> >> # parted /dev/pmem0 mklabel gpt >> >> Warning: Error fsyncing/closing /dev/pmem0: Input/output error >> >> >> >> Also I see an error when running "ndctl list": >> >> >> >> libndctl: __sysfs_device_parse: ndctl0: add_dev() failed >> > >> >Will look at this if it is related. >> >> This log still there. > >It looks to me libndctl needs to be taught about about virtio pmem >data parsing. But this is unrelated to kernel and qemu patch series. > Ok, if it works, we could list those regions? and change pmem0 mode to dax mode, right? >> >> > >> >Thanks, >> >Pankaj >> >> >> >> Would you mind letting me know which part I am wrong? >> >> >> >> > (qemu) info memory-devices >> >> > Memory device [virtio-pmem]: "nv1" >> >> > memaddr: 0x240000000 >> >> > size: 4294967296 >> >> > memdev: /objects/mem1 >> >> > >> >> > Implementation is divided into two parts: >> >> > New virtio pmem guest driver and qemu code changes for new >> >> > virtio pmem paravirtualized device. In this series we are >> >> > sharing Qemu device emulation. >> >> > >> >> >1. Guest virtio-pmem kernel driver >> >> >--------------------------------- >> >> > - Reads persistent memory range from paravirt device and >> >> > registers with 'nvdimm_bus'. >> >> > - 'nvdimm/pmem' driver uses this information to allocate >> >> > persistent memory region and setup filesystem operations >> >> > to the allocated memory. >> >> > - virtio pmem driver implements asynchronous flushing >> >> > interface to flush from guest to host. >> >> > >> >> >2. Qemu virtio-pmem device >> >> >--------------------------------- >> >> > - Creates virtio pmem device and exposes a memory range to >> >> > KVM guest. >> >> > - At host side this is file backed memory which acts as >> >> > persistent memory. >> >> > - Qemu side flush uses aio thread pool API's and virtio >> >> > for asynchronous guest multi request handling. >> >> > >> >> > Virtio-pmem security implications and suggested countermeasures: >> >> > --------------------------------------------------------------- >> >> > >> >> > In previous posting of kernel driver, there was discussion [7] >> >> > on possible implications of page cache side channel attacks with >> >> > virtio pmem. After thorough analysis of details of known side >> >> > channel attacks, below are the suggestions: >> >> > >> >> > - Depends entirely on how host backing image file is mapped >> >> > into guest address space. >> >> > >> >> > - virtio-pmem device emulation, by default shared mapping is used >> >> > to map host backing file. It is recommended to use separate >> >> > backing file at host side for every guest. This will prevent >> >> > any possibility of executing common code from multiple guests >> >> > and any chance of inferring guest local data based based on >> >> > execution time. >> >> > >> >> > - If backing file is required to be shared among multiple guests >> >> > it is recommended to don't support host page cache eviction >> >> > commands from the guest driver. This will avoid any possibility >> >> > of inferring guest local data or host data from another guest. >> >> > >> >> > - Proposed device specification [6] for virtio-pmem device with >> >> > details of possible security implications and suggested >> >> > countermeasures for device emulation. >> >> > >> >> >Changes from PATCH v1: >> >> > - Change proposed version from qemu 4.0 to 4.1 - Eric >> >> > - Remove virtio queue_add from unrealize function - Cornelia >> >> > >> >> >[1] https://lkml.org/lkml/2019/6/12/624 >> >> >[2] https://www.spinics.net/lists/kvm/msg149761.html >> >> >[3] https://www.spinics.net/lists/kvm/msg153095.html >> >> >[4] https://marc.info/?l=linux-kernel&m=153572228719237&w=2 >> >> >[5] https://marc.info/?l=qemu-devel&m=153555721901824&w=2 >> >> >[6] https://lists.oasis-open.org/archives/virtio-dev/201903/msg00083.html >> >> >[7] https://lkml.org/lkml/2019/1/9/1191 >> >> > >> >> > Pankaj Gupta (3): >> >> > virtio-pmem: add virtio device >> >> > virtio-pmem: sync linux headers >> >> > virtio-pci: proxy for virtio-pmem >> >> > >> >> > David Hildenbrand (4): >> >> > virtio-pci: Allow to specify additional interfaces for the base type >> >> > hmp: Handle virtio-pmem when printing memory device infos >> >> > numa: Handle virtio-pmem in NUMA stats >> >> > pc: Support for virtio-pmem-pci >> >> > >> >> > hmp.c | 27 ++- >> >> > hw/i386/Kconfig | 1 >> >> > hw/i386/pc.c | 72 ++++++++++ >> >> > hw/virtio/Kconfig | 10 + >> >> > hw/virtio/Makefile.objs | 2 >> >> > hw/virtio/virtio-pci.c | 1 >> >> > hw/virtio/virtio-pci.h | 1 >> >> > hw/virtio/virtio-pmem-pci.c | 131 ++++++++++++++++++ >> >> > hw/virtio/virtio-pmem-pci.h | 34 ++++ >> >> > hw/virtio/virtio-pmem.c | 189 >> >> > +++++++++++++++++++++++++++ >> >> > include/hw/pci/pci.h | 1 >> >> > include/hw/virtio/virtio-pmem.h | 49 +++++++ >> >> > include/standard-headers/linux/virtio_ids.h | 1 >> >> > include/standard-headers/linux/virtio_pmem.h | 35 +++++ >> >> > numa.c | 24 +-- >> >> > qapi/misc.json | 28 +++- >> >> > 16 files changed, 580 insertions(+), 26 deletions(-) >> >> >---- >> >> >> >> -- >> >> Wei Yang >> >> Help you, Help me >> >> >> >> >> >> -- >> Wei Yang >> Help you, Help me >> >>
> >> > > >> >> > >> >> On Wed, Jun 19, 2019 at 03:19:00PM +0530, Pankaj Gupta wrote: > >> >> > This patch series has implementation for "virtio pmem" > >> >> > device. "virtio pmem" is persistent memory(nvdimm) device in > >> >> > guest which allows to bypass the guest page cache. This > >> >> > also implements a VIRTIO based asynchronous flush mechanism. > >> >> > Details of project idea for 'virtio pmem' flushing interface > >> >> > is shared [2] & [3]. > >> >> > > >> >> > Sharing Qemu device emulation in this patchset. Tested with > >> >> > guest kernel driver [1]. This series is based on David's > >> >> > memory device refactoring [5] work with modified version of > >> >> > my initial virtio pmem [4] series. > >> >> > > >> >> > Usage: > >> >> > ./qemu -name test -machine pc -m 8G,slots=240,maxmem=20G > >> >> > -object memory-backend-file,id=mem1,share,mem-path=test.img, > >> >> > size=4G,share > >> >> > -device virtio-pmem-pci,memdev=mem1,id=nv1 > >> >> > > >> >> > >> >> Hi, Pankaj > >> > > >> >Hi Wei, > >> > > >> >> > >> >> I tried this series with v14 kernel driver, while getting some error on > >> >> using > >> >> this. Not sure this is my error configuration. > >> >> > >> >> The qemu command line is: > >> >> > >> >> -object > >> >> memory-backend-file,id=mem1,share=on,mem-path=/dev/dax0.0,size=1G,align=2M > >> >> -device virtio-pmem-pci,memdev=mem1,id=nv1 > >> > > >> >Are you using host backing on real NVDIMM? Currently, we only support > >> >backing image > >> >file on regular SSD. We have plans to support backing file on real NVDIMM > >> >device > >> >as well but this is part of future work. > >> > > >> >Can you please try by allocating a raw image file on regular SSD. It is > >> >working fine > >> >for me. > >> > > >> > >> I created a file with 2G on my disk. > >> > >> #ll -h 2G-file > >> -rw-r--r-- 1 richard richard 2.0G 6月 26 09:26 2G-file > >> > >> The command line is changed to: > >> > >> -object > >> memory-backend-file,id=mem1,share,mem-path=/home/richard/guest/2G-file,size=2G > >> -device virtio-pmem-pci,memdev=mem1,id=nv1 > >> > >> The behavior in guest is the same. > > > >Are you still facing an error with this? or its working fine for you? > > > > I can partition and mkfs with /dev/pmem0 after I use the SSD backend. Good to know. Thanks for the confirmation. > > >> > >> I took a look into the directory /sys/bus/nd/device. These files are > >> listed. > >> Compared with normal system, one device file is missed. > > > >virtio pmem does not support namespace/region mappings which ACPI NFIT > >supports. > > > > Ok, thanks. > > >> > >> btt0.0 dax0.0 namespace0.0 ndbus0 pfn0.0 region0 > >> > >> But the sysfs shows pmem0 block device is created. > >> > >> /sys/devices/pci0000:00/0000:00:04.0/virtio0/ndbus0/region0/namespace0.0/block/pmem0 > >> > >> Then I took a look into the pci device: > >> > >> # lspci -vs 00:04.0 > >> 00:04.0 Unclassified device [00ff]: Red Hat, Inc. Device 1013 > >> Subsystem: Red Hat, Inc. Device 001b > >> Physical Slot: 4 > >> Flags: bus master, fast devsel, latency 0, IRQ 11 > >> I/O ports at c040 [size=64] > >> Memory at fe000000 (64-bit, prefetchable) [size=16K] > >> Capabilities: [84] Vendor Specific Information: VirtIO: > >> Capabilities: [70] Vendor Specific Information: VirtIO: Notify > >> Capabilities: [60] Vendor Specific Information: VirtIO: DeviceCfg > >> Capabilities: [50] Vendor Specific Information: VirtIO: ISR > >> Capabilities: [40] Vendor Specific Information: VirtIO: CommonCfg > >> Kernel driver in use: virtio-pci > >> > >> This looks good. > > > >Good to know. > > > >> > >> >> > >> >> The guest boots up and I can see /dev/pmem0 device. But when I want to > >> >> partition this device, I got the error: > >> >> > >> >> # parted /dev/pmem0 mklabel gpt > >> >> Warning: Error fsyncing/closing /dev/pmem0: Input/output error > >> >> > >> >> Also I see an error when running "ndctl list": > >> >> > >> >> libndctl: __sysfs_device_parse: ndctl0: add_dev() failed > >> > > >> >Will look at this if it is related. > >> > >> This log still there. > > > >It looks to me libndctl needs to be taught about about virtio pmem > >data parsing. But this is unrelated to kernel and qemu patch series. > > > > Ok, if it works, we could list those regions? and change pmem0 mode to dax > mode, right? You mean fs dax? I don't think currently we support this because raw image wont have any metadata. Will have to think if this is at all possible or how we can achieve such behavior. Also, there is requirement to support host backing file on real NVDIMM and virtio. Once we have have first version of virtio pmem series merged upstream we will continue to work on advance features depending upon feasibility. Thanks, Pankaj > > >> > >> > > >> >Thanks, > >> >Pankaj > >> >> > >> >> Would you mind letting me know which part I am wrong? > >> >> > >> >> > (qemu) info memory-devices > >> >> > Memory device [virtio-pmem]: "nv1" > >> >> > memaddr: 0x240000000 > >> >> > size: 4294967296 > >> >> > memdev: /objects/mem1 > >> >> > > >> >> > Implementation is divided into two parts: > >> >> > New virtio pmem guest driver and qemu code changes for new > >> >> > virtio pmem paravirtualized device. In this series we are > >> >> > sharing Qemu device emulation. > >> >> > > >> >> >1. Guest virtio-pmem kernel driver > >> >> >--------------------------------- > >> >> > - Reads persistent memory range from paravirt device and > >> >> > registers with 'nvdimm_bus'. > >> >> > - 'nvdimm/pmem' driver uses this information to allocate > >> >> > persistent memory region and setup filesystem operations > >> >> > to the allocated memory. > >> >> > - virtio pmem driver implements asynchronous flushing > >> >> > interface to flush from guest to host. > >> >> > > >> >> >2. Qemu virtio-pmem device > >> >> >--------------------------------- > >> >> > - Creates virtio pmem device and exposes a memory range to > >> >> > KVM guest. > >> >> > - At host side this is file backed memory which acts as > >> >> > persistent memory. > >> >> > - Qemu side flush uses aio thread pool API's and virtio > >> >> > for asynchronous guest multi request handling. > >> >> > > >> >> > Virtio-pmem security implications and suggested countermeasures: > >> >> > --------------------------------------------------------------- > >> >> > > >> >> > In previous posting of kernel driver, there was discussion [7] > >> >> > on possible implications of page cache side channel attacks with > >> >> > virtio pmem. After thorough analysis of details of known side > >> >> > channel attacks, below are the suggestions: > >> >> > > >> >> > - Depends entirely on how host backing image file is mapped > >> >> > into guest address space. > >> >> > > >> >> > - virtio-pmem device emulation, by default shared mapping is used > >> >> > to map host backing file. It is recommended to use separate > >> >> > backing file at host side for every guest. This will prevent > >> >> > any possibility of executing common code from multiple guests > >> >> > and any chance of inferring guest local data based based on > >> >> > execution time. > >> >> > > >> >> > - If backing file is required to be shared among multiple guests > >> >> > it is recommended to don't support host page cache eviction > >> >> > commands from the guest driver. This will avoid any possibility > >> >> > of inferring guest local data or host data from another guest. > >> >> > > >> >> > - Proposed device specification [6] for virtio-pmem device with > >> >> > details of possible security implications and suggested > >> >> > countermeasures for device emulation. > >> >> > > >> >> >Changes from PATCH v1: > >> >> > - Change proposed version from qemu 4.0 to 4.1 - Eric > >> >> > - Remove virtio queue_add from unrealize function - Cornelia > >> >> > > >> >> >[1] https://lkml.org/lkml/2019/6/12/624 > >> >> >[2] https://www.spinics.net/lists/kvm/msg149761.html > >> >> >[3] https://www.spinics.net/lists/kvm/msg153095.html > >> >> >[4] https://marc.info/?l=linux-kernel&m=153572228719237&w=2 > >> >> >[5] https://marc.info/?l=qemu-devel&m=153555721901824&w=2 > >> >> >[6] > >> >> >https://lists.oasis-open.org/archives/virtio-dev/201903/msg00083.html > >> >> >[7] https://lkml.org/lkml/2019/1/9/1191 > >> >> > > >> >> > Pankaj Gupta (3): > >> >> > virtio-pmem: add virtio device > >> >> > virtio-pmem: sync linux headers > >> >> > virtio-pci: proxy for virtio-pmem > >> >> > > >> >> > David Hildenbrand (4): > >> >> > virtio-pci: Allow to specify additional interfaces for the base type > >> >> > hmp: Handle virtio-pmem when printing memory device infos > >> >> > numa: Handle virtio-pmem in NUMA stats > >> >> > pc: Support for virtio-pmem-pci > >> >> > > >> >> > hmp.c | 27 ++- > >> >> > hw/i386/Kconfig | 1 > >> >> > hw/i386/pc.c | 72 ++++++++++ > >> >> > hw/virtio/Kconfig | 10 + > >> >> > hw/virtio/Makefile.objs | 2 > >> >> > hw/virtio/virtio-pci.c | 1 > >> >> > hw/virtio/virtio-pci.h | 1 > >> >> > hw/virtio/virtio-pmem-pci.c | 131 > >> >> > ++++++++++++++++++ > >> >> > hw/virtio/virtio-pmem-pci.h | 34 ++++ > >> >> > hw/virtio/virtio-pmem.c | 189 > >> >> > +++++++++++++++++++++++++++ > >> >> > include/hw/pci/pci.h | 1 > >> >> > include/hw/virtio/virtio-pmem.h | 49 +++++++ > >> >> > include/standard-headers/linux/virtio_ids.h | 1 > >> >> > include/standard-headers/linux/virtio_pmem.h | 35 +++++ > >> >> > numa.c | 24 +-- > >> >> > qapi/misc.json | 28 +++- > >> >> > 16 files changed, 580 insertions(+), 26 deletions(-) > >> >> >---- > >> >> > >> >> -- > >> >> Wei Yang > >> >> Help you, Help me > >> >> > >> >> > >> > >> -- > >> Wei Yang > >> Help you, Help me > >> > >> > > -- > Wei Yang > Help you, Help me > >
On Tue, Jul 02, 2019 at 10:31:01PM -0400, Pankaj Gupta wrote: > > >> >> Ok, if it works, we could list those regions? and change pmem0 mode to dax >> mode, right? > >You mean fs dax? >I don't think currently we support this because raw image wont have any metadata. >Will have to think if this is at all possible or how we can achieve such behavior. > Ok, I got it. >Also, there is requirement to support host backing file on real NVDIMM and >virtio. Once we have have first version of virtio pmem series merged upstream >we will continue to work on advance features depending upon feasibility. > One curiosity, what difference make NVDIMM backend doesn't work now? The /dev/dax0.0 is a char file. The nvdimm device use mmap to map HVA to HPA. It looks a normal file to me. Would appreciated it if you would share some light on it. >Thanks, >Pankaj > >
> >> > >> Ok, if it works, we could list those regions? and change pmem0 mode to dax > >> mode, right? > > > >You mean fs dax? > >I don't think currently we support this because raw image wont have any > >metadata. > >Will have to think if this is at all possible or how we can achieve such > >behavior. > > > > Ok, I got it. > > >Also, there is requirement to support host backing file on real NVDIMM and > >virtio. Once we have have first version of virtio pmem series merged > >upstream > >we will continue to work on advance features depending upon feasibility. > > > > One curiosity, what difference make NVDIMM backend doesn't work now? > > The /dev/dax0.0 is a char file. The nvdimm device use mmap to map HVA to HPA. > It looks a normal file to me. Would appreciated it if you would share some > light on it. > It will require support of MAP_SYNC at host side without an explicit flush/fsync. We need to add the usecase in both guest and host drivers/device. Also, thorough testing is required to fix any issues. I will do the required changes for the next step. Thanks, Pankaj