358 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			Plaintext
		
	
	
	
	
	
			
		
		
	
	
			358 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			Plaintext
		
	
	
	
	
	
| # SPDX-License-Identifier: GPL-2.0-only
 | |
| menu "Xen driver support"
 | |
| 	depends on XEN
 | |
| 
 | |
| config XEN_BALLOON
 | |
| 	bool "Xen memory balloon driver"
 | |
| 	default y
 | |
| 	help
 | |
| 	  The balloon driver allows the Xen domain to request more memory from
 | |
| 	  the system to expand the domain's memory allocation, or alternatively
 | |
| 	  return unneeded memory to the system.
 | |
| 
 | |
| config XEN_BALLOON_MEMORY_HOTPLUG
 | |
| 	bool "Memory hotplug support for Xen balloon driver"
 | |
| 	depends on XEN_BALLOON && MEMORY_HOTPLUG
 | |
| 	default y
 | |
| 	help
 | |
| 	  Memory hotplug support for Xen balloon driver allows expanding memory
 | |
| 	  available for the system above limit declared at system startup.
 | |
| 	  It is very useful on critical systems which require long
 | |
| 	  run without rebooting.
 | |
| 
 | |
| 	  It's also very useful for non PV domains to obtain unpopulated physical
 | |
| 	  memory ranges to use in order to map foreign memory or grants.
 | |
| 
 | |
| 	  Memory could be hotplugged in following steps:
 | |
| 
 | |
| 	    1) target domain: ensure that memory auto online policy is in
 | |
| 	       effect by checking /sys/devices/system/memory/auto_online_blocks
 | |
| 	       file (should be 'online').
 | |
| 
 | |
| 	    2) control domain: xl mem-max <target-domain> <maxmem>
 | |
| 	       where <maxmem> is >= requested memory size,
 | |
| 
 | |
| 	    3) control domain: xl mem-set <target-domain> <memory>
 | |
| 	       where <memory> is requested memory size; alternatively memory
 | |
| 	       could be added by writing proper value to
 | |
| 	       /sys/devices/system/xen_memory/xen_memory0/target or
 | |
| 	       /sys/devices/system/xen_memory/xen_memory0/target_kb on the
 | |
| 	       target domain.
 | |
| 
 | |
| 	  Alternatively, if memory auto onlining was not requested at step 1
 | |
| 	  the newly added memory can be manually onlined in the target domain
 | |
| 	  by doing the following:
 | |
| 
 | |
| 		for i in /sys/devices/system/memory/memory*/state; do \
 | |
| 		  [ "`cat "$i"`" = offline ] && echo online > "$i"; done
 | |
| 
 | |
| 	  or by adding the following line to udev rules:
 | |
| 
 | |
| 	  SUBSYSTEM=="memory", ACTION=="add", RUN+="/bin/sh -c '[ -f /sys$devpath/state ] && echo online > /sys$devpath/state'"
 | |
| 
 | |
| config XEN_MEMORY_HOTPLUG_LIMIT
 | |
| 	int "Hotplugged memory limit (in GiB) for a PV guest"
 | |
| 	default 512
 | |
| 	depends on XEN_HAVE_PVMMU
 | |
| 	depends on MEMORY_HOTPLUG
 | |
| 	help
 | |
| 	  Maxmium amount of memory (in GiB) that a PV guest can be
 | |
| 	  expanded to when using memory hotplug.
 | |
| 
 | |
| 	  A PV guest can have more memory than this limit if is
 | |
| 	  started with a larger maximum.
 | |
| 
 | |
| 	  This value is used to allocate enough space in internal
 | |
| 	  tables needed for physical memory administration.
 | |
| 
 | |
| config XEN_SCRUB_PAGES_DEFAULT
 | |
| 	bool "Scrub pages before returning them to system by default"
 | |
| 	depends on XEN_BALLOON
 | |
| 	default y
 | |
| 	help
 | |
| 	  Scrub pages before returning them to the system for reuse by
 | |
| 	  other domains.  This makes sure that any confidential data
 | |
| 	  is not accidentally visible to other domains.  It is more
 | |
| 	  secure, but slightly less efficient. This can be controlled with
 | |
| 	  xen_scrub_pages=0 parameter and
 | |
| 	  /sys/devices/system/xen_memory/xen_memory0/scrub_pages.
 | |
| 	  This option only sets the default value.
 | |
| 
 | |
| 	  If in doubt, say yes.
 | |
| 
 | |
| config XEN_DEV_EVTCHN
 | |
| 	tristate "Xen /dev/xen/evtchn device"
 | |
| 	default y
 | |
| 	help
 | |
| 	  The evtchn driver allows a userspace process to trigger event
 | |
| 	  channels and to receive notification of an event channel
 | |
| 	  firing.
 | |
| 	  If in doubt, say yes.
 | |
| 
 | |
| config XEN_BACKEND
 | |
| 	bool "Backend driver support"
 | |
| 	default XEN_DOM0
 | |
| 	help
 | |
| 	  Support for backend device drivers that provide I/O services
 | |
| 	  to other virtual machines.
 | |
| 
 | |
| config XENFS
 | |
| 	tristate "Xen filesystem"
 | |
| 	select XEN_PRIVCMD
 | |
| 	default y
 | |
| 	help
 | |
| 	  The xen filesystem provides a way for domains to share
 | |
| 	  information with each other and with the hypervisor.
 | |
| 	  For example, by reading and writing the "xenbus" file, guests
 | |
| 	  may pass arbitrary information to the initial domain.
 | |
| 	  If in doubt, say yes.
 | |
| 
 | |
| config XEN_COMPAT_XENFS
 | |
| 	bool "Create compatibility mount point /proc/xen"
 | |
| 	depends on XENFS
 | |
| 	default y
 | |
| 	help
 | |
| 	  The old xenstore userspace tools expect to find "xenbus"
 | |
| 	  under /proc/xen, but "xenbus" is now found at the root of the
 | |
| 	  xenfs filesystem.  Selecting this causes the kernel to create
 | |
| 	  the compatibility mount point /proc/xen if it is running on
 | |
| 	  a xen platform.
 | |
| 	  If in doubt, say yes.
 | |
| 
 | |
| config XEN_SYS_HYPERVISOR
 | |
| 	bool "Create xen entries under /sys/hypervisor"
 | |
| 	depends on SYSFS
 | |
| 	select SYS_HYPERVISOR
 | |
| 	default y
 | |
| 	help
 | |
| 	  Create entries under /sys/hypervisor describing the Xen
 | |
| 	  hypervisor environment.  When running native or in another
 | |
| 	  virtual environment, /sys/hypervisor will still be present,
 | |
| 	  but will have no xen contents.
 | |
| 
 | |
| config XEN_XENBUS_FRONTEND
 | |
| 	tristate
 | |
| 
 | |
| config XEN_GNTDEV
 | |
| 	tristate "userspace grant access device driver"
 | |
| 	depends on XEN
 | |
| 	default m
 | |
| 	select MMU_NOTIFIER
 | |
| 	help
 | |
| 	  Allows userspace processes to use grants.
 | |
| 
 | |
| config XEN_GNTDEV_DMABUF
 | |
| 	bool "Add support for dma-buf grant access device driver extension"
 | |
| 	depends on XEN_GNTDEV && XEN_GRANT_DMA_ALLOC
 | |
| 	select DMA_SHARED_BUFFER
 | |
| 	help
 | |
| 	  Allows userspace processes and kernel modules to use Xen backed
 | |
| 	  dma-buf implementation. With this extension grant references to
 | |
| 	  the pages of an imported dma-buf can be exported for other domain
 | |
| 	  use and grant references coming from a foreign domain can be
 | |
| 	  converted into a local dma-buf for local export.
 | |
| 
 | |
| config XEN_GRANT_DEV_ALLOC
 | |
| 	tristate "User-space grant reference allocator driver"
 | |
| 	depends on XEN
 | |
| 	default m
 | |
| 	help
 | |
| 	  Allows userspace processes to create pages with access granted
 | |
| 	  to other domains. This can be used to implement frontend drivers
 | |
| 	  or as part of an inter-domain shared memory channel.
 | |
| 
 | |
| config XEN_GRANT_DMA_ALLOC
 | |
| 	bool "Allow allocating DMA capable buffers with grant reference module"
 | |
| 	depends on XEN && HAS_DMA
 | |
| 	help
 | |
| 	  Extends grant table module API to allow allocating DMA capable
 | |
| 	  buffers and mapping foreign grant references on top of it.
 | |
| 	  The resulting buffer is similar to one allocated by the balloon
 | |
| 	  driver in that proper memory reservation is made by
 | |
| 	  ({increase|decrease}_reservation and VA mappings are updated if
 | |
| 	  needed).
 | |
| 	  This is useful for sharing foreign buffers with HW drivers which
 | |
| 	  cannot work with scattered buffers provided by the balloon driver,
 | |
| 	  but require DMAable memory instead.
 | |
| 
 | |
| config SWIOTLB_XEN
 | |
| 	def_bool y
 | |
| 	depends on XEN_PV || ARM || ARM64
 | |
| 	select DMA_OPS
 | |
| 	select SWIOTLB
 | |
| 
 | |
| config XEN_PCI_STUB
 | |
| 	bool
 | |
| 
 | |
| config XEN_PCIDEV_STUB
 | |
| 	tristate "Xen PCI-device stub driver"
 | |
| 	depends on PCI && !X86 && XEN
 | |
| 	depends on XEN_BACKEND
 | |
| 	select XEN_PCI_STUB
 | |
| 	default m
 | |
| 	help
 | |
| 	  The PCI device stub driver provides limited version of the PCI
 | |
| 	  device backend driver without para-virtualized support for guests.
 | |
| 	  If you select this to be a module, you will need to make sure no
 | |
| 	  other driver has bound to the device(s) you want to make visible to
 | |
| 	  other guests.
 | |
| 
 | |
| 	  The "hide" parameter (only applicable if backend driver is compiled
 | |
| 	  into the kernel) allows you to bind the PCI devices to this module
 | |
| 	  from the default device drivers. The argument is the list of PCI BDFs:
 | |
| 	  xen-pciback.hide=(03:00.0)(04:00.0)
 | |
| 
 | |
| 	  If in doubt, say m.
 | |
| 
 | |
| config XEN_PCIDEV_BACKEND
 | |
| 	tristate "Xen PCI-device backend driver"
 | |
| 	depends on PCI && X86 && XEN
 | |
| 	depends on XEN_BACKEND
 | |
| 	select XEN_PCI_STUB
 | |
| 	default m
 | |
| 	help
 | |
| 	  The PCI device backend driver allows the kernel to export arbitrary
 | |
| 	  PCI devices to other guests. If you select this to be a module, you
 | |
| 	  will need to make sure no other driver has bound to the device(s)
 | |
| 	  you want to make visible to other guests.
 | |
| 
 | |
| 	  The parameter "passthrough" allows you specify how you want the PCI
 | |
| 	  devices to appear in the guest. You can choose the default (0) where
 | |
| 	  PCI topology starts at 00.00.0, or (1) for passthrough if you want
 | |
| 	  the PCI devices topology appear the same as in the host.
 | |
| 
 | |
| 	  The "hide" parameter (only applicable if backend driver is compiled
 | |
| 	  into the kernel) allows you to bind the PCI devices to this module
 | |
| 	  from the default device drivers. The argument is the list of PCI BDFs:
 | |
| 	  xen-pciback.hide=(03:00.0)(04:00.0)
 | |
| 
 | |
| 	  If in doubt, say m.
 | |
| 
 | |
| config XEN_PVCALLS_FRONTEND
 | |
| 	tristate "XEN PV Calls frontend driver"
 | |
| 	depends on INET && XEN
 | |
| 	select XEN_XENBUS_FRONTEND
 | |
| 	help
 | |
| 	  Experimental frontend for the Xen PV Calls protocol
 | |
| 	  (https://xenbits.xen.org/docs/unstable/misc/pvcalls.html). It
 | |
| 	  sends a small set of POSIX calls to the backend, which
 | |
| 	  implements them.
 | |
| 
 | |
| config XEN_PVCALLS_BACKEND
 | |
| 	bool "XEN PV Calls backend driver"
 | |
| 	depends on INET && XEN && XEN_BACKEND
 | |
| 	help
 | |
| 	  Experimental backend for the Xen PV Calls protocol
 | |
| 	  (https://xenbits.xen.org/docs/unstable/misc/pvcalls.html). It
 | |
| 	  allows PV Calls frontends to send POSIX calls to the backend,
 | |
| 	  which implements them.
 | |
| 
 | |
| 	  If in doubt, say n.
 | |
| 
 | |
| config XEN_SCSI_BACKEND
 | |
| 	tristate "XEN SCSI backend driver"
 | |
| 	depends on XEN && XEN_BACKEND && TARGET_CORE
 | |
| 	help
 | |
| 	  The SCSI backend driver allows the kernel to export its SCSI Devices
 | |
| 	  to other guests via a high-performance shared-memory interface.
 | |
| 	  Only needed for systems running as XEN driver domains (e.g. Dom0) and
 | |
| 	  if guests need generic access to SCSI devices.
 | |
| 
 | |
| config XEN_PRIVCMD
 | |
| 	tristate
 | |
| 	depends on XEN
 | |
| 	default m
 | |
| 
 | |
| config XEN_ACPI_PROCESSOR
 | |
| 	tristate "Xen ACPI processor"
 | |
| 	depends on XEN && XEN_PV_DOM0 && X86 && ACPI_PROCESSOR && CPU_FREQ
 | |
| 	default m
 | |
| 	help
 | |
| 	  This ACPI processor uploads Power Management information to the Xen
 | |
| 	  hypervisor.
 | |
| 
 | |
| 	  To do that the driver parses the Power Management data and uploads
 | |
| 	  said information to the Xen hypervisor. Then the Xen hypervisor can
 | |
| 	  select the proper Cx and Pxx states. It also registers itself as the
 | |
| 	  SMM so that other drivers (such as ACPI cpufreq scaling driver) will
 | |
| 	  not load.
 | |
| 
 | |
| 	  To compile this driver as a module, choose M here: the module will be
 | |
| 	  called xen_acpi_processor  If you do not know what to choose, select
 | |
| 	  M here. If the CPUFREQ drivers are built in, select Y here.
 | |
| 
 | |
| config XEN_MCE_LOG
 | |
| 	bool "Xen platform mcelog"
 | |
| 	depends on XEN_PV_DOM0 && X86_MCE
 | |
| 	help
 | |
| 	  Allow kernel fetching MCE error from Xen platform and
 | |
| 	  converting it into Linux mcelog format for mcelog tools
 | |
| 
 | |
| config XEN_HAVE_PVMMU
 | |
| 	bool
 | |
| 
 | |
| config XEN_EFI
 | |
| 	def_bool y
 | |
| 	depends on (ARM || ARM64 || X86_64) && EFI
 | |
| 
 | |
| config XEN_AUTO_XLATE
 | |
| 	def_bool y
 | |
| 	depends on ARM || ARM64 || XEN_PVHVM
 | |
| 	help
 | |
| 	  Support for auto-translated physmap guests.
 | |
| 
 | |
| config XEN_ACPI
 | |
| 	def_bool y
 | |
| 	depends on X86 && ACPI
 | |
| 
 | |
| config XEN_SYMS
 | |
| 	bool "Xen symbols"
 | |
| 	depends on X86 && XEN_DOM0 && XENFS
 | |
| 	default y if KALLSYMS
 | |
| 	help
 | |
| 	  Exports hypervisor symbols (along with their types and addresses) via
 | |
| 	  /proc/xen/xensyms file, similar to /proc/kallsyms
 | |
| 
 | |
| config XEN_HAVE_VPMU
 | |
| 	bool
 | |
| 
 | |
| config XEN_FRONT_PGDIR_SHBUF
 | |
| 	tristate
 | |
| 
 | |
| config XEN_UNPOPULATED_ALLOC
 | |
| 	bool "Use unpopulated memory ranges for guest mappings"
 | |
| 	depends on X86 && ZONE_DEVICE
 | |
| 	default XEN_BACKEND || XEN_GNTDEV || XEN_DOM0
 | |
| 	help
 | |
| 	  Use unpopulated memory ranges in order to create mappings for guest
 | |
| 	  memory regions, including grant maps and foreign pages. This avoids
 | |
| 	  having to balloon out RAM regions in order to obtain physical memory
 | |
| 	  space to create such mappings.
 | |
| 
 | |
| config XEN_VIRTIO
 | |
| 	bool "Xen virtio support"
 | |
| 	depends on VIRTIO
 | |
| 	select XEN_GRANT_DMA_OPS
 | |
| 	select XEN_GRANT_DMA_IOMMU if OF
 | |
| 	help
 | |
| 	  Enable virtio support for running as Xen guest. Depending on the
 | |
| 	  guest type this will require special support on the backend side
 | |
| 	  (qemu or kernel, depending on the virtio device types used).
 | |
| 
 | |
| 	  If in doubt, say n.
 | |
| 
 | |
| config XEN_GRANT_DMA_OPS
 | |
| 	bool
 | |
| 	select DMA_OPS
 | |
| 
 | |
| config XEN_VIRTIO_FORCE_GRANT
 | |
| 	bool "Require Xen virtio support to use grants"
 | |
| 	depends on XEN_VIRTIO
 | |
| 	help
 | |
| 	  Require virtio for Xen guests to use grant mappings.
 | |
| 	  This will avoid the need to give the backend the right to map all
 | |
| 	  of the guest memory. This will need support on the backend side
 | |
| 	  (e.g. qemu or kernel, depending on the virtio device types used).
 | |
| 
 | |
| endmenu
 |