245 lines
		
	
	
		
			7.2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			245 lines
		
	
	
		
			7.2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0-or-later
 | |
| /*
 | |
|  * Test TEST PROTECTION emulation.
 | |
|  *
 | |
|  * Copyright IBM Corp. 2021
 | |
|  */
 | |
| #include <sys/mman.h>
 | |
| #include "test_util.h"
 | |
| #include "kvm_util.h"
 | |
| #include "kselftest.h"
 | |
| #include "ucall_common.h"
 | |
| #include "processor.h"
 | |
| 
 | |
| #define CR0_FETCH_PROTECTION_OVERRIDE	(1UL << (63 - 38))
 | |
| #define CR0_STORAGE_PROTECTION_OVERRIDE	(1UL << (63 - 39))
 | |
| 
 | |
| static __aligned(PAGE_SIZE) uint8_t pages[2][PAGE_SIZE];
 | |
| static uint8_t *const page_store_prot = pages[0];
 | |
| static uint8_t *const page_fetch_prot = pages[1];
 | |
| 
 | |
| /* Nonzero return value indicates that address not mapped */
 | |
| static int set_storage_key(void *addr, uint8_t key)
 | |
| {
 | |
| 	int not_mapped = 0;
 | |
| 
 | |
| 	asm volatile (
 | |
| 		       "lra	%[addr], 0(0,%[addr])\n"
 | |
| 		"	jz	0f\n"
 | |
| 		"	llill	%[not_mapped],1\n"
 | |
| 		"	j	1f\n"
 | |
| 		"0:	sske	%[key], %[addr]\n"
 | |
| 		"1:"
 | |
| 		: [addr] "+&a" (addr), [not_mapped] "+r" (not_mapped)
 | |
| 		: [key] "r" (key)
 | |
| 		: "cc"
 | |
| 	);
 | |
| 	return -not_mapped;
 | |
| }
 | |
| 
 | |
| enum permission {
 | |
| 	READ_WRITE = 0,
 | |
| 	READ = 1,
 | |
| 	RW_PROTECTED = 2,
 | |
| 	TRANSL_UNAVAIL = 3,
 | |
| };
 | |
| 
 | |
| static enum permission test_protection(void *addr, uint8_t key)
 | |
| {
 | |
| 	uint64_t mask;
 | |
| 
 | |
| 	asm volatile (
 | |
| 		       "tprot	%[addr], 0(%[key])\n"
 | |
| 		"	ipm	%[mask]\n"
 | |
| 		: [mask] "=r" (mask)
 | |
| 		: [addr] "Q" (*(char *)addr),
 | |
| 		  [key] "a" (key)
 | |
| 		: "cc"
 | |
| 	);
 | |
| 
 | |
| 	return (enum permission)(mask >> 28);
 | |
| }
 | |
| 
 | |
| enum stage {
 | |
| 	STAGE_INIT_SIMPLE,
 | |
| 	TEST_SIMPLE,
 | |
| 	STAGE_INIT_FETCH_PROT_OVERRIDE,
 | |
| 	TEST_FETCH_PROT_OVERRIDE,
 | |
| 	TEST_STORAGE_PROT_OVERRIDE,
 | |
| 	STAGE_END	/* must be the last entry (it's the amount of tests) */
 | |
| };
 | |
| 
 | |
| struct test {
 | |
| 	enum stage stage;
 | |
| 	void *addr;
 | |
| 	uint8_t key;
 | |
| 	enum permission expected;
 | |
| } tests[] = {
 | |
| 	/*
 | |
| 	 * We perform each test in the array by executing TEST PROTECTION on
 | |
| 	 * the specified addr with the specified key and checking if the returned
 | |
| 	 * permissions match the expected value.
 | |
| 	 * Both guest and host cooperate to set up the required test conditions.
 | |
| 	 * A central condition is that the page targeted by addr has to be DAT
 | |
| 	 * protected in the host mappings, in order for KVM to emulate the
 | |
| 	 * TEST PROTECTION instruction.
 | |
| 	 * Since the page tables are shared, the host uses mprotect to achieve
 | |
| 	 * this.
 | |
| 	 *
 | |
| 	 * Test resulting in RW_PROTECTED/TRANSL_UNAVAIL will be interpreted
 | |
| 	 * by SIE, not KVM, but there is no harm in testing them also.
 | |
| 	 * See Enhanced Suppression-on-Protection Facilities in the
 | |
| 	 * Interpretive-Execution Mode
 | |
| 	 */
 | |
| 	/*
 | |
| 	 * guest: set storage key of page_store_prot to 1
 | |
| 	 *        storage key of page_fetch_prot to 9 and enable
 | |
| 	 *        protection for it
 | |
| 	 * STAGE_INIT_SIMPLE
 | |
| 	 * host: write protect both via mprotect
 | |
| 	 */
 | |
| 	/* access key 0 matches any storage key -> RW */
 | |
| 	{ TEST_SIMPLE, page_store_prot, 0x00, READ_WRITE },
 | |
| 	/* access key matches storage key -> RW */
 | |
| 	{ TEST_SIMPLE, page_store_prot, 0x10, READ_WRITE },
 | |
| 	/* mismatched keys, but no fetch protection -> RO */
 | |
| 	{ TEST_SIMPLE, page_store_prot, 0x20, READ },
 | |
| 	/* access key 0 matches any storage key -> RW */
 | |
| 	{ TEST_SIMPLE, page_fetch_prot, 0x00, READ_WRITE },
 | |
| 	/* access key matches storage key -> RW */
 | |
| 	{ TEST_SIMPLE, page_fetch_prot, 0x90, READ_WRITE },
 | |
| 	/* mismatched keys, fetch protection -> inaccessible */
 | |
| 	{ TEST_SIMPLE, page_fetch_prot, 0x10, RW_PROTECTED },
 | |
| 	/* page 0 not mapped yet -> translation not available */
 | |
| 	{ TEST_SIMPLE, (void *)0x00, 0x10, TRANSL_UNAVAIL },
 | |
| 	/*
 | |
| 	 * host: try to map page 0
 | |
| 	 * guest: set storage key of page 0 to 9 and enable fetch protection
 | |
| 	 * STAGE_INIT_FETCH_PROT_OVERRIDE
 | |
| 	 * host: write protect page 0
 | |
| 	 *       enable fetch protection override
 | |
| 	 */
 | |
| 	/* mismatched keys, fetch protection, but override applies -> RO */
 | |
| 	{ TEST_FETCH_PROT_OVERRIDE, (void *)0x00, 0x10, READ },
 | |
| 	/* mismatched keys, fetch protection, override applies to 0-2048 only -> inaccessible */
 | |
| 	{ TEST_FETCH_PROT_OVERRIDE, (void *)2049, 0x10, RW_PROTECTED },
 | |
| 	/*
 | |
| 	 * host: enable storage protection override
 | |
| 	 */
 | |
| 	/* mismatched keys, but override applies (storage key 9) -> RW */
 | |
| 	{ TEST_STORAGE_PROT_OVERRIDE, page_fetch_prot, 0x10, READ_WRITE },
 | |
| 	/* mismatched keys, no fetch protection, override doesn't apply -> RO */
 | |
| 	{ TEST_STORAGE_PROT_OVERRIDE, page_store_prot, 0x20, READ },
 | |
| 	/* mismatched keys, but override applies (storage key 9) -> RW */
 | |
| 	{ TEST_STORAGE_PROT_OVERRIDE, (void *)2049, 0x10, READ_WRITE },
 | |
| 	/* end marker */
 | |
| 	{ STAGE_END, 0, 0, 0 },
 | |
| };
 | |
| 
 | |
| static enum stage perform_next_stage(int *i, bool mapped_0)
 | |
| {
 | |
| 	enum stage stage = tests[*i].stage;
 | |
| 	enum permission result;
 | |
| 	bool skip;
 | |
| 
 | |
| 	for (; tests[*i].stage == stage; (*i)++) {
 | |
| 		/*
 | |
| 		 * Some fetch protection override tests require that page 0
 | |
| 		 * be mapped, however, when the hosts tries to map that page via
 | |
| 		 * vm_vaddr_alloc, it may happen that some other page gets mapped
 | |
| 		 * instead.
 | |
| 		 * In order to skip these tests we detect this inside the guest
 | |
| 		 */
 | |
| 		skip = tests[*i].addr < (void *)PAGE_SIZE &&
 | |
| 		       tests[*i].expected != TRANSL_UNAVAIL &&
 | |
| 		       !mapped_0;
 | |
| 		if (!skip) {
 | |
| 			result = test_protection(tests[*i].addr, tests[*i].key);
 | |
| 			__GUEST_ASSERT(result == tests[*i].expected,
 | |
| 				       "Wanted %u, got %u, for i = %u",
 | |
| 				       tests[*i].expected, result, *i);
 | |
| 		}
 | |
| 	}
 | |
| 	return stage;
 | |
| }
 | |
| 
 | |
| static void guest_code(void)
 | |
| {
 | |
| 	bool mapped_0;
 | |
| 	int i = 0;
 | |
| 
 | |
| 	GUEST_ASSERT_EQ(set_storage_key(page_store_prot, 0x10), 0);
 | |
| 	GUEST_ASSERT_EQ(set_storage_key(page_fetch_prot, 0x98), 0);
 | |
| 	GUEST_SYNC(STAGE_INIT_SIMPLE);
 | |
| 	GUEST_SYNC(perform_next_stage(&i, false));
 | |
| 
 | |
| 	/* Fetch-protection override */
 | |
| 	mapped_0 = !set_storage_key((void *)0, 0x98);
 | |
| 	GUEST_SYNC(STAGE_INIT_FETCH_PROT_OVERRIDE);
 | |
| 	GUEST_SYNC(perform_next_stage(&i, mapped_0));
 | |
| 
 | |
| 	/* Storage-protection override */
 | |
| 	GUEST_SYNC(perform_next_stage(&i, mapped_0));
 | |
| }
 | |
| 
 | |
| #define HOST_SYNC_NO_TAP(vcpup, stage)				\
 | |
| ({								\
 | |
| 	struct kvm_vcpu *__vcpu = (vcpup);			\
 | |
| 	struct ucall uc;					\
 | |
| 	int __stage = (stage);					\
 | |
| 								\
 | |
| 	vcpu_run(__vcpu);					\
 | |
| 	get_ucall(__vcpu, &uc);					\
 | |
| 	if (uc.cmd == UCALL_ABORT)				\
 | |
| 		REPORT_GUEST_ASSERT(uc);			\
 | |
| 	TEST_ASSERT_EQ(uc.cmd, UCALL_SYNC);			\
 | |
| 	TEST_ASSERT_EQ(uc.args[1], __stage);			\
 | |
| })
 | |
| 
 | |
| #define HOST_SYNC(vcpu, stage)			\
 | |
| ({						\
 | |
| 	HOST_SYNC_NO_TAP(vcpu, stage);		\
 | |
| 	ksft_test_result_pass("" #stage "\n");	\
 | |
| })
 | |
| 
 | |
| int main(int argc, char *argv[])
 | |
| {
 | |
| 	struct kvm_vcpu *vcpu;
 | |
| 	struct kvm_vm *vm;
 | |
| 	struct kvm_run *run;
 | |
| 	vm_vaddr_t guest_0_page;
 | |
| 
 | |
| 	ksft_print_header();
 | |
| 	ksft_set_plan(STAGE_END);
 | |
| 
 | |
| 	vm = vm_create_with_one_vcpu(&vcpu, guest_code);
 | |
| 	run = vcpu->run;
 | |
| 
 | |
| 	HOST_SYNC(vcpu, STAGE_INIT_SIMPLE);
 | |
| 	mprotect(addr_gva2hva(vm, (vm_vaddr_t)pages), PAGE_SIZE * 2, PROT_READ);
 | |
| 	HOST_SYNC(vcpu, TEST_SIMPLE);
 | |
| 
 | |
| 	guest_0_page = vm_vaddr_alloc(vm, PAGE_SIZE, 0);
 | |
| 	if (guest_0_page != 0) {
 | |
| 		/* Use NO_TAP so we don't get a PASS print */
 | |
| 		HOST_SYNC_NO_TAP(vcpu, STAGE_INIT_FETCH_PROT_OVERRIDE);
 | |
| 		ksft_test_result_skip("STAGE_INIT_FETCH_PROT_OVERRIDE - "
 | |
| 				      "Did not allocate page at 0\n");
 | |
| 	} else {
 | |
| 		HOST_SYNC(vcpu, STAGE_INIT_FETCH_PROT_OVERRIDE);
 | |
| 	}
 | |
| 	if (guest_0_page == 0)
 | |
| 		mprotect(addr_gva2hva(vm, (vm_vaddr_t)0), PAGE_SIZE, PROT_READ);
 | |
| 	run->s.regs.crs[0] |= CR0_FETCH_PROTECTION_OVERRIDE;
 | |
| 	run->kvm_dirty_regs = KVM_SYNC_CRS;
 | |
| 	HOST_SYNC(vcpu, TEST_FETCH_PROT_OVERRIDE);
 | |
| 
 | |
| 	run->s.regs.crs[0] |= CR0_STORAGE_PROTECTION_OVERRIDE;
 | |
| 	run->kvm_dirty_regs = KVM_SYNC_CRS;
 | |
| 	HOST_SYNC(vcpu, TEST_STORAGE_PROT_OVERRIDE);
 | |
| 
 | |
| 	kvm_vm_free(vm);
 | |
| 
 | |
| 	ksft_finished();	/* Print results and exit() accordingly */
 | |
| }
 |