1 // SPDX-License-Identifier: GPL-2.0 2 /* Copyright(c) 2016-20 Intel Corporation. */ 3 4 #include <cpuid.h> 5 #include <elf.h> 6 #include <errno.h> 7 #include <fcntl.h> 8 #include <stdbool.h> 9 #include <stdio.h> 10 #include <stdint.h> 11 #include <stdlib.h> 12 #include <string.h> 13 #include <unistd.h> 14 #include <sys/ioctl.h> 15 #include <sys/mman.h> 16 #include <sys/stat.h> 17 #include <sys/time.h> 18 #include <sys/types.h> 19 #include <sys/auxv.h> 20 #include "defines.h" 21 #include "../kselftest_harness.h" 22 #include "main.h" 23 24 static const uint64_t MAGIC = 0x1122334455667788ULL; 25 static const uint64_t MAGIC2 = 0x8877665544332211ULL; 26 vdso_sgx_enter_enclave_t vdso_sgx_enter_enclave; 27 28 /* 29 * Security Information (SECINFO) data structure needed by a few SGX 30 * instructions (eg. ENCLU[EACCEPT] and ENCLU[EMODPE]) holds meta-data 31 * about an enclave page. &enum sgx_secinfo_page_state specifies the 32 * secinfo flags used for page state. 33 */ 34 enum sgx_secinfo_page_state { 35 SGX_SECINFO_PENDING = (1 << 3), 36 SGX_SECINFO_MODIFIED = (1 << 4), 37 SGX_SECINFO_PR = (1 << 5), 38 }; 39 40 struct vdso_symtab { 41 Elf64_Sym *elf_symtab; 42 const char *elf_symstrtab; 43 Elf64_Word *elf_hashtab; 44 }; 45 46 static Elf64_Dyn *vdso_get_dyntab(void *addr) 47 { 48 Elf64_Ehdr *ehdr = addr; 49 Elf64_Phdr *phdrtab = addr + ehdr->e_phoff; 50 int i; 51 52 for (i = 0; i < ehdr->e_phnum; i++) 53 if (phdrtab[i].p_type == PT_DYNAMIC) 54 return addr + phdrtab[i].p_offset; 55 56 return NULL; 57 } 58 59 static void *vdso_get_dyn(void *addr, Elf64_Dyn *dyntab, Elf64_Sxword tag) 60 { 61 int i; 62 63 for (i = 0; dyntab[i].d_tag != DT_NULL; i++) 64 if (dyntab[i].d_tag == tag) 65 return addr + dyntab[i].d_un.d_ptr; 66 67 return NULL; 68 } 69 70 static bool vdso_get_symtab(void *addr, struct vdso_symtab *symtab) 71 { 72 Elf64_Dyn *dyntab = vdso_get_dyntab(addr); 73 74 symtab->elf_symtab = vdso_get_dyn(addr, dyntab, DT_SYMTAB); 75 if (!symtab->elf_symtab) 76 return false; 77 78 symtab->elf_symstrtab = vdso_get_dyn(addr, dyntab, DT_STRTAB); 79 if (!symtab->elf_symstrtab) 80 return false; 81 82 symtab->elf_hashtab = vdso_get_dyn(addr, dyntab, DT_HASH); 83 if (!symtab->elf_hashtab) 84 return false; 85 86 return true; 87 } 88 89 static inline int sgx2_supported(void) 90 { 91 unsigned int eax, ebx, ecx, edx; 92 93 __cpuid_count(SGX_CPUID, 0x0, eax, ebx, ecx, edx); 94 95 return eax & 0x2; 96 } 97 98 static unsigned long elf_sym_hash(const char *name) 99 { 100 unsigned long h = 0, high; 101 102 while (*name) { 103 h = (h << 4) + *name++; 104 high = h & 0xf0000000; 105 106 if (high) 107 h ^= high >> 24; 108 109 h &= ~high; 110 } 111 112 return h; 113 } 114 115 static Elf64_Sym *vdso_symtab_get(struct vdso_symtab *symtab, const char *name) 116 { 117 Elf64_Word bucketnum = symtab->elf_hashtab[0]; 118 Elf64_Word *buckettab = &symtab->elf_hashtab[2]; 119 Elf64_Word *chaintab = &symtab->elf_hashtab[2 + bucketnum]; 120 Elf64_Sym *sym; 121 Elf64_Word i; 122 123 for (i = buckettab[elf_sym_hash(name) % bucketnum]; i != STN_UNDEF; 124 i = chaintab[i]) { 125 sym = &symtab->elf_symtab[i]; 126 if (!strcmp(name, &symtab->elf_symstrtab[sym->st_name])) 127 return sym; 128 } 129 130 return NULL; 131 } 132 133 /* 134 * Return the offset in the enclave where the TCS segment can be found. 135 * The first RW segment loaded is the TCS. 136 */ 137 static off_t encl_get_tcs_offset(struct encl *encl) 138 { 139 int i; 140 141 for (i = 0; i < encl->nr_segments; i++) { 142 struct encl_segment *seg = &encl->segment_tbl[i]; 143 144 if (i == 0 && seg->prot == (PROT_READ | PROT_WRITE)) 145 return seg->offset; 146 } 147 148 return -1; 149 } 150 151 /* 152 * Return the offset in the enclave where the data segment can be found. 153 * The first RW segment loaded is the TCS, skip that to get info on the 154 * data segment. 155 */ 156 static off_t encl_get_data_offset(struct encl *encl) 157 { 158 int i; 159 160 for (i = 1; i < encl->nr_segments; i++) { 161 struct encl_segment *seg = &encl->segment_tbl[i]; 162 163 if (seg->prot == (PROT_READ | PROT_WRITE)) 164 return seg->offset; 165 } 166 167 return -1; 168 } 169 170 FIXTURE(enclave) { 171 struct encl encl; 172 struct sgx_enclave_run run; 173 }; 174 175 static bool setup_test_encl(unsigned long heap_size, struct encl *encl, 176 struct __test_metadata *_metadata) 177 { 178 Elf64_Sym *sgx_enter_enclave_sym = NULL; 179 struct vdso_symtab symtab; 180 struct encl_segment *seg; 181 char maps_line[256]; 182 FILE *maps_file; 183 unsigned int i; 184 void *addr; 185 186 if (!encl_load("test_encl.elf", encl, heap_size)) { 187 encl_delete(encl); 188 TH_LOG("Failed to load the test enclave."); 189 return false; 190 } 191 192 if (!encl_measure(encl)) 193 goto err; 194 195 if (!encl_build(encl)) 196 goto err; 197 198 /* 199 * An enclave consumer only must do this. 200 */ 201 for (i = 0; i < encl->nr_segments; i++) { 202 struct encl_segment *seg = &encl->segment_tbl[i]; 203 204 addr = mmap((void *)encl->encl_base + seg->offset, seg->size, 205 seg->prot, MAP_SHARED | MAP_FIXED, encl->fd, 0); 206 EXPECT_NE(addr, MAP_FAILED); 207 if (addr == MAP_FAILED) 208 goto err; 209 } 210 211 /* Get vDSO base address */ 212 addr = (void *)getauxval(AT_SYSINFO_EHDR); 213 if (!addr) 214 goto err; 215 216 if (!vdso_get_symtab(addr, &symtab)) 217 goto err; 218 219 sgx_enter_enclave_sym = vdso_symtab_get(&symtab, "__vdso_sgx_enter_enclave"); 220 if (!sgx_enter_enclave_sym) 221 goto err; 222 223 vdso_sgx_enter_enclave = addr + sgx_enter_enclave_sym->st_value; 224 225 return true; 226 227 err: 228 for (i = 0; i < encl->nr_segments; i++) { 229 seg = &encl->segment_tbl[i]; 230 231 TH_LOG("0x%016lx 0x%016lx 0x%02x", seg->offset, seg->size, seg->prot); 232 } 233 234 maps_file = fopen("/proc/self/maps", "r"); 235 if (maps_file != NULL) { 236 while (fgets(maps_line, sizeof(maps_line), maps_file) != NULL) { 237 maps_line[strlen(maps_line) - 1] = '\0'; 238 239 if (strstr(maps_line, "/dev/sgx_enclave")) 240 TH_LOG("%s", maps_line); 241 } 242 243 fclose(maps_file); 244 } 245 246 TH_LOG("Failed to initialize the test enclave."); 247 248 encl_delete(encl); 249 250 return false; 251 } 252 253 FIXTURE_SETUP(enclave) 254 { 255 } 256 257 FIXTURE_TEARDOWN(enclave) 258 { 259 encl_delete(&self->encl); 260 } 261 262 #define ENCL_CALL(op, run, clobbered) \ 263 ({ \ 264 int ret; \ 265 if ((clobbered)) \ 266 ret = vdso_sgx_enter_enclave((unsigned long)(op), 0, 0, \ 267 EENTER, 0, 0, (run)); \ 268 else \ 269 ret = sgx_enter_enclave((void *)(op), NULL, 0, EENTER, NULL, NULL, \ 270 (run)); \ 271 ret; \ 272 }) 273 274 #define EXPECT_EEXIT(run) \ 275 do { \ 276 EXPECT_EQ((run)->function, EEXIT); \ 277 if ((run)->function != EEXIT) \ 278 TH_LOG("0x%02x 0x%02x 0x%016llx", (run)->exception_vector, \ 279 (run)->exception_error_code, (run)->exception_addr); \ 280 } while (0) 281 282 TEST_F(enclave, unclobbered_vdso) 283 { 284 struct encl_op_get_from_buf get_op; 285 struct encl_op_put_to_buf put_op; 286 287 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 288 289 memset(&self->run, 0, sizeof(self->run)); 290 self->run.tcs = self->encl.encl_base; 291 292 put_op.header.type = ENCL_OP_PUT_TO_BUFFER; 293 put_op.value = MAGIC; 294 295 EXPECT_EQ(ENCL_CALL(&put_op, &self->run, false), 0); 296 297 EXPECT_EEXIT(&self->run); 298 EXPECT_EQ(self->run.user_data, 0); 299 300 get_op.header.type = ENCL_OP_GET_FROM_BUFFER; 301 get_op.value = 0; 302 303 EXPECT_EQ(ENCL_CALL(&get_op, &self->run, false), 0); 304 305 EXPECT_EQ(get_op.value, MAGIC); 306 EXPECT_EEXIT(&self->run); 307 EXPECT_EQ(self->run.user_data, 0); 308 } 309 310 /* 311 * A section metric is concatenated in a way that @low bits 12-31 define the 312 * bits 12-31 of the metric and @high bits 0-19 define the bits 32-51 of the 313 * metric. 314 */ 315 static unsigned long sgx_calc_section_metric(unsigned int low, 316 unsigned int high) 317 { 318 return (low & GENMASK_ULL(31, 12)) + 319 ((high & GENMASK_ULL(19, 0)) << 32); 320 } 321 322 /* 323 * Sum total available physical SGX memory across all EPC sections 324 * 325 * Return: total available physical SGX memory available on system 326 */ 327 static unsigned long get_total_epc_mem(void) 328 { 329 unsigned int eax, ebx, ecx, edx; 330 unsigned long total_size = 0; 331 unsigned int type; 332 int section = 0; 333 334 while (true) { 335 __cpuid_count(SGX_CPUID, section + SGX_CPUID_EPC, eax, ebx, ecx, edx); 336 337 type = eax & SGX_CPUID_EPC_MASK; 338 if (type == SGX_CPUID_EPC_INVALID) 339 break; 340 341 if (type != SGX_CPUID_EPC_SECTION) 342 break; 343 344 total_size += sgx_calc_section_metric(ecx, edx); 345 346 section++; 347 } 348 349 return total_size; 350 } 351 352 TEST_F(enclave, unclobbered_vdso_oversubscribed) 353 { 354 struct encl_op_get_from_buf get_op; 355 struct encl_op_put_to_buf put_op; 356 unsigned long total_mem; 357 358 total_mem = get_total_epc_mem(); 359 ASSERT_NE(total_mem, 0); 360 ASSERT_TRUE(setup_test_encl(total_mem, &self->encl, _metadata)); 361 362 memset(&self->run, 0, sizeof(self->run)); 363 self->run.tcs = self->encl.encl_base; 364 365 put_op.header.type = ENCL_OP_PUT_TO_BUFFER; 366 put_op.value = MAGIC; 367 368 EXPECT_EQ(ENCL_CALL(&put_op, &self->run, false), 0); 369 370 EXPECT_EEXIT(&self->run); 371 EXPECT_EQ(self->run.user_data, 0); 372 373 get_op.header.type = ENCL_OP_GET_FROM_BUFFER; 374 get_op.value = 0; 375 376 EXPECT_EQ(ENCL_CALL(&get_op, &self->run, false), 0); 377 378 EXPECT_EQ(get_op.value, MAGIC); 379 EXPECT_EEXIT(&self->run); 380 EXPECT_EQ(self->run.user_data, 0); 381 } 382 383 TEST_F_TIMEOUT(enclave, unclobbered_vdso_oversubscribed_remove, 900) 384 { 385 struct sgx_enclave_remove_pages remove_ioc; 386 struct sgx_enclave_modify_types modt_ioc; 387 struct encl_op_get_from_buf get_op; 388 struct encl_op_eaccept eaccept_op; 389 struct encl_op_put_to_buf put_op; 390 struct encl_segment *heap; 391 unsigned long total_mem; 392 int ret, errno_save; 393 unsigned long addr; 394 unsigned long i; 395 396 /* 397 * Create enclave with additional heap that is as big as all 398 * available physical SGX memory. 399 */ 400 total_mem = get_total_epc_mem(); 401 ASSERT_NE(total_mem, 0); 402 TH_LOG("Creating an enclave with %lu bytes heap may take a while ...", 403 total_mem); 404 ASSERT_TRUE(setup_test_encl(total_mem, &self->encl, _metadata)); 405 406 /* 407 * Hardware (SGX2) and kernel support is needed for this test. Start 408 * with check that test has a chance of succeeding. 409 */ 410 memset(&modt_ioc, 0, sizeof(modt_ioc)); 411 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &modt_ioc); 412 413 if (ret == -1) { 414 if (errno == ENOTTY) 415 SKIP(return, 416 "Kernel does not support SGX_IOC_ENCLAVE_MODIFY_TYPES ioctl()"); 417 else if (errno == ENODEV) 418 SKIP(return, "System does not support SGX2"); 419 } 420 421 /* 422 * Invalid parameters were provided during sanity check, 423 * expect command to fail. 424 */ 425 EXPECT_EQ(ret, -1); 426 427 /* SGX2 is supported by kernel and hardware, test can proceed. */ 428 memset(&self->run, 0, sizeof(self->run)); 429 self->run.tcs = self->encl.encl_base; 430 431 heap = &self->encl.segment_tbl[self->encl.nr_segments - 1]; 432 433 put_op.header.type = ENCL_OP_PUT_TO_BUFFER; 434 put_op.value = MAGIC; 435 436 EXPECT_EQ(ENCL_CALL(&put_op, &self->run, false), 0); 437 438 EXPECT_EEXIT(&self->run); 439 EXPECT_EQ(self->run.user_data, 0); 440 441 get_op.header.type = ENCL_OP_GET_FROM_BUFFER; 442 get_op.value = 0; 443 444 EXPECT_EQ(ENCL_CALL(&get_op, &self->run, false), 0); 445 446 EXPECT_EQ(get_op.value, MAGIC); 447 EXPECT_EEXIT(&self->run); 448 EXPECT_EQ(self->run.user_data, 0); 449 450 /* Trim entire heap. */ 451 memset(&modt_ioc, 0, sizeof(modt_ioc)); 452 453 modt_ioc.offset = heap->offset; 454 modt_ioc.length = heap->size; 455 modt_ioc.page_type = SGX_PAGE_TYPE_TRIM; 456 457 TH_LOG("Changing type of %zd bytes to trimmed may take a while ...", 458 heap->size); 459 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &modt_ioc); 460 errno_save = ret == -1 ? errno : 0; 461 462 EXPECT_EQ(ret, 0); 463 EXPECT_EQ(errno_save, 0); 464 EXPECT_EQ(modt_ioc.result, 0); 465 EXPECT_EQ(modt_ioc.count, heap->size); 466 467 /* EACCEPT all removed pages. */ 468 addr = self->encl.encl_base + heap->offset; 469 470 eaccept_op.flags = SGX_SECINFO_TRIM | SGX_SECINFO_MODIFIED; 471 eaccept_op.header.type = ENCL_OP_EACCEPT; 472 473 TH_LOG("Entering enclave to run EACCEPT for each page of %zd bytes may take a while ...", 474 heap->size); 475 for (i = 0; i < heap->size; i += 4096) { 476 eaccept_op.epc_addr = addr + i; 477 eaccept_op.ret = 0; 478 479 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 480 481 EXPECT_EQ(self->run.exception_vector, 0); 482 EXPECT_EQ(self->run.exception_error_code, 0); 483 EXPECT_EQ(self->run.exception_addr, 0); 484 ASSERT_EQ(eaccept_op.ret, 0); 485 ASSERT_EQ(self->run.function, EEXIT); 486 } 487 488 /* Complete page removal. */ 489 memset(&remove_ioc, 0, sizeof(remove_ioc)); 490 491 remove_ioc.offset = heap->offset; 492 remove_ioc.length = heap->size; 493 494 TH_LOG("Removing %zd bytes from enclave may take a while ...", 495 heap->size); 496 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_REMOVE_PAGES, &remove_ioc); 497 errno_save = ret == -1 ? errno : 0; 498 499 EXPECT_EQ(ret, 0); 500 EXPECT_EQ(errno_save, 0); 501 EXPECT_EQ(remove_ioc.count, heap->size); 502 } 503 504 TEST_F(enclave, clobbered_vdso) 505 { 506 struct encl_op_get_from_buf get_op; 507 struct encl_op_put_to_buf put_op; 508 509 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 510 511 memset(&self->run, 0, sizeof(self->run)); 512 self->run.tcs = self->encl.encl_base; 513 514 put_op.header.type = ENCL_OP_PUT_TO_BUFFER; 515 put_op.value = MAGIC; 516 517 EXPECT_EQ(ENCL_CALL(&put_op, &self->run, true), 0); 518 519 EXPECT_EEXIT(&self->run); 520 EXPECT_EQ(self->run.user_data, 0); 521 522 get_op.header.type = ENCL_OP_GET_FROM_BUFFER; 523 get_op.value = 0; 524 525 EXPECT_EQ(ENCL_CALL(&get_op, &self->run, true), 0); 526 527 EXPECT_EQ(get_op.value, MAGIC); 528 EXPECT_EEXIT(&self->run); 529 EXPECT_EQ(self->run.user_data, 0); 530 } 531 532 static int test_handler(long rdi, long rsi, long rdx, long ursp, long r8, long r9, 533 struct sgx_enclave_run *run) 534 { 535 run->user_data = 0; 536 537 return 0; 538 } 539 540 TEST_F(enclave, clobbered_vdso_and_user_function) 541 { 542 struct encl_op_get_from_buf get_op; 543 struct encl_op_put_to_buf put_op; 544 545 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 546 547 memset(&self->run, 0, sizeof(self->run)); 548 self->run.tcs = self->encl.encl_base; 549 550 self->run.user_handler = (__u64)test_handler; 551 self->run.user_data = 0xdeadbeef; 552 553 put_op.header.type = ENCL_OP_PUT_TO_BUFFER; 554 put_op.value = MAGIC; 555 556 EXPECT_EQ(ENCL_CALL(&put_op, &self->run, true), 0); 557 558 EXPECT_EEXIT(&self->run); 559 EXPECT_EQ(self->run.user_data, 0); 560 561 get_op.header.type = ENCL_OP_GET_FROM_BUFFER; 562 get_op.value = 0; 563 564 EXPECT_EQ(ENCL_CALL(&get_op, &self->run, true), 0); 565 566 EXPECT_EQ(get_op.value, MAGIC); 567 EXPECT_EEXIT(&self->run); 568 EXPECT_EQ(self->run.user_data, 0); 569 } 570 571 /* 572 * Sanity check that it is possible to enter either of the two hardcoded TCS 573 */ 574 TEST_F(enclave, tcs_entry) 575 { 576 struct encl_op_header op; 577 578 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 579 580 memset(&self->run, 0, sizeof(self->run)); 581 self->run.tcs = self->encl.encl_base; 582 583 op.type = ENCL_OP_NOP; 584 585 EXPECT_EQ(ENCL_CALL(&op, &self->run, true), 0); 586 587 EXPECT_EEXIT(&self->run); 588 EXPECT_EQ(self->run.exception_vector, 0); 589 EXPECT_EQ(self->run.exception_error_code, 0); 590 EXPECT_EQ(self->run.exception_addr, 0); 591 592 /* Move to the next TCS. */ 593 self->run.tcs = self->encl.encl_base + PAGE_SIZE; 594 595 EXPECT_EQ(ENCL_CALL(&op, &self->run, true), 0); 596 597 EXPECT_EEXIT(&self->run); 598 EXPECT_EQ(self->run.exception_vector, 0); 599 EXPECT_EQ(self->run.exception_error_code, 0); 600 EXPECT_EQ(self->run.exception_addr, 0); 601 } 602 603 /* 604 * Second page of .data segment is used to test changing PTE permissions. 605 * This spans the local encl_buffer within the test enclave. 606 * 607 * 1) Start with a sanity check: a value is written to the target page within 608 * the enclave and read back to ensure target page can be written to. 609 * 2) Change PTE permissions (RW -> RO) of target page within enclave. 610 * 3) Repeat (1) - this time expecting a regular #PF communicated via the 611 * vDSO. 612 * 4) Change PTE permissions of target page within enclave back to be RW. 613 * 5) Repeat (1) by resuming enclave, now expected to be possible to write to 614 * and read from target page within enclave. 615 */ 616 TEST_F(enclave, pte_permissions) 617 { 618 struct encl_op_get_from_addr get_addr_op; 619 struct encl_op_put_to_addr put_addr_op; 620 unsigned long data_start; 621 int ret; 622 623 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 624 625 memset(&self->run, 0, sizeof(self->run)); 626 self->run.tcs = self->encl.encl_base; 627 628 data_start = self->encl.encl_base + 629 encl_get_data_offset(&self->encl) + 630 PAGE_SIZE; 631 632 /* 633 * Sanity check to ensure it is possible to write to page that will 634 * have its permissions manipulated. 635 */ 636 637 /* Write MAGIC to page */ 638 put_addr_op.value = MAGIC; 639 put_addr_op.addr = data_start; 640 put_addr_op.header.type = ENCL_OP_PUT_TO_ADDRESS; 641 642 EXPECT_EQ(ENCL_CALL(&put_addr_op, &self->run, true), 0); 643 644 EXPECT_EEXIT(&self->run); 645 EXPECT_EQ(self->run.exception_vector, 0); 646 EXPECT_EQ(self->run.exception_error_code, 0); 647 EXPECT_EQ(self->run.exception_addr, 0); 648 649 /* 650 * Read memory that was just written to, confirming that it is the 651 * value previously written (MAGIC). 652 */ 653 get_addr_op.value = 0; 654 get_addr_op.addr = data_start; 655 get_addr_op.header.type = ENCL_OP_GET_FROM_ADDRESS; 656 657 EXPECT_EQ(ENCL_CALL(&get_addr_op, &self->run, true), 0); 658 659 EXPECT_EQ(get_addr_op.value, MAGIC); 660 EXPECT_EEXIT(&self->run); 661 EXPECT_EQ(self->run.exception_vector, 0); 662 EXPECT_EQ(self->run.exception_error_code, 0); 663 EXPECT_EQ(self->run.exception_addr, 0); 664 665 /* Change PTE permissions of target page within the enclave */ 666 ret = mprotect((void *)data_start, PAGE_SIZE, PROT_READ); 667 if (ret) 668 perror("mprotect"); 669 670 /* 671 * PTE permissions of target page changed to read-only, EPCM 672 * permissions unchanged (EPCM permissions are RW), attempt to 673 * write to the page, expecting a regular #PF. 674 */ 675 676 put_addr_op.value = MAGIC2; 677 678 EXPECT_EQ(ENCL_CALL(&put_addr_op, &self->run, true), 0); 679 680 EXPECT_EQ(self->run.exception_vector, 14); 681 EXPECT_EQ(self->run.exception_error_code, 0x7); 682 EXPECT_EQ(self->run.exception_addr, data_start); 683 684 self->run.exception_vector = 0; 685 self->run.exception_error_code = 0; 686 self->run.exception_addr = 0; 687 688 /* 689 * Change PTE permissions back to enable enclave to write to the 690 * target page and resume enclave - do not expect any exceptions this 691 * time. 692 */ 693 ret = mprotect((void *)data_start, PAGE_SIZE, PROT_READ | PROT_WRITE); 694 if (ret) 695 perror("mprotect"); 696 697 EXPECT_EQ(vdso_sgx_enter_enclave((unsigned long)&put_addr_op, 0, 698 0, ERESUME, 0, 0, &self->run), 699 0); 700 701 EXPECT_EEXIT(&self->run); 702 EXPECT_EQ(self->run.exception_vector, 0); 703 EXPECT_EQ(self->run.exception_error_code, 0); 704 EXPECT_EQ(self->run.exception_addr, 0); 705 706 get_addr_op.value = 0; 707 708 EXPECT_EQ(ENCL_CALL(&get_addr_op, &self->run, true), 0); 709 710 EXPECT_EQ(get_addr_op.value, MAGIC2); 711 EXPECT_EEXIT(&self->run); 712 EXPECT_EQ(self->run.exception_vector, 0); 713 EXPECT_EQ(self->run.exception_error_code, 0); 714 EXPECT_EQ(self->run.exception_addr, 0); 715 } 716 717 /* 718 * Modifying permissions of TCS page should not be possible. 719 */ 720 TEST_F(enclave, tcs_permissions) 721 { 722 struct sgx_enclave_restrict_permissions ioc; 723 int ret, errno_save; 724 725 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 726 727 memset(&self->run, 0, sizeof(self->run)); 728 self->run.tcs = self->encl.encl_base; 729 730 memset(&ioc, 0, sizeof(ioc)); 731 732 /* 733 * Ensure kernel supports needed ioctl() and system supports needed 734 * commands. 735 */ 736 737 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_RESTRICT_PERMISSIONS, &ioc); 738 errno_save = ret == -1 ? errno : 0; 739 740 /* 741 * Invalid parameters were provided during sanity check, 742 * expect command to fail. 743 */ 744 ASSERT_EQ(ret, -1); 745 746 /* ret == -1 */ 747 if (errno_save == ENOTTY) 748 SKIP(return, 749 "Kernel does not support SGX_IOC_ENCLAVE_RESTRICT_PERMISSIONS ioctl()"); 750 else if (errno_save == ENODEV) 751 SKIP(return, "System does not support SGX2"); 752 753 /* 754 * Attempt to make TCS page read-only. This is not allowed and 755 * should be prevented by the kernel. 756 */ 757 ioc.offset = encl_get_tcs_offset(&self->encl); 758 ioc.length = PAGE_SIZE; 759 ioc.permissions = SGX_SECINFO_R; 760 761 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_RESTRICT_PERMISSIONS, &ioc); 762 errno_save = ret == -1 ? errno : 0; 763 764 EXPECT_EQ(ret, -1); 765 EXPECT_EQ(errno_save, EINVAL); 766 EXPECT_EQ(ioc.result, 0); 767 EXPECT_EQ(ioc.count, 0); 768 } 769 770 /* 771 * Enclave page permission test. 772 * 773 * Modify and restore enclave page's EPCM (enclave) permissions from 774 * outside enclave (ENCLS[EMODPR] via kernel) as well as from within 775 * enclave (via ENCLU[EMODPE]). Check for page fault if 776 * VMA allows access but EPCM permissions do not. 777 */ 778 TEST_F(enclave, epcm_permissions) 779 { 780 struct sgx_enclave_restrict_permissions restrict_ioc; 781 struct encl_op_get_from_addr get_addr_op; 782 struct encl_op_put_to_addr put_addr_op; 783 struct encl_op_eaccept eaccept_op; 784 struct encl_op_emodpe emodpe_op; 785 unsigned long data_start; 786 int ret, errno_save; 787 788 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 789 790 memset(&self->run, 0, sizeof(self->run)); 791 self->run.tcs = self->encl.encl_base; 792 793 /* 794 * Ensure kernel supports needed ioctl() and system supports needed 795 * commands. 796 */ 797 memset(&restrict_ioc, 0, sizeof(restrict_ioc)); 798 799 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_RESTRICT_PERMISSIONS, 800 &restrict_ioc); 801 errno_save = ret == -1 ? errno : 0; 802 803 /* 804 * Invalid parameters were provided during sanity check, 805 * expect command to fail. 806 */ 807 ASSERT_EQ(ret, -1); 808 809 /* ret == -1 */ 810 if (errno_save == ENOTTY) 811 SKIP(return, 812 "Kernel does not support SGX_IOC_ENCLAVE_RESTRICT_PERMISSIONS ioctl()"); 813 else if (errno_save == ENODEV) 814 SKIP(return, "System does not support SGX2"); 815 816 /* 817 * Page that will have its permissions changed is the second data 818 * page in the .data segment. This forms part of the local encl_buffer 819 * within the enclave. 820 * 821 * At start of test @data_start should have EPCM as well as PTE and 822 * VMA permissions of RW. 823 */ 824 825 data_start = self->encl.encl_base + 826 encl_get_data_offset(&self->encl) + PAGE_SIZE; 827 828 /* 829 * Sanity check that page at @data_start is writable before making 830 * any changes to page permissions. 831 * 832 * Start by writing MAGIC to test page. 833 */ 834 put_addr_op.value = MAGIC; 835 put_addr_op.addr = data_start; 836 put_addr_op.header.type = ENCL_OP_PUT_TO_ADDRESS; 837 838 EXPECT_EQ(ENCL_CALL(&put_addr_op, &self->run, true), 0); 839 840 EXPECT_EEXIT(&self->run); 841 EXPECT_EQ(self->run.exception_vector, 0); 842 EXPECT_EQ(self->run.exception_error_code, 0); 843 EXPECT_EQ(self->run.exception_addr, 0); 844 845 /* 846 * Read memory that was just written to, confirming that 847 * page is writable. 848 */ 849 get_addr_op.value = 0; 850 get_addr_op.addr = data_start; 851 get_addr_op.header.type = ENCL_OP_GET_FROM_ADDRESS; 852 853 EXPECT_EQ(ENCL_CALL(&get_addr_op, &self->run, true), 0); 854 855 EXPECT_EQ(get_addr_op.value, MAGIC); 856 EXPECT_EEXIT(&self->run); 857 EXPECT_EQ(self->run.exception_vector, 0); 858 EXPECT_EQ(self->run.exception_error_code, 0); 859 EXPECT_EQ(self->run.exception_addr, 0); 860 861 /* 862 * Change EPCM permissions to read-only. Kernel still considers 863 * the page writable. 864 */ 865 memset(&restrict_ioc, 0, sizeof(restrict_ioc)); 866 867 restrict_ioc.offset = encl_get_data_offset(&self->encl) + PAGE_SIZE; 868 restrict_ioc.length = PAGE_SIZE; 869 restrict_ioc.permissions = SGX_SECINFO_R; 870 871 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_RESTRICT_PERMISSIONS, 872 &restrict_ioc); 873 errno_save = ret == -1 ? errno : 0; 874 875 EXPECT_EQ(ret, 0); 876 EXPECT_EQ(errno_save, 0); 877 EXPECT_EQ(restrict_ioc.result, 0); 878 EXPECT_EQ(restrict_ioc.count, 4096); 879 880 /* 881 * EPCM permissions changed from kernel, need to EACCEPT from enclave. 882 */ 883 eaccept_op.epc_addr = data_start; 884 eaccept_op.flags = SGX_SECINFO_R | SGX_SECINFO_REG | SGX_SECINFO_PR; 885 eaccept_op.ret = 0; 886 eaccept_op.header.type = ENCL_OP_EACCEPT; 887 888 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 889 890 EXPECT_EEXIT(&self->run); 891 EXPECT_EQ(self->run.exception_vector, 0); 892 EXPECT_EQ(self->run.exception_error_code, 0); 893 EXPECT_EQ(self->run.exception_addr, 0); 894 EXPECT_EQ(eaccept_op.ret, 0); 895 896 /* 897 * EPCM permissions of page is now read-only, expect #PF 898 * on EPCM when attempting to write to page from within enclave. 899 */ 900 put_addr_op.value = MAGIC2; 901 902 EXPECT_EQ(ENCL_CALL(&put_addr_op, &self->run, true), 0); 903 904 EXPECT_EQ(self->run.function, ERESUME); 905 EXPECT_EQ(self->run.exception_vector, 14); 906 EXPECT_EQ(self->run.exception_error_code, 0x8007); 907 EXPECT_EQ(self->run.exception_addr, data_start); 908 909 self->run.exception_vector = 0; 910 self->run.exception_error_code = 0; 911 self->run.exception_addr = 0; 912 913 /* 914 * Received AEX but cannot return to enclave at same entrypoint, 915 * need different TCS from where EPCM permission can be made writable 916 * again. 917 */ 918 self->run.tcs = self->encl.encl_base + PAGE_SIZE; 919 920 /* 921 * Enter enclave at new TCS to change EPCM permissions to be 922 * writable again and thus fix the page fault that triggered the 923 * AEX. 924 */ 925 926 emodpe_op.epc_addr = data_start; 927 emodpe_op.flags = SGX_SECINFO_R | SGX_SECINFO_W; 928 emodpe_op.header.type = ENCL_OP_EMODPE; 929 930 EXPECT_EQ(ENCL_CALL(&emodpe_op, &self->run, true), 0); 931 932 EXPECT_EEXIT(&self->run); 933 EXPECT_EQ(self->run.exception_vector, 0); 934 EXPECT_EQ(self->run.exception_error_code, 0); 935 EXPECT_EQ(self->run.exception_addr, 0); 936 937 /* 938 * Attempt to return to main TCS to resume execution at faulting 939 * instruction, PTE should continue to allow writing to the page. 940 */ 941 self->run.tcs = self->encl.encl_base; 942 943 /* 944 * Wrong page permissions that caused original fault has 945 * now been fixed via EPCM permissions. 946 * Resume execution in main TCS to re-attempt the memory access. 947 */ 948 self->run.tcs = self->encl.encl_base; 949 950 EXPECT_EQ(vdso_sgx_enter_enclave((unsigned long)&put_addr_op, 0, 0, 951 ERESUME, 0, 0, 952 &self->run), 953 0); 954 955 EXPECT_EEXIT(&self->run); 956 EXPECT_EQ(self->run.exception_vector, 0); 957 EXPECT_EQ(self->run.exception_error_code, 0); 958 EXPECT_EQ(self->run.exception_addr, 0); 959 960 get_addr_op.value = 0; 961 962 EXPECT_EQ(ENCL_CALL(&get_addr_op, &self->run, true), 0); 963 964 EXPECT_EQ(get_addr_op.value, MAGIC2); 965 EXPECT_EEXIT(&self->run); 966 EXPECT_EQ(self->run.user_data, 0); 967 EXPECT_EQ(self->run.exception_vector, 0); 968 EXPECT_EQ(self->run.exception_error_code, 0); 969 EXPECT_EQ(self->run.exception_addr, 0); 970 } 971 972 /* 973 * Test the addition of pages to an initialized enclave via writing to 974 * a page belonging to the enclave's address space but was not added 975 * during enclave creation. 976 */ 977 TEST_F(enclave, augment) 978 { 979 struct encl_op_get_from_addr get_addr_op; 980 struct encl_op_put_to_addr put_addr_op; 981 struct encl_op_eaccept eaccept_op; 982 size_t total_size = 0; 983 void *addr; 984 int i; 985 986 if (!sgx2_supported()) 987 SKIP(return, "SGX2 not supported"); 988 989 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 990 991 memset(&self->run, 0, sizeof(self->run)); 992 self->run.tcs = self->encl.encl_base; 993 994 for (i = 0; i < self->encl.nr_segments; i++) { 995 struct encl_segment *seg = &self->encl.segment_tbl[i]; 996 997 total_size += seg->size; 998 } 999 1000 /* 1001 * Actual enclave size is expected to be larger than the loaded 1002 * test enclave since enclave size must be a power of 2 in bytes 1003 * and test_encl does not consume it all. 1004 */ 1005 EXPECT_LT(total_size + PAGE_SIZE, self->encl.encl_size); 1006 1007 /* 1008 * Create memory mapping for the page that will be added. New 1009 * memory mapping is for one page right after all existing 1010 * mappings. 1011 * Kernel will allow new mapping using any permissions if it 1012 * falls into the enclave's address range but not backed 1013 * by existing enclave pages. 1014 */ 1015 addr = mmap((void *)self->encl.encl_base + total_size, PAGE_SIZE, 1016 PROT_READ | PROT_WRITE | PROT_EXEC, 1017 MAP_SHARED | MAP_FIXED, self->encl.fd, 0); 1018 EXPECT_NE(addr, MAP_FAILED); 1019 1020 self->run.exception_vector = 0; 1021 self->run.exception_error_code = 0; 1022 self->run.exception_addr = 0; 1023 1024 /* 1025 * Attempt to write to the new page from within enclave. 1026 * Expected to fail since page is not (yet) part of the enclave. 1027 * The first #PF will trigger the addition of the page to the 1028 * enclave, but since the new page needs an EACCEPT from within the 1029 * enclave before it can be used it would not be possible 1030 * to successfully return to the failing instruction. This is the 1031 * cause of the second #PF captured here having the SGX bit set, 1032 * it is from hardware preventing the page from being used. 1033 */ 1034 put_addr_op.value = MAGIC; 1035 put_addr_op.addr = (unsigned long)addr; 1036 put_addr_op.header.type = ENCL_OP_PUT_TO_ADDRESS; 1037 1038 EXPECT_EQ(ENCL_CALL(&put_addr_op, &self->run, true), 0); 1039 1040 EXPECT_EQ(self->run.function, ERESUME); 1041 EXPECT_EQ(self->run.exception_vector, 14); 1042 EXPECT_EQ(self->run.exception_addr, (unsigned long)addr); 1043 1044 if (self->run.exception_error_code == 0x6) { 1045 munmap(addr, PAGE_SIZE); 1046 SKIP(return, "Kernel does not support adding pages to initialized enclave"); 1047 } 1048 1049 EXPECT_EQ(self->run.exception_error_code, 0x8007); 1050 1051 self->run.exception_vector = 0; 1052 self->run.exception_error_code = 0; 1053 self->run.exception_addr = 0; 1054 1055 /* Handle AEX by running EACCEPT from new entry point. */ 1056 self->run.tcs = self->encl.encl_base + PAGE_SIZE; 1057 1058 eaccept_op.epc_addr = self->encl.encl_base + total_size; 1059 eaccept_op.flags = SGX_SECINFO_R | SGX_SECINFO_W | SGX_SECINFO_REG | SGX_SECINFO_PENDING; 1060 eaccept_op.ret = 0; 1061 eaccept_op.header.type = ENCL_OP_EACCEPT; 1062 1063 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 1064 1065 EXPECT_EEXIT(&self->run); 1066 EXPECT_EQ(self->run.exception_vector, 0); 1067 EXPECT_EQ(self->run.exception_error_code, 0); 1068 EXPECT_EQ(self->run.exception_addr, 0); 1069 EXPECT_EQ(eaccept_op.ret, 0); 1070 1071 /* Can now return to main TCS to resume execution. */ 1072 self->run.tcs = self->encl.encl_base; 1073 1074 EXPECT_EQ(vdso_sgx_enter_enclave((unsigned long)&put_addr_op, 0, 0, 1075 ERESUME, 0, 0, 1076 &self->run), 1077 0); 1078 1079 EXPECT_EEXIT(&self->run); 1080 EXPECT_EQ(self->run.exception_vector, 0); 1081 EXPECT_EQ(self->run.exception_error_code, 0); 1082 EXPECT_EQ(self->run.exception_addr, 0); 1083 1084 /* 1085 * Read memory from newly added page that was just written to, 1086 * confirming that data previously written (MAGIC) is present. 1087 */ 1088 get_addr_op.value = 0; 1089 get_addr_op.addr = (unsigned long)addr; 1090 get_addr_op.header.type = ENCL_OP_GET_FROM_ADDRESS; 1091 1092 EXPECT_EQ(ENCL_CALL(&get_addr_op, &self->run, true), 0); 1093 1094 EXPECT_EQ(get_addr_op.value, MAGIC); 1095 EXPECT_EEXIT(&self->run); 1096 EXPECT_EQ(self->run.exception_vector, 0); 1097 EXPECT_EQ(self->run.exception_error_code, 0); 1098 EXPECT_EQ(self->run.exception_addr, 0); 1099 1100 munmap(addr, PAGE_SIZE); 1101 } 1102 1103 /* 1104 * Test for the addition of pages to an initialized enclave via a 1105 * pre-emptive run of EACCEPT on page to be added. 1106 */ 1107 TEST_F(enclave, augment_via_eaccept) 1108 { 1109 struct encl_op_get_from_addr get_addr_op; 1110 struct encl_op_put_to_addr put_addr_op; 1111 struct encl_op_eaccept eaccept_op; 1112 size_t total_size = 0; 1113 void *addr; 1114 int i; 1115 1116 if (!sgx2_supported()) 1117 SKIP(return, "SGX2 not supported"); 1118 1119 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 1120 1121 memset(&self->run, 0, sizeof(self->run)); 1122 self->run.tcs = self->encl.encl_base; 1123 1124 for (i = 0; i < self->encl.nr_segments; i++) { 1125 struct encl_segment *seg = &self->encl.segment_tbl[i]; 1126 1127 total_size += seg->size; 1128 } 1129 1130 /* 1131 * Actual enclave size is expected to be larger than the loaded 1132 * test enclave since enclave size must be a power of 2 in bytes while 1133 * test_encl does not consume it all. 1134 */ 1135 EXPECT_LT(total_size + PAGE_SIZE, self->encl.encl_size); 1136 1137 /* 1138 * mmap() a page at end of existing enclave to be used for dynamic 1139 * EPC page. 1140 * 1141 * Kernel will allow new mapping using any permissions if it 1142 * falls into the enclave's address range but not backed 1143 * by existing enclave pages. 1144 */ 1145 1146 addr = mmap((void *)self->encl.encl_base + total_size, PAGE_SIZE, 1147 PROT_READ | PROT_WRITE | PROT_EXEC, MAP_SHARED | MAP_FIXED, 1148 self->encl.fd, 0); 1149 EXPECT_NE(addr, MAP_FAILED); 1150 1151 self->run.exception_vector = 0; 1152 self->run.exception_error_code = 0; 1153 self->run.exception_addr = 0; 1154 1155 /* 1156 * Run EACCEPT on new page to trigger the #PF->EAUG->EACCEPT(again 1157 * without a #PF). All should be transparent to userspace. 1158 */ 1159 eaccept_op.epc_addr = self->encl.encl_base + total_size; 1160 eaccept_op.flags = SGX_SECINFO_R | SGX_SECINFO_W | SGX_SECINFO_REG | SGX_SECINFO_PENDING; 1161 eaccept_op.ret = 0; 1162 eaccept_op.header.type = ENCL_OP_EACCEPT; 1163 1164 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 1165 1166 if (self->run.exception_vector == 14 && 1167 self->run.exception_error_code == 4 && 1168 self->run.exception_addr == self->encl.encl_base + total_size) { 1169 munmap(addr, PAGE_SIZE); 1170 SKIP(return, "Kernel does not support adding pages to initialized enclave"); 1171 } 1172 1173 EXPECT_EEXIT(&self->run); 1174 EXPECT_EQ(self->run.exception_vector, 0); 1175 EXPECT_EQ(self->run.exception_error_code, 0); 1176 EXPECT_EQ(self->run.exception_addr, 0); 1177 EXPECT_EQ(eaccept_op.ret, 0); 1178 1179 /* 1180 * New page should be accessible from within enclave - attempt to 1181 * write to it. 1182 */ 1183 put_addr_op.value = MAGIC; 1184 put_addr_op.addr = (unsigned long)addr; 1185 put_addr_op.header.type = ENCL_OP_PUT_TO_ADDRESS; 1186 1187 EXPECT_EQ(ENCL_CALL(&put_addr_op, &self->run, true), 0); 1188 1189 EXPECT_EEXIT(&self->run); 1190 EXPECT_EQ(self->run.exception_vector, 0); 1191 EXPECT_EQ(self->run.exception_error_code, 0); 1192 EXPECT_EQ(self->run.exception_addr, 0); 1193 1194 /* 1195 * Read memory from newly added page that was just written to, 1196 * confirming that data previously written (MAGIC) is present. 1197 */ 1198 get_addr_op.value = 0; 1199 get_addr_op.addr = (unsigned long)addr; 1200 get_addr_op.header.type = ENCL_OP_GET_FROM_ADDRESS; 1201 1202 EXPECT_EQ(ENCL_CALL(&get_addr_op, &self->run, true), 0); 1203 1204 EXPECT_EQ(get_addr_op.value, MAGIC); 1205 EXPECT_EEXIT(&self->run); 1206 EXPECT_EQ(self->run.exception_vector, 0); 1207 EXPECT_EQ(self->run.exception_error_code, 0); 1208 EXPECT_EQ(self->run.exception_addr, 0); 1209 1210 munmap(addr, PAGE_SIZE); 1211 } 1212 1213 /* 1214 * SGX2 page type modification test in two phases: 1215 * Phase 1: 1216 * Create a new TCS, consisting out of three new pages (stack page with regular 1217 * page type, SSA page with regular page type, and TCS page with TCS page 1218 * type) in an initialized enclave and run a simple workload within it. 1219 * Phase 2: 1220 * Remove the three pages added in phase 1, add a new regular page at the 1221 * same address that previously hosted the TCS page and verify that it can 1222 * be modified. 1223 */ 1224 TEST_F(enclave, tcs_create) 1225 { 1226 struct encl_op_init_tcs_page init_tcs_page_op; 1227 struct sgx_enclave_remove_pages remove_ioc; 1228 struct encl_op_get_from_addr get_addr_op; 1229 struct sgx_enclave_modify_types modt_ioc; 1230 struct encl_op_put_to_addr put_addr_op; 1231 struct encl_op_get_from_buf get_buf_op; 1232 struct encl_op_put_to_buf put_buf_op; 1233 void *addr, *tcs, *stack_end, *ssa; 1234 struct encl_op_eaccept eaccept_op; 1235 size_t total_size = 0; 1236 uint64_t val_64; 1237 int errno_save; 1238 int ret, i; 1239 1240 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, 1241 _metadata)); 1242 1243 memset(&self->run, 0, sizeof(self->run)); 1244 self->run.tcs = self->encl.encl_base; 1245 1246 /* 1247 * Hardware (SGX2) and kernel support is needed for this test. Start 1248 * with check that test has a chance of succeeding. 1249 */ 1250 memset(&modt_ioc, 0, sizeof(modt_ioc)); 1251 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &modt_ioc); 1252 1253 if (ret == -1) { 1254 if (errno == ENOTTY) 1255 SKIP(return, 1256 "Kernel does not support SGX_IOC_ENCLAVE_MODIFY_TYPES ioctl()"); 1257 else if (errno == ENODEV) 1258 SKIP(return, "System does not support SGX2"); 1259 } 1260 1261 /* 1262 * Invalid parameters were provided during sanity check, 1263 * expect command to fail. 1264 */ 1265 EXPECT_EQ(ret, -1); 1266 1267 /* 1268 * Add three regular pages via EAUG: one will be the TCS stack, one 1269 * will be the TCS SSA, and one will be the new TCS. The stack and 1270 * SSA will remain as regular pages, the TCS page will need its 1271 * type changed after populated with needed data. 1272 */ 1273 for (i = 0; i < self->encl.nr_segments; i++) { 1274 struct encl_segment *seg = &self->encl.segment_tbl[i]; 1275 1276 total_size += seg->size; 1277 } 1278 1279 /* 1280 * Actual enclave size is expected to be larger than the loaded 1281 * test enclave since enclave size must be a power of 2 in bytes while 1282 * test_encl does not consume it all. 1283 */ 1284 EXPECT_LT(total_size + 3 * PAGE_SIZE, self->encl.encl_size); 1285 1286 /* 1287 * mmap() three pages at end of existing enclave to be used for the 1288 * three new pages. 1289 */ 1290 addr = mmap((void *)self->encl.encl_base + total_size, 3 * PAGE_SIZE, 1291 PROT_READ | PROT_WRITE, MAP_SHARED | MAP_FIXED, 1292 self->encl.fd, 0); 1293 EXPECT_NE(addr, MAP_FAILED); 1294 1295 self->run.exception_vector = 0; 1296 self->run.exception_error_code = 0; 1297 self->run.exception_addr = 0; 1298 1299 stack_end = (void *)self->encl.encl_base + total_size; 1300 tcs = (void *)self->encl.encl_base + total_size + PAGE_SIZE; 1301 ssa = (void *)self->encl.encl_base + total_size + 2 * PAGE_SIZE; 1302 1303 /* 1304 * Run EACCEPT on each new page to trigger the 1305 * EACCEPT->(#PF)->EAUG->EACCEPT(again without a #PF) flow. 1306 */ 1307 1308 eaccept_op.epc_addr = (unsigned long)stack_end; 1309 eaccept_op.flags = SGX_SECINFO_R | SGX_SECINFO_W | SGX_SECINFO_REG | SGX_SECINFO_PENDING; 1310 eaccept_op.ret = 0; 1311 eaccept_op.header.type = ENCL_OP_EACCEPT; 1312 1313 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 1314 1315 if (self->run.exception_vector == 14 && 1316 self->run.exception_error_code == 4 && 1317 self->run.exception_addr == (unsigned long)stack_end) { 1318 munmap(addr, 3 * PAGE_SIZE); 1319 SKIP(return, "Kernel does not support adding pages to initialized enclave"); 1320 } 1321 1322 EXPECT_EEXIT(&self->run); 1323 EXPECT_EQ(self->run.exception_vector, 0); 1324 EXPECT_EQ(self->run.exception_error_code, 0); 1325 EXPECT_EQ(self->run.exception_addr, 0); 1326 EXPECT_EQ(eaccept_op.ret, 0); 1327 1328 eaccept_op.epc_addr = (unsigned long)ssa; 1329 1330 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 1331 1332 EXPECT_EEXIT(&self->run); 1333 EXPECT_EQ(self->run.exception_vector, 0); 1334 EXPECT_EQ(self->run.exception_error_code, 0); 1335 EXPECT_EQ(self->run.exception_addr, 0); 1336 EXPECT_EQ(eaccept_op.ret, 0); 1337 1338 eaccept_op.epc_addr = (unsigned long)tcs; 1339 1340 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 1341 1342 EXPECT_EEXIT(&self->run); 1343 EXPECT_EQ(self->run.exception_vector, 0); 1344 EXPECT_EQ(self->run.exception_error_code, 0); 1345 EXPECT_EQ(self->run.exception_addr, 0); 1346 EXPECT_EQ(eaccept_op.ret, 0); 1347 1348 /* 1349 * Three new pages added to enclave. Now populate the TCS page with 1350 * needed data. This should be done from within enclave. Provide 1351 * the function that will do the actual data population with needed 1352 * data. 1353 */ 1354 1355 /* 1356 * New TCS will use the "encl_dyn_entry" entrypoint that expects 1357 * stack to begin in page before TCS page. 1358 */ 1359 val_64 = encl_get_entry(&self->encl, "encl_dyn_entry"); 1360 EXPECT_NE(val_64, 0); 1361 1362 init_tcs_page_op.tcs_page = (unsigned long)tcs; 1363 init_tcs_page_op.ssa = (unsigned long)total_size + 2 * PAGE_SIZE; 1364 init_tcs_page_op.entry = val_64; 1365 init_tcs_page_op.header.type = ENCL_OP_INIT_TCS_PAGE; 1366 1367 EXPECT_EQ(ENCL_CALL(&init_tcs_page_op, &self->run, true), 0); 1368 1369 EXPECT_EEXIT(&self->run); 1370 EXPECT_EQ(self->run.exception_vector, 0); 1371 EXPECT_EQ(self->run.exception_error_code, 0); 1372 EXPECT_EQ(self->run.exception_addr, 0); 1373 1374 /* Change TCS page type to TCS. */ 1375 memset(&modt_ioc, 0, sizeof(modt_ioc)); 1376 1377 modt_ioc.offset = total_size + PAGE_SIZE; 1378 modt_ioc.length = PAGE_SIZE; 1379 modt_ioc.page_type = SGX_PAGE_TYPE_TCS; 1380 1381 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &modt_ioc); 1382 errno_save = ret == -1 ? errno : 0; 1383 1384 EXPECT_EQ(ret, 0); 1385 EXPECT_EQ(errno_save, 0); 1386 EXPECT_EQ(modt_ioc.result, 0); 1387 EXPECT_EQ(modt_ioc.count, 4096); 1388 1389 /* EACCEPT new TCS page from enclave. */ 1390 eaccept_op.epc_addr = (unsigned long)tcs; 1391 eaccept_op.flags = SGX_SECINFO_TCS | SGX_SECINFO_MODIFIED; 1392 eaccept_op.ret = 0; 1393 eaccept_op.header.type = ENCL_OP_EACCEPT; 1394 1395 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 1396 1397 EXPECT_EEXIT(&self->run); 1398 EXPECT_EQ(self->run.exception_vector, 0); 1399 EXPECT_EQ(self->run.exception_error_code, 0); 1400 EXPECT_EQ(self->run.exception_addr, 0); 1401 EXPECT_EQ(eaccept_op.ret, 0); 1402 1403 /* Run workload from new TCS. */ 1404 self->run.tcs = (unsigned long)tcs; 1405 1406 /* 1407 * Simple workload to write to data buffer and read value back. 1408 */ 1409 put_buf_op.header.type = ENCL_OP_PUT_TO_BUFFER; 1410 put_buf_op.value = MAGIC; 1411 1412 EXPECT_EQ(ENCL_CALL(&put_buf_op, &self->run, true), 0); 1413 1414 EXPECT_EEXIT(&self->run); 1415 EXPECT_EQ(self->run.exception_vector, 0); 1416 EXPECT_EQ(self->run.exception_error_code, 0); 1417 EXPECT_EQ(self->run.exception_addr, 0); 1418 1419 get_buf_op.header.type = ENCL_OP_GET_FROM_BUFFER; 1420 get_buf_op.value = 0; 1421 1422 EXPECT_EQ(ENCL_CALL(&get_buf_op, &self->run, true), 0); 1423 1424 EXPECT_EQ(get_buf_op.value, MAGIC); 1425 EXPECT_EEXIT(&self->run); 1426 EXPECT_EQ(self->run.exception_vector, 0); 1427 EXPECT_EQ(self->run.exception_error_code, 0); 1428 EXPECT_EQ(self->run.exception_addr, 0); 1429 1430 /* 1431 * Phase 2 of test: 1432 * Remove pages associated with new TCS, create a regular page 1433 * where TCS page used to be and verify it can be used as a regular 1434 * page. 1435 */ 1436 1437 /* Start page removal by requesting change of page type to PT_TRIM. */ 1438 memset(&modt_ioc, 0, sizeof(modt_ioc)); 1439 1440 modt_ioc.offset = total_size; 1441 modt_ioc.length = 3 * PAGE_SIZE; 1442 modt_ioc.page_type = SGX_PAGE_TYPE_TRIM; 1443 1444 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &modt_ioc); 1445 errno_save = ret == -1 ? errno : 0; 1446 1447 EXPECT_EQ(ret, 0); 1448 EXPECT_EQ(errno_save, 0); 1449 EXPECT_EQ(modt_ioc.result, 0); 1450 EXPECT_EQ(modt_ioc.count, 3 * PAGE_SIZE); 1451 1452 /* 1453 * Enter enclave via TCS #1 and approve page removal by sending 1454 * EACCEPT for each of three removed pages. 1455 */ 1456 self->run.tcs = self->encl.encl_base; 1457 1458 eaccept_op.epc_addr = (unsigned long)stack_end; 1459 eaccept_op.flags = SGX_SECINFO_TRIM | SGX_SECINFO_MODIFIED; 1460 eaccept_op.ret = 0; 1461 eaccept_op.header.type = ENCL_OP_EACCEPT; 1462 1463 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 1464 1465 EXPECT_EEXIT(&self->run); 1466 EXPECT_EQ(self->run.exception_vector, 0); 1467 EXPECT_EQ(self->run.exception_error_code, 0); 1468 EXPECT_EQ(self->run.exception_addr, 0); 1469 EXPECT_EQ(eaccept_op.ret, 0); 1470 1471 eaccept_op.epc_addr = (unsigned long)tcs; 1472 eaccept_op.ret = 0; 1473 1474 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 1475 1476 EXPECT_EEXIT(&self->run); 1477 EXPECT_EQ(self->run.exception_vector, 0); 1478 EXPECT_EQ(self->run.exception_error_code, 0); 1479 EXPECT_EQ(self->run.exception_addr, 0); 1480 EXPECT_EQ(eaccept_op.ret, 0); 1481 1482 eaccept_op.epc_addr = (unsigned long)ssa; 1483 eaccept_op.ret = 0; 1484 1485 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 1486 1487 EXPECT_EEXIT(&self->run); 1488 EXPECT_EQ(self->run.exception_vector, 0); 1489 EXPECT_EQ(self->run.exception_error_code, 0); 1490 EXPECT_EQ(self->run.exception_addr, 0); 1491 EXPECT_EQ(eaccept_op.ret, 0); 1492 1493 /* Send final ioctl() to complete page removal. */ 1494 memset(&remove_ioc, 0, sizeof(remove_ioc)); 1495 1496 remove_ioc.offset = total_size; 1497 remove_ioc.length = 3 * PAGE_SIZE; 1498 1499 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_REMOVE_PAGES, &remove_ioc); 1500 errno_save = ret == -1 ? errno : 0; 1501 1502 EXPECT_EQ(ret, 0); 1503 EXPECT_EQ(errno_save, 0); 1504 EXPECT_EQ(remove_ioc.count, 3 * PAGE_SIZE); 1505 1506 /* 1507 * Enter enclave via TCS #1 and access location where TCS #3 was to 1508 * trigger dynamic add of regular page at that location. 1509 */ 1510 eaccept_op.epc_addr = (unsigned long)tcs; 1511 eaccept_op.flags = SGX_SECINFO_R | SGX_SECINFO_W | SGX_SECINFO_REG | SGX_SECINFO_PENDING; 1512 eaccept_op.ret = 0; 1513 eaccept_op.header.type = ENCL_OP_EACCEPT; 1514 1515 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 1516 1517 EXPECT_EEXIT(&self->run); 1518 EXPECT_EQ(self->run.exception_vector, 0); 1519 EXPECT_EQ(self->run.exception_error_code, 0); 1520 EXPECT_EQ(self->run.exception_addr, 0); 1521 EXPECT_EQ(eaccept_op.ret, 0); 1522 1523 /* 1524 * New page should be accessible from within enclave - write to it. 1525 */ 1526 put_addr_op.value = MAGIC; 1527 put_addr_op.addr = (unsigned long)tcs; 1528 put_addr_op.header.type = ENCL_OP_PUT_TO_ADDRESS; 1529 1530 EXPECT_EQ(ENCL_CALL(&put_addr_op, &self->run, true), 0); 1531 1532 EXPECT_EEXIT(&self->run); 1533 EXPECT_EQ(self->run.exception_vector, 0); 1534 EXPECT_EQ(self->run.exception_error_code, 0); 1535 EXPECT_EQ(self->run.exception_addr, 0); 1536 1537 /* 1538 * Read memory from newly added page that was just written to, 1539 * confirming that data previously written (MAGIC) is present. 1540 */ 1541 get_addr_op.value = 0; 1542 get_addr_op.addr = (unsigned long)tcs; 1543 get_addr_op.header.type = ENCL_OP_GET_FROM_ADDRESS; 1544 1545 EXPECT_EQ(ENCL_CALL(&get_addr_op, &self->run, true), 0); 1546 1547 EXPECT_EQ(get_addr_op.value, MAGIC); 1548 EXPECT_EEXIT(&self->run); 1549 EXPECT_EQ(self->run.exception_vector, 0); 1550 EXPECT_EQ(self->run.exception_error_code, 0); 1551 EXPECT_EQ(self->run.exception_addr, 0); 1552 1553 munmap(addr, 3 * PAGE_SIZE); 1554 } 1555 1556 /* 1557 * Ensure sane behavior if user requests page removal, does not run 1558 * EACCEPT from within enclave but still attempts to finalize page removal 1559 * with the SGX_IOC_ENCLAVE_REMOVE_PAGES ioctl(). The latter should fail 1560 * because the removal was not EACCEPTed from within the enclave. 1561 */ 1562 TEST_F(enclave, remove_added_page_no_eaccept) 1563 { 1564 struct sgx_enclave_remove_pages remove_ioc; 1565 struct encl_op_get_from_addr get_addr_op; 1566 struct sgx_enclave_modify_types modt_ioc; 1567 struct encl_op_put_to_addr put_addr_op; 1568 unsigned long data_start; 1569 int ret, errno_save; 1570 1571 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 1572 1573 memset(&self->run, 0, sizeof(self->run)); 1574 self->run.tcs = self->encl.encl_base; 1575 1576 /* 1577 * Hardware (SGX2) and kernel support is needed for this test. Start 1578 * with check that test has a chance of succeeding. 1579 */ 1580 memset(&modt_ioc, 0, sizeof(modt_ioc)); 1581 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &modt_ioc); 1582 1583 if (ret == -1) { 1584 if (errno == ENOTTY) 1585 SKIP(return, 1586 "Kernel does not support SGX_IOC_ENCLAVE_MODIFY_TYPES ioctl()"); 1587 else if (errno == ENODEV) 1588 SKIP(return, "System does not support SGX2"); 1589 } 1590 1591 /* 1592 * Invalid parameters were provided during sanity check, 1593 * expect command to fail. 1594 */ 1595 EXPECT_EQ(ret, -1); 1596 1597 /* 1598 * Page that will be removed is the second data page in the .data 1599 * segment. This forms part of the local encl_buffer within the 1600 * enclave. 1601 */ 1602 data_start = self->encl.encl_base + 1603 encl_get_data_offset(&self->encl) + PAGE_SIZE; 1604 1605 /* 1606 * Sanity check that page at @data_start is writable before 1607 * removing it. 1608 * 1609 * Start by writing MAGIC to test page. 1610 */ 1611 put_addr_op.value = MAGIC; 1612 put_addr_op.addr = data_start; 1613 put_addr_op.header.type = ENCL_OP_PUT_TO_ADDRESS; 1614 1615 EXPECT_EQ(ENCL_CALL(&put_addr_op, &self->run, true), 0); 1616 1617 EXPECT_EEXIT(&self->run); 1618 EXPECT_EQ(self->run.exception_vector, 0); 1619 EXPECT_EQ(self->run.exception_error_code, 0); 1620 EXPECT_EQ(self->run.exception_addr, 0); 1621 1622 /* 1623 * Read memory that was just written to, confirming that data 1624 * previously written (MAGIC) is present. 1625 */ 1626 get_addr_op.value = 0; 1627 get_addr_op.addr = data_start; 1628 get_addr_op.header.type = ENCL_OP_GET_FROM_ADDRESS; 1629 1630 EXPECT_EQ(ENCL_CALL(&get_addr_op, &self->run, true), 0); 1631 1632 EXPECT_EQ(get_addr_op.value, MAGIC); 1633 EXPECT_EEXIT(&self->run); 1634 EXPECT_EQ(self->run.exception_vector, 0); 1635 EXPECT_EQ(self->run.exception_error_code, 0); 1636 EXPECT_EQ(self->run.exception_addr, 0); 1637 1638 /* Start page removal by requesting change of page type to PT_TRIM */ 1639 memset(&modt_ioc, 0, sizeof(modt_ioc)); 1640 1641 modt_ioc.offset = encl_get_data_offset(&self->encl) + PAGE_SIZE; 1642 modt_ioc.length = PAGE_SIZE; 1643 modt_ioc.page_type = SGX_PAGE_TYPE_TRIM; 1644 1645 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &modt_ioc); 1646 errno_save = ret == -1 ? errno : 0; 1647 1648 EXPECT_EQ(ret, 0); 1649 EXPECT_EQ(errno_save, 0); 1650 EXPECT_EQ(modt_ioc.result, 0); 1651 EXPECT_EQ(modt_ioc.count, 4096); 1652 1653 /* Skip EACCEPT */ 1654 1655 /* Send final ioctl() to complete page removal */ 1656 memset(&remove_ioc, 0, sizeof(remove_ioc)); 1657 1658 remove_ioc.offset = encl_get_data_offset(&self->encl) + PAGE_SIZE; 1659 remove_ioc.length = PAGE_SIZE; 1660 1661 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_REMOVE_PAGES, &remove_ioc); 1662 errno_save = ret == -1 ? errno : 0; 1663 1664 /* Operation not permitted since EACCEPT was omitted. */ 1665 EXPECT_EQ(ret, -1); 1666 EXPECT_EQ(errno_save, EPERM); 1667 EXPECT_EQ(remove_ioc.count, 0); 1668 } 1669 1670 /* 1671 * Request enclave page removal but instead of correctly following with 1672 * EACCEPT a read attempt to page is made from within the enclave. 1673 */ 1674 TEST_F(enclave, remove_added_page_invalid_access) 1675 { 1676 struct encl_op_get_from_addr get_addr_op; 1677 struct encl_op_put_to_addr put_addr_op; 1678 struct sgx_enclave_modify_types ioc; 1679 unsigned long data_start; 1680 int ret, errno_save; 1681 1682 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 1683 1684 memset(&self->run, 0, sizeof(self->run)); 1685 self->run.tcs = self->encl.encl_base; 1686 1687 /* 1688 * Hardware (SGX2) and kernel support is needed for this test. Start 1689 * with check that test has a chance of succeeding. 1690 */ 1691 memset(&ioc, 0, sizeof(ioc)); 1692 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &ioc); 1693 1694 if (ret == -1) { 1695 if (errno == ENOTTY) 1696 SKIP(return, 1697 "Kernel does not support SGX_IOC_ENCLAVE_MODIFY_TYPES ioctl()"); 1698 else if (errno == ENODEV) 1699 SKIP(return, "System does not support SGX2"); 1700 } 1701 1702 /* 1703 * Invalid parameters were provided during sanity check, 1704 * expect command to fail. 1705 */ 1706 EXPECT_EQ(ret, -1); 1707 1708 /* 1709 * Page that will be removed is the second data page in the .data 1710 * segment. This forms part of the local encl_buffer within the 1711 * enclave. 1712 */ 1713 data_start = self->encl.encl_base + 1714 encl_get_data_offset(&self->encl) + PAGE_SIZE; 1715 1716 /* 1717 * Sanity check that page at @data_start is writable before 1718 * removing it. 1719 * 1720 * Start by writing MAGIC to test page. 1721 */ 1722 put_addr_op.value = MAGIC; 1723 put_addr_op.addr = data_start; 1724 put_addr_op.header.type = ENCL_OP_PUT_TO_ADDRESS; 1725 1726 EXPECT_EQ(ENCL_CALL(&put_addr_op, &self->run, true), 0); 1727 1728 EXPECT_EEXIT(&self->run); 1729 EXPECT_EQ(self->run.exception_vector, 0); 1730 EXPECT_EQ(self->run.exception_error_code, 0); 1731 EXPECT_EQ(self->run.exception_addr, 0); 1732 1733 /* 1734 * Read memory that was just written to, confirming that data 1735 * previously written (MAGIC) is present. 1736 */ 1737 get_addr_op.value = 0; 1738 get_addr_op.addr = data_start; 1739 get_addr_op.header.type = ENCL_OP_GET_FROM_ADDRESS; 1740 1741 EXPECT_EQ(ENCL_CALL(&get_addr_op, &self->run, true), 0); 1742 1743 EXPECT_EQ(get_addr_op.value, MAGIC); 1744 EXPECT_EEXIT(&self->run); 1745 EXPECT_EQ(self->run.exception_vector, 0); 1746 EXPECT_EQ(self->run.exception_error_code, 0); 1747 EXPECT_EQ(self->run.exception_addr, 0); 1748 1749 /* Start page removal by requesting change of page type to PT_TRIM. */ 1750 memset(&ioc, 0, sizeof(ioc)); 1751 1752 ioc.offset = encl_get_data_offset(&self->encl) + PAGE_SIZE; 1753 ioc.length = PAGE_SIZE; 1754 ioc.page_type = SGX_PAGE_TYPE_TRIM; 1755 1756 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &ioc); 1757 errno_save = ret == -1 ? errno : 0; 1758 1759 EXPECT_EQ(ret, 0); 1760 EXPECT_EQ(errno_save, 0); 1761 EXPECT_EQ(ioc.result, 0); 1762 EXPECT_EQ(ioc.count, 4096); 1763 1764 /* 1765 * Read from page that was just removed. 1766 */ 1767 get_addr_op.value = 0; 1768 1769 EXPECT_EQ(ENCL_CALL(&get_addr_op, &self->run, true), 0); 1770 1771 /* 1772 * From kernel perspective the page is present but according to SGX the 1773 * page should not be accessible so a #PF with SGX bit set is 1774 * expected. 1775 */ 1776 1777 EXPECT_EQ(self->run.function, ERESUME); 1778 EXPECT_EQ(self->run.exception_vector, 14); 1779 EXPECT_EQ(self->run.exception_error_code, 0x8005); 1780 EXPECT_EQ(self->run.exception_addr, data_start); 1781 } 1782 1783 /* 1784 * Request enclave page removal and correctly follow with 1785 * EACCEPT but do not follow with removal ioctl() but instead a read attempt 1786 * to removed page is made from within the enclave. 1787 */ 1788 TEST_F(enclave, remove_added_page_invalid_access_after_eaccept) 1789 { 1790 struct encl_op_get_from_addr get_addr_op; 1791 struct encl_op_put_to_addr put_addr_op; 1792 struct sgx_enclave_modify_types ioc; 1793 struct encl_op_eaccept eaccept_op; 1794 unsigned long data_start; 1795 int ret, errno_save; 1796 1797 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 1798 1799 memset(&self->run, 0, sizeof(self->run)); 1800 self->run.tcs = self->encl.encl_base; 1801 1802 /* 1803 * Hardware (SGX2) and kernel support is needed for this test. Start 1804 * with check that test has a chance of succeeding. 1805 */ 1806 memset(&ioc, 0, sizeof(ioc)); 1807 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &ioc); 1808 1809 if (ret == -1) { 1810 if (errno == ENOTTY) 1811 SKIP(return, 1812 "Kernel does not support SGX_IOC_ENCLAVE_MODIFY_TYPES ioctl()"); 1813 else if (errno == ENODEV) 1814 SKIP(return, "System does not support SGX2"); 1815 } 1816 1817 /* 1818 * Invalid parameters were provided during sanity check, 1819 * expect command to fail. 1820 */ 1821 EXPECT_EQ(ret, -1); 1822 1823 /* 1824 * Page that will be removed is the second data page in the .data 1825 * segment. This forms part of the local encl_buffer within the 1826 * enclave. 1827 */ 1828 data_start = self->encl.encl_base + 1829 encl_get_data_offset(&self->encl) + PAGE_SIZE; 1830 1831 /* 1832 * Sanity check that page at @data_start is writable before 1833 * removing it. 1834 * 1835 * Start by writing MAGIC to test page. 1836 */ 1837 put_addr_op.value = MAGIC; 1838 put_addr_op.addr = data_start; 1839 put_addr_op.header.type = ENCL_OP_PUT_TO_ADDRESS; 1840 1841 EXPECT_EQ(ENCL_CALL(&put_addr_op, &self->run, true), 0); 1842 1843 EXPECT_EEXIT(&self->run); 1844 EXPECT_EQ(self->run.exception_vector, 0); 1845 EXPECT_EQ(self->run.exception_error_code, 0); 1846 EXPECT_EQ(self->run.exception_addr, 0); 1847 1848 /* 1849 * Read memory that was just written to, confirming that data 1850 * previously written (MAGIC) is present. 1851 */ 1852 get_addr_op.value = 0; 1853 get_addr_op.addr = data_start; 1854 get_addr_op.header.type = ENCL_OP_GET_FROM_ADDRESS; 1855 1856 EXPECT_EQ(ENCL_CALL(&get_addr_op, &self->run, true), 0); 1857 1858 EXPECT_EQ(get_addr_op.value, MAGIC); 1859 EXPECT_EEXIT(&self->run); 1860 EXPECT_EQ(self->run.exception_vector, 0); 1861 EXPECT_EQ(self->run.exception_error_code, 0); 1862 EXPECT_EQ(self->run.exception_addr, 0); 1863 1864 /* Start page removal by requesting change of page type to PT_TRIM. */ 1865 memset(&ioc, 0, sizeof(ioc)); 1866 1867 ioc.offset = encl_get_data_offset(&self->encl) + PAGE_SIZE; 1868 ioc.length = PAGE_SIZE; 1869 ioc.page_type = SGX_PAGE_TYPE_TRIM; 1870 1871 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &ioc); 1872 errno_save = ret == -1 ? errno : 0; 1873 1874 EXPECT_EQ(ret, 0); 1875 EXPECT_EQ(errno_save, 0); 1876 EXPECT_EQ(ioc.result, 0); 1877 EXPECT_EQ(ioc.count, 4096); 1878 1879 eaccept_op.epc_addr = (unsigned long)data_start; 1880 eaccept_op.ret = 0; 1881 eaccept_op.flags = SGX_SECINFO_TRIM | SGX_SECINFO_MODIFIED; 1882 eaccept_op.header.type = ENCL_OP_EACCEPT; 1883 1884 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 1885 1886 EXPECT_EEXIT(&self->run); 1887 EXPECT_EQ(self->run.exception_vector, 0); 1888 EXPECT_EQ(self->run.exception_error_code, 0); 1889 EXPECT_EQ(self->run.exception_addr, 0); 1890 EXPECT_EQ(eaccept_op.ret, 0); 1891 1892 /* Skip ioctl() to remove page. */ 1893 1894 /* 1895 * Read from page that was just removed. 1896 */ 1897 get_addr_op.value = 0; 1898 1899 EXPECT_EQ(ENCL_CALL(&get_addr_op, &self->run, true), 0); 1900 1901 /* 1902 * From kernel perspective the page is present but according to SGX the 1903 * page should not be accessible so a #PF with SGX bit set is 1904 * expected. 1905 */ 1906 1907 EXPECT_EQ(self->run.function, ERESUME); 1908 EXPECT_EQ(self->run.exception_vector, 14); 1909 EXPECT_EQ(self->run.exception_error_code, 0x8005); 1910 EXPECT_EQ(self->run.exception_addr, data_start); 1911 } 1912 1913 TEST_F(enclave, remove_untouched_page) 1914 { 1915 struct sgx_enclave_remove_pages remove_ioc; 1916 struct sgx_enclave_modify_types modt_ioc; 1917 struct encl_op_eaccept eaccept_op; 1918 unsigned long data_start; 1919 int ret, errno_save; 1920 1921 ASSERT_TRUE(setup_test_encl(ENCL_HEAP_SIZE_DEFAULT, &self->encl, _metadata)); 1922 1923 /* 1924 * Hardware (SGX2) and kernel support is needed for this test. Start 1925 * with check that test has a chance of succeeding. 1926 */ 1927 memset(&modt_ioc, 0, sizeof(modt_ioc)); 1928 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &modt_ioc); 1929 1930 if (ret == -1) { 1931 if (errno == ENOTTY) 1932 SKIP(return, 1933 "Kernel does not support SGX_IOC_ENCLAVE_MODIFY_TYPES ioctl()"); 1934 else if (errno == ENODEV) 1935 SKIP(return, "System does not support SGX2"); 1936 } 1937 1938 /* 1939 * Invalid parameters were provided during sanity check, 1940 * expect command to fail. 1941 */ 1942 EXPECT_EQ(ret, -1); 1943 1944 /* SGX2 is supported by kernel and hardware, test can proceed. */ 1945 memset(&self->run, 0, sizeof(self->run)); 1946 self->run.tcs = self->encl.encl_base; 1947 1948 data_start = self->encl.encl_base + 1949 encl_get_data_offset(&self->encl) + PAGE_SIZE; 1950 1951 memset(&modt_ioc, 0, sizeof(modt_ioc)); 1952 1953 modt_ioc.offset = encl_get_data_offset(&self->encl) + PAGE_SIZE; 1954 modt_ioc.length = PAGE_SIZE; 1955 modt_ioc.page_type = SGX_PAGE_TYPE_TRIM; 1956 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_MODIFY_TYPES, &modt_ioc); 1957 errno_save = ret == -1 ? errno : 0; 1958 1959 EXPECT_EQ(ret, 0); 1960 EXPECT_EQ(errno_save, 0); 1961 EXPECT_EQ(modt_ioc.result, 0); 1962 EXPECT_EQ(modt_ioc.count, 4096); 1963 1964 /* 1965 * Enter enclave via TCS #1 and approve page removal by sending 1966 * EACCEPT for removed page. 1967 */ 1968 1969 eaccept_op.epc_addr = data_start; 1970 eaccept_op.flags = SGX_SECINFO_TRIM | SGX_SECINFO_MODIFIED; 1971 eaccept_op.ret = 0; 1972 eaccept_op.header.type = ENCL_OP_EACCEPT; 1973 1974 EXPECT_EQ(ENCL_CALL(&eaccept_op, &self->run, true), 0); 1975 EXPECT_EEXIT(&self->run); 1976 EXPECT_EQ(self->run.exception_vector, 0); 1977 EXPECT_EQ(self->run.exception_error_code, 0); 1978 EXPECT_EQ(self->run.exception_addr, 0); 1979 EXPECT_EQ(eaccept_op.ret, 0); 1980 1981 memset(&remove_ioc, 0, sizeof(remove_ioc)); 1982 1983 remove_ioc.offset = encl_get_data_offset(&self->encl) + PAGE_SIZE; 1984 remove_ioc.length = PAGE_SIZE; 1985 ret = ioctl(self->encl.fd, SGX_IOC_ENCLAVE_REMOVE_PAGES, &remove_ioc); 1986 errno_save = ret == -1 ? errno : 0; 1987 1988 EXPECT_EQ(ret, 0); 1989 EXPECT_EQ(errno_save, 0); 1990 EXPECT_EQ(remove_ioc.count, 4096); 1991 } 1992 1993 TEST_HARNESS_MAIN 1994