Lines Matching +full:0 +full:x33

44 		"	sllg 1,1,12(0)\n"  in guest_do_one_essa()
46 " .insn rrf,0xb9ab0000,2,1,1,0\n" in guest_do_one_essa()
48 " diag 0,0,0x501\n" in guest_do_one_essa()
49 "0: j 0b" in guest_do_one_essa()
71 "1: sllg 2,1,12(0)\n" in guest_dirty_test_data()
73 " .insn rrf,0xb9ab0000,4,2,1,0\n" in guest_dirty_test_data()
79 " diag 0,0,0x501\n" in guest_dirty_test_data()
80 "0: j 0b" in guest_dirty_test_data()
101 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, 0, 0, MAIN_PAGE_COUNT, 0); in create_main_memslot()
103 for (i = 0; i < NR_MEM_REGIONS; i++) in create_main_memslot()
104 vm->memslots[i] = 0; in create_main_memslot()
114 0 in create_test_memslot()
126 * | 0 | | in create_memslots()
145 slot0 = memslot2region(vm, 0); in finish_vm_setup()
174 vm_mem_region_set_flags(vm, 0, KVM_MEM_LOG_DIRTY_PAGES); in enable_dirty_tracking()
213 errno = 0; in vm_get_cmma_bits()
216 .start_gfn = 0, in vm_get_cmma_bits()
219 .values = (__u64)&cmma_value_buf[0] in vm_get_cmma_bits()
234 rc = vm_get_cmma_bits(vm, 0, &errno_out); in test_get_cmma_basic()
244 rc = vm_get_cmma_bits(vm, 0, &errno_out); in test_get_cmma_basic()
250 TEST_ASSERT_EQ(rc, 0); in test_get_cmma_basic()
251 TEST_ASSERT_EQ(errno_out, 0); in test_get_cmma_basic()
257 rc = vm_get_cmma_bits(vm, 0xfeedc0fe, &errno_out); in test_get_cmma_basic()
268 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipa, 0x8300); in assert_exit_was_hypercall()
269 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipb, 0x5010000); in assert_exit_was_hypercall()
284 errno = 0; in test_migration_mode()
305 errno = 0; in test_migration_mode()
312 TEST_ASSERT_EQ(rc, 0); in test_migration_mode()
314 errno = 0; in test_migration_mode()
331 0 in test_migration_mode()
348 TEST_ASSERT_EQ(rc, 0); in test_migration_mode()
350 errno = 0; in test_migration_mode()
357 vm_mem_region_set_flags(vm, TEST_DATA_TWO_MEMSLOT, 0); in test_migration_mode()
385 .start_gfn = 0, in assert_all_slots_cmma_dirty()
387 .flags = 0, in assert_all_slots_cmma_dirty()
388 .values = (__u64)&cmma_value_buf[0] in assert_all_slots_cmma_dirty()
390 memset(cmma_value_buf, 0xff, sizeof(cmma_value_buf)); in assert_all_slots_cmma_dirty()
394 TEST_ASSERT_EQ(args.start_gfn, 0); in assert_all_slots_cmma_dirty()
400 .flags = 0, in assert_all_slots_cmma_dirty()
401 .values = (__u64)&cmma_value_buf[0] in assert_all_slots_cmma_dirty()
403 memset(cmma_value_buf, 0xff, sizeof(cmma_value_buf)); in assert_all_slots_cmma_dirty()
407 TEST_ASSERT_EQ(args.remaining, 0); in assert_all_slots_cmma_dirty()
413 .flags = 0, in assert_all_slots_cmma_dirty()
414 .values = (__u64)&cmma_value_buf[0] in assert_all_slots_cmma_dirty()
416 memset(cmma_value_buf, 0xff, sizeof(cmma_value_buf)); in assert_all_slots_cmma_dirty()
418 TEST_ASSERT_EQ(args.count, 0); in assert_all_slots_cmma_dirty()
419 TEST_ASSERT_EQ(args.start_gfn, 0); in assert_all_slots_cmma_dirty()
420 TEST_ASSERT_EQ(args.remaining, 0); in assert_all_slots_cmma_dirty()
430 /* If we start from GFN 0 again, nothing should be dirty. */ in assert_no_pages_cmma_dirty()
432 .start_gfn = 0, in assert_no_pages_cmma_dirty()
434 .flags = 0, in assert_no_pages_cmma_dirty()
435 .values = (__u64)&cmma_value_buf[0] in assert_no_pages_cmma_dirty()
437 memset(cmma_value_buf, 0xff, sizeof(cmma_value_buf)); in assert_no_pages_cmma_dirty()
440 TEST_FAIL("pages are still dirty start_gfn=0x%llx count=%u remaining=%llu", in assert_no_pages_cmma_dirty()
480 .flags = 0, in query_cmma_range()
481 .values = (__u64)&cmma_value_buf[0] in query_cmma_range()
483 memset(cmma_value_buf, 0xff, sizeof(cmma_value_buf)); in query_cmma_range()
498 for (size_t i = 0; i < dirty_gfn_count; i++) in assert_cmma_dirty()
499 TEST_ASSERT_EQ(cmma_value_buf[0], 0x0); /* stable state */ in assert_cmma_dirty()
500 TEST_ASSERT_EQ(cmma_value_buf[dirty_gfn_count], 0xff); /* not touched */ in assert_cmma_dirty()
535 * Query CMMA attributes of one page, starting at page 0. Since the in test_get_skip_holes()
539 * 0: not dirty in test_get_skip_holes()
540 * [0x1, 0x200): dirty in test_get_skip_holes()
542 query_cmma_range(vm, 0, 1, &log); in test_get_skip_holes()
547 * Query CMMA attributes of 32 (0x20) pages past the end of the TEST_DATA in test_get_skip_holes()
551 * [0, 0x21): not dirty in test_get_skip_holes()
552 * [0x21, 0x200): dirty in test_get_skip_holes()
554 query_cmma_range(vm, TEST_DATA_START_GFN + TEST_DATA_PAGE_COUNT, 0x20, &log); in test_get_skip_holes()
555 assert_cmma_dirty(gfn_offset, 0x20, &log); in test_get_skip_holes()
556 gfn_offset += 0x20; in test_get_skip_holes()
559 gfn_offset += 0x20; in test_get_skip_holes()
562 * After skipping 32 pages, query the next 32 (0x20) pages. in test_get_skip_holes()
564 * [0, 0x21): not dirty in test_get_skip_holes()
565 * [0x21, 0x41): dirty in test_get_skip_holes()
566 * [0x41, 0x61): not dirty in test_get_skip_holes()
567 * [0x61, 0x200): dirty in test_get_skip_holes()
569 query_cmma_range(vm, gfn_offset, 0x20, &log); in test_get_skip_holes()
570 assert_cmma_dirty(gfn_offset, 0x20, &log); in test_get_skip_holes()
571 gfn_offset += 0x20; in test_get_skip_holes()
575 * yield page 0x21. in test_get_skip_holes()
577 * [0, 0x22): not dirty in test_get_skip_holes()
578 * [0x22, 0x41): dirty in test_get_skip_holes()
579 * [0x41, 0x61): not dirty in test_get_skip_holes()
580 * [0x61, 0x200): dirty in test_get_skip_holes()
583 assert_cmma_dirty(TEST_DATA_START_GFN + 0x21, 1, &log); in test_get_skip_holes()
587 * Query 15 (0xF) pages from page 0x23 in TEST_DATA memslot. in test_get_skip_holes()
588 * This should yield pages [0x23, 0x33). in test_get_skip_holes()
590 * [0, 0x22): not dirty in test_get_skip_holes()
591 * 0x22: dirty in test_get_skip_holes()
592 * [0x23, 0x33): not dirty in test_get_skip_holes()
593 * [0x33, 0x41): dirty in test_get_skip_holes()
594 * [0x41, 0x61): not dirty in test_get_skip_holes()
595 * [0x61, 0x200): dirty in test_get_skip_holes()
597 gfn_offset = TEST_DATA_START_GFN + 0x23; in test_get_skip_holes()
602 * Query 17 (0x11) pages from page 0x22 in TEST_DATA memslot. in test_get_skip_holes()
603 * This should yield page [0x22, 0x33) in test_get_skip_holes()
605 * [0, 0x33): not dirty in test_get_skip_holes()
606 * [0x33, 0x41): dirty in test_get_skip_holes()
607 * [0x41, 0x61): not dirty in test_get_skip_holes()
608 * [0x61, 0x200): dirty in test_get_skip_holes()
610 gfn_offset = TEST_DATA_START_GFN + 0x22; in test_get_skip_holes()
615 * Query 25 (0x19) pages from page 0x40 in TEST_DATA memslot. in test_get_skip_holes()
616 * This should yield page 0x40 and nothing more, since there are more in test_get_skip_holes()
617 * than 16 non-dirty pages after page 0x40. in test_get_skip_holes()
619 * [0, 0x33): not dirty in test_get_skip_holes()
620 * [0x33, 0x40): dirty in test_get_skip_holes()
621 * [0x40, 0x61): not dirty in test_get_skip_holes()
622 * [0x61, 0x200): dirty in test_get_skip_holes()
624 gfn_offset = TEST_DATA_START_GFN + 0x40; in test_get_skip_holes()
629 * Query pages [0x33, 0x40). in test_get_skip_holes()
631 * [0, 0x61): not dirty in test_get_skip_holes()
632 * [0x61, 0x200): dirty in test_get_skip_holes()
634 gfn_offset = TEST_DATA_START_GFN + 0x33; in test_get_skip_holes()
635 query_cmma_range(vm, gfn_offset, 0x40 - 0x33, &log); in test_get_skip_holes()
636 assert_cmma_dirty(gfn_offset, 0x40 - 0x33, &log); in test_get_skip_holes()
639 * Query the remaining pages [0x61, 0x200). in test_get_skip_holes()
642 query_cmma_range(vm, gfn_offset, TEST_DATA_PAGE_COUNT - 0x61, &log); in test_get_skip_holes()
643 assert_cmma_dirty(TEST_DATA_START_GFN + 0x61, TEST_DATA_PAGE_COUNT - 0x61, &log); in test_get_skip_holes()
689 for (idx = 0; idx < ARRAY_SIZE(testlist); idx++) { in main()