Lines Matching refs:cur_gfn

682 	gfn_t cur_gfn, last_gfn;  in kvm_arch_sync_dirty_log()  local
688 cur_gfn = memslot->base_gfn; in kvm_arch_sync_dirty_log()
690 for (; cur_gfn <= last_gfn; cur_gfn += _PAGE_ENTRIES) { in kvm_arch_sync_dirty_log()
691 gaddr = gfn_to_gpa(cur_gfn); in kvm_arch_sync_dirty_log()
692 vmaddr = gfn_to_hva_memslot(memslot, cur_gfn); in kvm_arch_sync_dirty_log()
700 mark_page_dirty(kvm, cur_gfn + i); in kvm_arch_sync_dirty_log()
2204 unsigned long pgstev, hva, cur_gfn = args->start_gfn; in kvm_s390_peek_cmma() local
2208 hva = gfn_to_hva(kvm, cur_gfn); in kvm_s390_peek_cmma()
2218 cur_gfn++; in kvm_s390_peek_cmma()
2231 unsigned long cur_gfn) in kvm_s390_next_dirty_cmma() argument
2233 struct kvm_memory_slot *ms = gfn_to_memslot_approx(slots, cur_gfn); in kvm_s390_next_dirty_cmma()
2234 unsigned long ofs = cur_gfn - ms->base_gfn; in kvm_s390_next_dirty_cmma()
2237 if (ms->base_gfn + ms->npages <= cur_gfn) { in kvm_s390_next_dirty_cmma()
2247 if (cur_gfn < ms->base_gfn) in kvm_s390_next_dirty_cmma()
2261 unsigned long mem_end, cur_gfn, next_gfn, hva, pgstev; in kvm_s390_get_cmma() local
2268 cur_gfn = kvm_s390_next_dirty_cmma(slots, args->start_gfn); in kvm_s390_get_cmma()
2269 ms = gfn_to_memslot(kvm, cur_gfn); in kvm_s390_get_cmma()
2271 args->start_gfn = cur_gfn; in kvm_s390_get_cmma()
2274 next_gfn = kvm_s390_next_dirty_cmma(slots, cur_gfn + 1); in kvm_s390_get_cmma()
2278 hva = gfn_to_hva(kvm, cur_gfn); in kvm_s390_get_cmma()
2282 if (test_and_clear_bit(cur_gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms))) in kvm_s390_get_cmma()
2289 if (next_gfn > cur_gfn + KVM_S390_MAX_BIT_DISTANCE) in kvm_s390_get_cmma()
2292 if (cur_gfn == next_gfn) in kvm_s390_get_cmma()
2293 next_gfn = kvm_s390_next_dirty_cmma(slots, cur_gfn + 1); in kvm_s390_get_cmma()
2298 cur_gfn++; in kvm_s390_get_cmma()
2300 if (cur_gfn - ms->base_gfn >= ms->npages) { in kvm_s390_get_cmma()
2301 ms = gfn_to_memslot(kvm, cur_gfn); in kvm_s390_get_cmma()