Lines Matching refs:evcn

65 	CLST evcn = le64_to_cpu(attr->nres.evcn);  in attr_load_runs()  local
69 if (svcn >= evcn + 1 || run_is_mapped_full(run, svcn, evcn)) in attr_load_runs()
72 if (vcn && (evcn < *vcn || *vcn < svcn)) in attr_load_runs()
81 err = run_unpack_ex(run, ni->mi.sbi, ni->mi.rno, svcn, evcn, in attr_load_runs()
416 CLST alen, vcn, lcn, new_alen, old_alen, svcn, evcn; in attr_set_size() local
473 evcn = le64_to_cpu(attr_b->nres.evcn); in attr_set_size()
475 if (svcn <= vcn && vcn <= evcn) { in attr_set_size()
493 evcn = le64_to_cpu(attr->nres.evcn); in attr_set_size()
599 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_set_size()
620 evcn = old_alen - 1; in attr_set_size()
677 run_truncate_head(run, evcn + 1); in attr_set_size()
680 evcn = le64_to_cpu(attr->nres.evcn); in attr_set_size()
730 attr->nres.evcn = cpu_to_le64((u64)vcn - 1); in attr_set_size()
753 err = run_deallocate_ex(sbi, run, vcn, evcn - vcn + 1, &dlen, in attr_set_size()
775 evcn = svcn - 1; in attr_set_size()
843 svcn <= le64_to_cpu(attr_b->nres.evcn)) { in attr_set_size()
859 if (mi_pack_runs(mi, attr, run, evcn - svcn + 1)) in attr_set_size()
896 CLST alloc, evcn; in attr_data_get_block() local
956 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_data_get_block()
970 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_data_get_block()
1040 evcn1 = le64_to_cpu(attr2->nres.evcn) + 1; in attr_data_get_block()
1106 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_data_get_block()
1160 evcn = le64_to_cpu(attr->nres.evcn); in attr_data_get_block()
1164 while (end > evcn) { in attr_data_get_block()
1173 if (evcn + 1 >= alloc) { in attr_data_get_block()
1175 evcn1 = evcn + 1; in attr_data_get_block()
1190 evcn = le64_to_cpu(attr->nres.evcn); in attr_data_get_block()
1200 evcn1 = evcn + 1; in attr_data_get_block()
1209 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_data_get_block()
1310 CLST svcn, evcn; in attr_load_runs_vcn() local
1325 evcn = le64_to_cpu(attr->nres.evcn); in attr_load_runs_vcn()
1327 if (evcn < vcn || vcn < svcn) { in attr_load_runs_vcn()
1337 err = run_unpack_ex(run, ni->mi.sbi, ni->mi.rno, svcn, evcn, svcn, in attr_load_runs_vcn()
1655 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_allocate_frame()
1674 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_allocate_frame()
1740 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_allocate_frame()
1776 CLST evcn = le64_to_cpu(attr->nres.evcn); in attr_allocate_frame() local
1780 while (end > evcn) { in attr_allocate_frame()
1789 if (evcn + 1 >= alloc) { in attr_allocate_frame()
1791 evcn1 = evcn + 1; in attr_allocate_frame()
1807 evcn = le64_to_cpu(attr->nres.evcn); in attr_allocate_frame()
1817 evcn1 = evcn + 1; in attr_allocate_frame()
1827 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_allocate_frame()
1930 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_collapse_range()
1949 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_collapse_range()
1956 attr->nres.evcn = cpu_to_le64(evcn1 - 1 - len); in attr_collapse_range()
1995 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_collapse_range()
2067 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_collapse_range()
2192 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_punch_hole()
2212 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_punch_hole()
2252 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_punch_hole()
2284 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_punch_hole()
2431 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_insert_range()
2450 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_insert_range()
2468 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_insert_range()
2473 le64_add_cpu(&attr->nres.evcn, len); in attr_insert_range()
2534 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_insert_range()
2551 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_insert_range()
2566 le64_sub_cpu(&attr->nres.evcn, len); in attr_insert_range()