/linux-6.12.1/fs/xfs/ |
D | xfs_extent_busy.c | 159 xfs_agblock_t fend = fbno + flen; in xfs_extent_busy_update_extent() local 186 if (bbno < fbno && bend > fend) { in xfs_extent_busy_update_extent() 205 } else if (bbno >= fbno && bend <= fend) { in xfs_extent_busy_update_extent() 247 } else if (fend < bend) { in xfs_extent_busy_update_extent() 262 busyp->bno = fend; in xfs_extent_busy_update_extent() 263 busyp->length = bend - fend; in xfs_extent_busy_update_extent() 366 xfs_agblock_t fend = fbno + flen; in xfs_extent_busy_trim() local 370 if (fend <= bbno) { in xfs_extent_busy_trim() 408 if (fend <= bend) in xfs_extent_busy_trim() 429 } else if (bend >= fend) { in xfs_extent_busy_trim() [all …]
|
/linux-6.12.1/mm/ |
D | msync.c | 66 loff_t fstart, fend; in SYSCALL_DEFINE3() local 90 fend = fstart + (min(end, vma->vm_end) - start) - 1; in SYSCALL_DEFINE3() 96 error = vfs_fsync_range(file, fstart, fend, 1); in SYSCALL_DEFINE3()
|
/linux-6.12.1/fs/netfs/ |
D | misc.c | 253 unsigned int fend = fstart + finfo->dirty_len; in netfs_invalidate_folio() local 256 if (offset >= fend) in netfs_invalidate_folio() 266 if (iend >= fend) in netfs_invalidate_folio() 269 finfo->dirty_len = fend - iend; in netfs_invalidate_folio() 277 if (iend >= fend) { in netfs_invalidate_folio()
|
D | read_pgpriv2.c | 218 unsigned long long fpos, fend; in netfs_pgpriv2_unlock_copied_folios() local 231 fend = min_t(unsigned long long, fpos + flen, wreq->i_size); in netfs_pgpriv2_unlock_copied_folios() 233 trace_netfs_collect_folio(wreq, folio, fend, collected_to); in netfs_pgpriv2_unlock_copied_folios() 236 if (collected_to < fend) in netfs_pgpriv2_unlock_copied_folios()
|
D | write_collect.c | 41 unsigned long long fend; in netfs_folio_written_back() local 43 fend = folio_pos(folio) + finfo->dirty_offset + finfo->dirty_len; in netfs_folio_written_back() 44 if (fend > ictx->zero_point) in netfs_folio_written_back() 45 ictx->zero_point = fend; in netfs_folio_written_back() 104 unsigned long long fpos, fend; in netfs_writeback_unlock_folios() local 118 fend = min_t(unsigned long long, fpos + flen, wreq->i_size); in netfs_writeback_unlock_folios() 120 trace_netfs_collect_folio(wreq, folio, fend, collected_to); in netfs_writeback_unlock_folios() 123 if (collected_to < fend) in netfs_writeback_unlock_folios()
|
D | read_collect.c | 95 loff_t fend; in netfs_consume_read_data() local 107 fend = fpos + fsize; in netfs_consume_read_data() 125 fpos, fend - 1, folio_pos(folio), folio_order(folio), in netfs_consume_read_data() 173 if (start + avail >= fend) { in netfs_consume_read_data() 176 subreq->consumed = fend - subreq->start; in netfs_consume_read_data() 182 excess = fend - subreq->start; in netfs_consume_read_data() 309 printk("folio: %llx-%llx\n", fpos, fend - 1); in netfs_consume_read_data()
|
/linux-6.12.1/drivers/net/ethernet/netronome/nfp/nfdk/ |
D | rings.c | 17 const skb_frag_t *frag, *fend; in nfp_nfdk_tx_ring_reset() local 42 fend = frag + nr_frags; in nfp_nfdk_tx_ring_reset() 43 for (; frag < fend; frag++) { in nfp_nfdk_tx_ring_reset()
|
D | dp.c | 123 const skb_frag_t *frag, *fend; in nfp_nfdk_tx_maybe_close_block() local 133 fend = frag + nr_frags; in nfp_nfdk_tx_maybe_close_block() 134 for (; frag < fend; frag++) in nfp_nfdk_tx_maybe_close_block() 260 const skb_frag_t *frag, *fend; in nfp_nfdk_tx() local 355 fend = frag + nr_frags; in nfp_nfdk_tx() 370 if (frag >= fend) in nfp_nfdk_tx() 496 const skb_frag_t *frag, *fend; in nfp_nfdk_tx_complete() local 519 fend = frag + skb_shinfo(skb)->nr_frags; in nfp_nfdk_tx_complete() 520 for (; frag < fend; frag++) { in nfp_nfdk_tx_complete()
|
/linux-6.12.1/fs/smb/client/ |
D | cifsfs.c | 1207 unsigned long long fpos, fend; in cifs_flush_folio() local 1218 fend = fpos + size - 1; in cifs_flush_folio() 1220 *_fend = max_t(unsigned long long, *_fend, fend); in cifs_flush_folio() 1221 if ((first && pos == fpos) || (!first && pos == fend)) in cifs_flush_folio() 1224 rc = filemap_write_and_wait_range(inode->i_mapping, fpos, fend); in cifs_flush_folio() 1241 unsigned long long destend, fstart, fend, old_size, new_size; in cifs_remap_file_range() local 1298 fend = destend; in cifs_remap_file_range() 1300 rc = cifs_flush_folio(target_inode, destoff, &fstart, &fend, true); in cifs_remap_file_range() 1303 rc = cifs_flush_folio(target_inode, destend, &fstart, &fend, false); in cifs_remap_file_range() 1306 if (fend > target_cifsi->netfs.zero_point) in cifs_remap_file_range() [all …]
|
/linux-6.12.1/arch/alpha/kernel/ |
D | pci_iommu.c | 725 dma_addr_t fbeg, fend; in alpha_pci_unmap_sg() local 738 fbeg = -1, fend = 0; in alpha_pci_unmap_sg() 777 if (fend < tend) fend = tend; in alpha_pci_unmap_sg() 783 if ((fend - arena->dma_base) >> PAGE_SHIFT >= arena->next_entry) in alpha_pci_unmap_sg() 784 alpha_mv.mv_pci_tbi(hose, fbeg, fend); in alpha_pci_unmap_sg()
|
/linux-6.12.1/include/trace/events/ |
D | netfs.h | 576 unsigned long long fend, 579 TP_ARGS(wreq, folio, fend, collected_to), 584 __field(unsigned long long, fend ) 592 __entry->fend = fend; 599 (unsigned long long)__entry->index * PAGE_SIZE, __entry->fend,
|
/linux-6.12.1/arch/powerpc/lib/ |
D | feature-fixups.c | 111 struct fixup_entry *fcur, *fend; in do_feature_fixups_mask() local 114 fend = fixup_end; in do_feature_fixups_mask() 116 for (; fcur < fend; fcur++) { in do_feature_fixups_mask()
|
/linux-6.12.1/drivers/macintosh/ |
D | smu.c | 113 unsigned long faddr, fend; in smu_start_cmd() local 135 fend = faddr + smu->cmd_buf->length + 2; in smu_start_cmd() 136 flush_dcache_range(faddr, fend); in smu_start_cmd()
|
/linux-6.12.1/kernel/cgroup/ |
D | cgroup.c | 7024 int fstart = 0, fend = 0, flen; in cgroup_parse_float() local 7026 if (!sscanf(input, "%lld.%n%lld%n", &whole, &fstart, &frac, &fend)) in cgroup_parse_float() 7031 flen = fend > fstart ? fend - fstart : 0; in cgroup_parse_float()
|