/linux-6.12.1/arch/powerpc/mm/ |
D | dma-noncoherent.c | 62 size_t seg_size = min((size_t)(PAGE_SIZE - offset), size); in __dma_sync_page_highmem() local 63 size_t cur_size = seg_size; in __dma_sync_page_highmem() 65 int nr_segs = 1 + ((size - seg_size) + PAGE_SIZE - 1)/PAGE_SIZE; in __dma_sync_page_highmem() 74 __dma_sync((void *)start, seg_size, direction); in __dma_sync_page_highmem() 79 seg_size = min((size_t)PAGE_SIZE, size - cur_size); in __dma_sync_page_highmem() 82 cur_size += seg_size; in __dma_sync_page_highmem()
|
/linux-6.12.1/drivers/gpu/drm/amd/amdgpu/ |
D | gfxhub_v1_1.c | 50 u64 seg_size; in gfxhub_v1_1_get_xgmi_info() local 54 seg_size = REG_GET_FIELD( in gfxhub_v1_1_get_xgmi_info() 61 seg_size = REG_GET_FIELD( in gfxhub_v1_1_get_xgmi_info() 107 adev->gmc.xgmi.node_segment_size = seg_size; in gfxhub_v1_1_get_xgmi_info()
|
D | gfxhub_v1_2.c | 591 u64 seg_size; in gfxhub_v1_2_get_xgmi_info() local 594 seg_size = REG_GET_FIELD( in gfxhub_v1_2_get_xgmi_info() 619 adev->gmc.xgmi.node_segment_size = seg_size; in gfxhub_v1_2_get_xgmi_info()
|
/linux-6.12.1/drivers/mmc/core/ |
D | sdio_ops.c | 123 unsigned int seg_size = card->host->max_seg_size; in mmc_io_rw_extended() local 152 nents = DIV_ROUND_UP(left_size, seg_size); in mmc_io_rw_extended() 161 sg_set_buf(sg_ptr, buf + i * seg_size, in mmc_io_rw_extended() 162 min(seg_size, left_size)); in mmc_io_rw_extended() 163 left_size -= seg_size; in mmc_io_rw_extended()
|
/linux-6.12.1/tools/testing/selftests/kvm/lib/ |
D | elf.c | 163 size_t seg_size = seg_vend - seg_vstart + 1; in kvm_vm_elf_load() local 165 vm_vaddr_t vaddr = __vm_vaddr_alloc(vm, seg_size, seg_vstart, in kvm_vm_elf_load() 173 memset(addr_gva2hva(vm, vaddr), 0, seg_size); in kvm_vm_elf_load()
|
/linux-6.12.1/drivers/bluetooth/ |
D | hci_aml.c | 203 u32 seg_size, in aml_send_firmware_segment() argument 214 op_addr, (u32 *)seg, seg_size); in aml_send_firmware_segment() 220 u32 seg_size = 0; in aml_send_firmware() local 230 seg_size = (fw_size > AML_FIRMWARE_OPERATION_SIZE) ? in aml_send_firmware() 233 seg_size, offset)) { in aml_send_firmware() 238 seg_off += seg_size; in aml_send_firmware() 239 fw_size -= seg_size; in aml_send_firmware() 240 offset += seg_size; in aml_send_firmware()
|
D | btqca.c | 444 static int qca_tlv_send_segment(struct hci_dev *hdev, int seg_size, in qca_tlv_send_segment() argument 458 cmd[1] = seg_size; in qca_tlv_send_segment() 459 memcpy(cmd + 2, data, seg_size); in qca_tlv_send_segment() 462 return __hci_cmd_send(hdev, EDL_PATCH_CMD_OPCODE, seg_size + 2, in qca_tlv_send_segment() 475 skb = __hci_cmd_sync_ev(hdev, EDL_PATCH_CMD_OPCODE, seg_size + 2, cmd, in qca_tlv_send_segment()
|
/linux-6.12.1/drivers/net/wireless/ath/ath10k/ |
D | swap.c | 76 u32 seg_size; in ath10k_swap_code_seg_free() local 84 seg_size = __le32_to_cpu(seg_info->seg_hw_info.size); in ath10k_swap_code_seg_free() 85 dma_free_coherent(ar->dev, seg_size, seg_info->virt_address[0], in ath10k_swap_code_seg_free()
|
/linux-6.12.1/block/ |
D | blk-integrity.c | 31 unsigned int seg_size = 0; in blk_rq_count_integrity_sg() local 40 if (seg_size + iv.bv_len > queue_max_segment_size(q)) in blk_rq_count_integrity_sg() 43 seg_size += iv.bv_len; in blk_rq_count_integrity_sg() 47 seg_size = iv.bv_len; in blk_rq_count_integrity_sg()
|
D | blk-merge.c | 280 unsigned seg_size = 0; in bvec_split_segs() local 283 seg_size = get_max_segment_size(lim, bvec_phys(bv) + total_len, len); in bvec_split_segs() 286 total_len += seg_size; in bvec_split_segs() 287 len -= seg_size; in bvec_split_segs()
|
/linux-6.12.1/arch/x86/kernel/ |
D | amd_gart_64.c | 383 unsigned int seg_size; in gart_map_sg() local 393 seg_size = 0; in gart_map_sg() 413 (s->length + seg_size > max_seg_size) || in gart_map_sg() 421 seg_size = 0; in gart_map_sg() 429 seg_size += s->length; in gart_map_sg()
|
/linux-6.12.1/drivers/infiniband/core/ |
D | user_mad.c | 309 size_t seg_size; in copy_recv_mad() local 312 seg_size = packet->recv_wc->mad_seg_size; in copy_recv_mad() 315 if ((packet->length <= seg_size && in copy_recv_mad() 317 (packet->length > seg_size && in copy_recv_mad() 318 count < hdr_size(file) + seg_size)) in copy_recv_mad() 325 seg_payload = min_t(int, packet->length, seg_size); in copy_recv_mad() 342 max_seg_payload = seg_size - offset; in copy_recv_mad() 454 seg++, left -= msg->seg_size, buf += msg->seg_size) { in copy_rmpp_mad() 456 min(left, msg->seg_size))) in copy_rmpp_mad()
|
D | mad.c | 768 int seg_size, pad; in get_pad_size() local 770 seg_size = mad_size - hdr_len; in get_pad_size() 771 if (data_len && seg_size) { in get_pad_size() 772 pad = seg_size - data_len % seg_size; in get_pad_size() 773 return pad == seg_size ? 0 : pad; in get_pad_size() 775 return seg_size; in get_pad_size() 794 int left, seg_size, pad; in alloc_send_rmpp_list() local 796 send_buf->seg_size = mad_size - send_buf->hdr_len; in alloc_send_rmpp_list() 798 seg_size = send_buf->seg_size; in alloc_send_rmpp_list() 802 for (left = send_buf->data_len + pad; left > 0; left -= seg_size) { in alloc_send_rmpp_list() [all …]
|
/linux-6.12.1/arch/powerpc/kvm/ |
D | book3s_64_mmu.c | 452 u64 seg_size; in kvmppc_mmu_book3s_64_slbie() local 467 seg_size = 1ull << kvmppc_slb_sid_shift(slbe); in kvmppc_mmu_book3s_64_slbie() 468 kvmppc_mmu_flush_segment(vcpu, ea & ~(seg_size - 1), seg_size); in kvmppc_mmu_book3s_64_slbie()
|
D | book3s_64_mmu_host.c | 358 void kvmppc_mmu_flush_segment(struct kvm_vcpu *vcpu, ulong ea, ulong seg_size) in kvmppc_mmu_flush_segment() argument 361 ulong seg_mask = -seg_size; in kvmppc_mmu_flush_segment()
|
/linux-6.12.1/fs/afs/ |
D | xdr_fs.h | 25 __be32 seg_size; member
|
/linux-6.12.1/drivers/bus/mhi/host/ |
D | boot.c | 317 size_t seg_size = mhi_cntrl->seg_len; in mhi_alloc_bhie_table() local 318 int segments = DIV_ROUND_UP(alloc_size, seg_size) + 1; in mhi_alloc_bhie_table() 336 size_t vec_size = seg_size; in mhi_alloc_bhie_table()
|
/linux-6.12.1/drivers/usb/early/ |
D | xhci-dbc.h | 61 __le32 seg_size; member
|
/linux-6.12.1/drivers/video/fbdev/via/ |
D | via-core.c | 150 u32 seg_size; /* Size, 16-byte units */ member 260 descr->seg_size = sg_dma_len(sgentry) >> 4; in viafb_dma_copy_out_sg()
|
/linux-6.12.1/drivers/net/ethernet/marvell/octeontx2/nic/ |
D | otx2_txrx.c | 348 u16 *seg_size; in otx2_rcv_pkt_handler() local 369 seg_size = (void *)sg; in otx2_rcv_pkt_handler() 372 seg_size[seg], parse, cq->cq_idx)) in otx2_rcv_pkt_handler() 1428 cqe->sg.seg_size, false); in otx2_xdp_rcv_pkt_handler() 1439 cqe->sg.seg_size, qidx); in otx2_xdp_rcv_pkt_handler()
|
D | otx2_struct.h | 150 u64 seg_size : 16; /* W0 */ member
|
/linux-6.12.1/drivers/net/ethernet/intel/ice/ |
D | ice_ddp.h | 118 __le32 seg_size; member
|
/linux-6.12.1/include/rdma/ |
D | ib_mad.h | 474 int seg_size; member
|
/linux-6.12.1/drivers/usb/mtu3/ |
D | mtu3_core.c | 23 static int ep_fifo_alloc(struct mtu3_ep *mep, u32 seg_size) in ep_fifo_alloc() argument 26 u32 num_bits = DIV_ROUND_UP(seg_size, MTU3_EP_FIFO_UNIT); in ep_fifo_alloc()
|
/linux-6.12.1/net/ipv4/ |
D | tcp_output.c | 4333 unsigned int seg_size = tcp_wnd_end(tp) - TCP_SKB_CB(skb)->seq; in tcp_write_wakeup() local 4342 if (seg_size < TCP_SKB_CB(skb)->end_seq - TCP_SKB_CB(skb)->seq || in tcp_write_wakeup() 4344 seg_size = min(seg_size, mss); in tcp_write_wakeup() 4347 skb, seg_size, mss, GFP_ATOMIC)) in tcp_write_wakeup()
|