/linux-6.12.1/arch/s390/mm/ |
D | extmem.c | 165 query_segment_type (struct dcss_segment *seg) in query_segment_type() argument 183 memcpy (qin->qname, seg->dcss_name, 8); in query_segment_type() 203 seg->vm_segtype = qout->range[0].start & 0xff; in query_segment_type() 222 seg->vm_segtype = SEG_TYPE_EWEN; in query_segment_type() 225 /* analyze diag output and update seg */ in query_segment_type() 226 seg->start_addr = qout->segstart; in query_segment_type() 227 seg->end = qout->segend; in query_segment_type() 229 memcpy (seg->range, qout->range, 6*sizeof(struct qrange)); in query_segment_type() 230 seg->segcnt = qout->segcnt; in query_segment_type() 254 struct dcss_segment seg; in segment_type() local [all …]
|
/linux-6.12.1/drivers/infiniband/hw/mlx5/ |
D | wr.c | 55 void **seg, int *size, void **cur_edge) in set_eth_seg() argument 57 struct mlx5_wqe_eth_seg *eseg = *seg; in set_eth_seg() 85 *seg += stride; in set_eth_seg() 88 handle_post_send_edge(&qp->sq, seg, *size, cur_edge); in set_eth_seg() 91 mlx5r_memcpy_send_wqe(&qp->sq, cur_edge, seg, size, in set_eth_seg() 98 *seg += sizeof(struct mlx5_wqe_eth_seg); in set_eth_seg() 189 static void set_reg_mkey_seg(struct mlx5_mkey_seg *seg, in set_reg_mkey_seg() argument 195 memset(seg, 0, sizeof(*seg)); in set_reg_mkey_seg() 198 seg->log2_page_size = ilog2(mr->ibmr.page_size); in set_reg_mkey_seg() 203 seg->flags = get_umr_flags(access) | mr->access_mode; in set_reg_mkey_seg() [all …]
|
D | wr.h | 47 * @seg: Current WQE position (16B aligned). 51 static inline void handle_post_send_edge(struct mlx5_ib_wq *sq, void **seg, in handle_post_send_edge() argument 56 if (likely(*seg != *cur_edge)) in handle_post_send_edge() 62 *seg = mlx5_frag_buf_get_wqe(&sq->fbc, idx); in handle_post_send_edge() 66 * WQ's pointers. At the end @seg is aligned to 16B regardless the copied size. 69 * @seg: Current WQE position (16B aligned). 75 void **seg, u32 *wqe_sz, in mlx5r_memcpy_send_wqe() argument 79 size_t leftlen = *cur_edge - *seg; in mlx5r_memcpy_send_wqe() 83 memcpy(*seg, src, copysz); in mlx5r_memcpy_send_wqe() 88 *seg += stride; in mlx5r_memcpy_send_wqe() [all …]
|
/linux-6.12.1/ipc/ |
D | msgutil.c | 74 struct msg_msgseg *seg; in alloc_msg() local 79 seg = kmalloc(sizeof(*seg) + alen, GFP_KERNEL_ACCOUNT); in alloc_msg() 80 if (seg == NULL) in alloc_msg() 82 *pseg = seg; in alloc_msg() 83 seg->next = NULL; in alloc_msg() 84 pseg = &seg->next; in alloc_msg() 98 struct msg_msgseg *seg; in load_msg() local 110 for (seg = msg->next; seg != NULL; seg = seg->next) { in load_msg() 114 if (copy_from_user(seg + 1, src, alen)) in load_msg() 164 struct msg_msgseg *seg; in store_msg() local [all …]
|
/linux-6.12.1/tools/testing/selftests/sgx/ |
D | load.c | 106 static bool encl_ioc_add_pages(struct encl *encl, struct encl_segment *seg) in encl_ioc_add_pages() argument 113 secinfo.flags = seg->flags; in encl_ioc_add_pages() 115 ioc.src = (uint64_t)seg->src; in encl_ioc_add_pages() 116 ioc.offset = seg->offset; in encl_ioc_add_pages() 117 ioc.length = seg->size; in encl_ioc_add_pages() 119 if (seg->measure) in encl_ioc_add_pages() 180 struct encl_segment *seg; in encl_load() local 253 seg = &encl->segment_tbl[j]; in encl_load() 273 seg->prot = PROT_READ | PROT_WRITE; in encl_load() 274 seg->flags = SGX_PAGE_TYPE_TCS << 8; in encl_load() [all …]
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
D | mlx5hws_buddy.c | 87 unsigned int seg, order_iter, m; in hws_buddy_find_free_seg() local 95 seg = find_first_bit(buddy->bitmap[order_iter], m); in hws_buddy_find_free_seg() 97 if (WARN(seg >= m, in hws_buddy_find_free_seg() 108 *segment = seg; in hws_buddy_find_free_seg() 115 u32 seg, order_iter, err; in mlx5hws_buddy_alloc_mem() local 117 err = hws_buddy_find_free_seg(buddy, order, &seg, &order_iter); in mlx5hws_buddy_alloc_mem() 121 bitmap_clear(buddy->bitmap[order_iter], seg, 1); in mlx5hws_buddy_alloc_mem() 126 seg <<= 1; in mlx5hws_buddy_alloc_mem() 127 bitmap_set(buddy->bitmap[order_iter], seg ^ 1, 1); in mlx5hws_buddy_alloc_mem() 131 seg <<= order; in mlx5hws_buddy_alloc_mem() [all …]
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | dr_buddy.c | 77 unsigned int seg, order_iter, m; in dr_buddy_find_free_seg() local 85 seg = find_first_bit(buddy->bitmap[order_iter], m); in dr_buddy_find_free_seg() 87 if (WARN(seg >= m, in dr_buddy_find_free_seg() 98 *segment = seg; in dr_buddy_find_free_seg() 123 unsigned int seg, order_iter; in mlx5dr_buddy_alloc_mem() local 126 err = dr_buddy_find_free_seg(buddy, order, &seg, &order_iter); in mlx5dr_buddy_alloc_mem() 130 bitmap_clear(buddy->bitmap[order_iter], seg, 1); in mlx5dr_buddy_alloc_mem() 139 seg <<= 1; in mlx5dr_buddy_alloc_mem() 140 bitmap_set(buddy->bitmap[order_iter], seg ^ 1, 1); in mlx5dr_buddy_alloc_mem() 144 seg <<= order; in mlx5dr_buddy_alloc_mem() [all …]
|
D | dr_icm_pool.c | 13 unsigned int seg; member 84 return (u64)offset * chunk->seg; in mlx5dr_icm_pool_get_chunk_mr_addr() 96 return (u64)chunk->buddy_mem->icm_mr->icm_start_addr + size * chunk->seg; in mlx5dr_icm_pool_get_chunk_icm_addr() 325 unsigned int seg) in dr_icm_chunk_init() argument 329 chunk->seg = seg; in dr_icm_chunk_init() 334 offset = mlx5dr_icm_pool_dm_type_to_entry_size(pool->icm_type) * seg; in dr_icm_chunk_init() 355 hot_chunk->seg, ilog2(num_entries)); in dr_icm_pool_clear_hot_chunks_arr() 389 unsigned int *seg) in dr_icm_handle_buddies_get_mem() argument 399 chunk_size, seg); in dr_icm_handle_buddies_get_mem() 440 unsigned int seg; in mlx5dr_icm_alloc_chunk() local [all …]
|
/linux-6.12.1/net/ipv4/ |
D | udp_offload.c | 192 static void __udpv4_gso_segment_csum(struct sk_buff *seg, in __udpv4_gso_segment_csum() argument 202 uh = udp_hdr(seg); in __udpv4_gso_segment_csum() 203 iph = ip_hdr(seg); in __udpv4_gso_segment_csum() 206 inet_proto_csum_replace4(&uh->check, seg, *oldip, *newip, in __udpv4_gso_segment_csum() 208 inet_proto_csum_replace2(&uh->check, seg, *oldport, *newport, in __udpv4_gso_segment_csum() 221 struct sk_buff *seg; in __udpv4_gso_segment_list_csum() local 225 seg = segs; in __udpv4_gso_segment_list_csum() 226 uh = udp_hdr(seg); in __udpv4_gso_segment_list_csum() 227 iph = ip_hdr(seg); in __udpv4_gso_segment_list_csum() 229 if ((udp_hdr(seg)->dest == udp_hdr(seg->next)->dest) && in __udpv4_gso_segment_list_csum() [all …]
|
D | tcp_offload.c | 31 static void __tcpv4_gso_segment_csum(struct sk_buff *seg, in __tcpv4_gso_segment_csum() argument 41 th = tcp_hdr(seg); in __tcpv4_gso_segment_csum() 42 iph = ip_hdr(seg); in __tcpv4_gso_segment_csum() 44 inet_proto_csum_replace4(&th->check, seg, *oldip, newip, true); in __tcpv4_gso_segment_csum() 45 inet_proto_csum_replace2(&th->check, seg, *oldport, newport, false); in __tcpv4_gso_segment_csum() 56 struct sk_buff *seg; in __tcpv4_gso_segment_list_csum() local 60 seg = segs; in __tcpv4_gso_segment_list_csum() 61 th = tcp_hdr(seg); in __tcpv4_gso_segment_list_csum() 62 iph = ip_hdr(seg); in __tcpv4_gso_segment_list_csum() 63 th2 = tcp_hdr(seg->next); in __tcpv4_gso_segment_list_csum() [all …]
|
/linux-6.12.1/drivers/firmware/google/ |
D | memconsole-coreboot.c | 45 struct seg { /* describes ring buffer segments in logical order */ in memconsole_coreboot_read() struct 48 } seg[2] = { {0}, {0} }; in memconsole_coreboot_read() local 55 seg[0] = (struct seg){.phys = cursor, .len = size - cursor}; in memconsole_coreboot_read() 56 seg[1] = (struct seg){.phys = 0, .len = cursor}; in memconsole_coreboot_read() 58 seg[0] = (struct seg){.phys = 0, .len = min(cursor, size)}; in memconsole_coreboot_read() 61 for (i = 0; i < ARRAY_SIZE(seg) && count > done; i++) { in memconsole_coreboot_read() 63 cbmem_console->body + seg[i].phys, seg[i].len); in memconsole_coreboot_read() 64 pos -= seg[i].len; in memconsole_coreboot_read()
|
/linux-6.12.1/net/ipv6/ |
D | tcpv6_offload.c | 96 static void __tcpv6_gso_segment_csum(struct sk_buff *seg, in __tcpv6_gso_segment_csum() argument 104 th = tcp_hdr(seg); in __tcpv6_gso_segment_csum() 105 inet_proto_csum_replace2(&th->check, seg, *oldport, newport, false); in __tcpv6_gso_segment_csum() 113 struct sk_buff *seg; in __tcpv6_gso_segment_list_csum() local 117 seg = segs; in __tcpv6_gso_segment_list_csum() 118 th = tcp_hdr(seg); in __tcpv6_gso_segment_list_csum() 119 iph = ipv6_hdr(seg); in __tcpv6_gso_segment_list_csum() 120 th2 = tcp_hdr(seg->next); in __tcpv6_gso_segment_list_csum() 121 iph2 = ipv6_hdr(seg->next); in __tcpv6_gso_segment_list_csum() 128 while ((seg = seg->next)) { in __tcpv6_gso_segment_list_csum() [all …]
|
/linux-6.12.1/arch/m68k/sun3/ |
D | mmu_emu.c | 129 unsigned long seg, num; in mmu_emu_init() local 152 for(seg = bootmem_end; seg < 0x0f800000; seg += SUN3_PMEG_SIZE) { in mmu_emu_init() 153 i = sun3_get_segmap(seg); in mmu_emu_init() 158 print_pte_vaddr (seg); in mmu_emu_init() 160 sun3_put_segmap(seg, SUN3_INVALID_PMEG); in mmu_emu_init() 165 for (num=0, seg=0x0F800000; seg<0x10000000; seg+=16*PAGE_SIZE) { in mmu_emu_init() 166 if (sun3_get_segmap (seg) != SUN3_INVALID_PMEG) { in mmu_emu_init() 170 print_pte_vaddr (seg + (i*PAGE_SIZE)); in mmu_emu_init() 177 m68k_vmalloc_end = seg; in mmu_emu_init() 182 pmeg_alloc[sun3_get_segmap(seg)] = 2; in mmu_emu_init() [all …]
|
/linux-6.12.1/drivers/net/ethernet/intel/ice/ |
D | ice_ethtool_fdir.c | 627 * @seg: protocol header description pointer 632 ice_fdir_set_hw_fltr_rule(struct ice_pf *pf, struct ice_flow_seg_info *seg, in ice_fdir_set_hw_fltr_rule() argument 666 if (!memcmp(old_seg, seg, sizeof(*seg))) in ice_fdir_set_hw_fltr_rule() 691 err = ice_flow_add_prof(hw, ICE_BLK_FD, ICE_FLOW_RX, seg, in ice_fdir_set_hw_fltr_rule() 697 seg, &entry1_h); in ice_fdir_set_hw_fltr_rule() 702 seg, &entry2_h); in ice_fdir_set_hw_fltr_rule() 706 hw_prof->fdir_seg[tun] = seg; in ice_fdir_set_hw_fltr_rule() 726 ICE_FLOW_PRIO_NORMAL, seg, in ice_fdir_set_hw_fltr_rule() 782 * @seg: flow segment for programming 791 ice_set_init_fdir_seg(struct ice_flow_seg_info *seg, in ice_set_init_fdir_seg() argument [all …]
|
D | ice_ddp.c | 60 const struct ice_generic_seg_hdr *seg; in ice_verify_pkg() local 63 if (len < off + sizeof(*seg)) in ice_verify_pkg() 66 seg = (void *)pkg + off; in ice_verify_pkg() 69 if (len < off + le32_to_cpu(seg->seg_size)) in ice_verify_pkg() 90 hw->seg = NULL; in ice_free_seg() 381 if (!hw->seg) in ice_get_prof_index_max() 384 ice_seg = hw->seg; in ice_get_prof_index_max() 785 ice_seg = hw->seg; in ice_get_sw_fv_bitmap() 830 if (!lkups->n_val_words || !hw->seg) in ice_get_sw_fv_list() 833 ice_seg = hw->seg; in ice_get_sw_fv_list() [all …]
|
/linux-6.12.1/drivers/usb/cdns3/ |
D | cdnsp-mem.c | 34 struct cdnsp_segment *seg; in cdnsp_segment_alloc() local 38 seg = kzalloc(sizeof(*seg), flags); in cdnsp_segment_alloc() 39 if (!seg) in cdnsp_segment_alloc() 42 seg->trbs = dma_pool_zalloc(pdev->segment_pool, flags, &dma); in cdnsp_segment_alloc() 43 if (!seg->trbs) { in cdnsp_segment_alloc() 44 kfree(seg); in cdnsp_segment_alloc() 49 seg->bounce_buf = kzalloc(max_packet, flags | GFP_DMA); in cdnsp_segment_alloc() 50 if (!seg->bounce_buf) in cdnsp_segment_alloc() 57 seg->trbs[i].link.control |= cpu_to_le32(TRB_CYCLE); in cdnsp_segment_alloc() 59 seg->dma = dma; in cdnsp_segment_alloc() [all …]
|
/linux-6.12.1/drivers/acpi/ |
D | pci_mcfg.c | 46 #define AL_ECAM(table_id, rev, seg, ops) \ argument 47 { "AMAZON", table_id, rev, seg, MCFG_BUS_ANY, ops } 58 #define QCOM_ECAM32(seg) \ argument 59 { "QCOM ", "QDF2432 ", 1, seg, MCFG_BUS_ANY, &pci_32b_ops } 70 #define HISI_QUAD_DOM(table_id, seg, ops) \ argument 71 { "HISI ", table_id, 0, (seg) + 0, MCFG_BUS_ANY, ops }, \ 72 { "HISI ", table_id, 0, (seg) + 1, MCFG_BUS_ANY, ops }, \ 73 { "HISI ", table_id, 0, (seg) + 2, MCFG_BUS_ANY, ops }, \ 74 { "HISI ", table_id, 0, (seg) + 3, MCFG_BUS_ANY, ops } 100 #define THUNDER_ECAM_QUIRK(rev, seg) \ argument [all …]
|
/linux-6.12.1/drivers/media/test-drivers/visl/ |
D | visl-trace-vp9.h | 25 "seg.feature_data {%s}\n" 26 "seg.feature_enabled %s\n" 27 "seg.tree_probs %s\n" 28 "seg.pred_probs %s\n" 29 "seg.flags %s\n" 65 __entry->f.seg.feature_data, 66 sizeof(__entry->f.seg.feature_data), 68 __print_array(__entry->f.seg.feature_enabled, 69 ARRAY_SIZE(__entry->f.seg.feature_enabled), 70 sizeof(__entry->f.seg.feature_enabled[0])), [all …]
|
/linux-6.12.1/arch/m68k/include/asm/ |
D | tlbflush.h | 142 unsigned char seg; in flush_tlb_mm() local 149 seg = sun3_get_segmap(i); in flush_tlb_mm() 150 if (seg == SUN3_INVALID_PMEG) in flush_tlb_mm() 154 pmeg_alloc[seg] = 0; in flush_tlb_mm() 155 pmeg_ctx[seg] = 0; in flush_tlb_mm() 156 pmeg_vaddr[seg] = 0; in flush_tlb_mm() 189 unsigned char seg, oldctx; in flush_tlb_range() local 198 if((seg = sun3_get_segmap(start)) == SUN3_INVALID_PMEG) in flush_tlb_range() 200 if(pmeg_ctx[seg] == mm->context) { in flush_tlb_range() 201 pmeg_alloc[seg] = 0; in flush_tlb_range() [all …]
|
/linux-6.12.1/drivers/net/ethernet/marvell/octeontx2/nic/ |
D | otx2_txrx.c | 84 struct sk_buff *skb, int seg, int *len) in otx2_dma_map_skb_frag() argument 91 if (!seg) { in otx2_dma_map_skb_frag() 96 frag = &skb_shinfo(skb)->frags[seg - 1]; in otx2_dma_map_skb_frag() 106 int seg; in otx2_dma_unmap_skb_frags() local 108 for (seg = 0; seg < sg->num_segs; seg++) { in otx2_dma_unmap_skb_frags() 109 otx2_dma_unmap_page(pfvf, sg->dma_addr[seg], in otx2_dma_unmap_skb_frags() 110 sg->size[seg], DMA_TO_DEVICE); in otx2_dma_unmap_skb_frags() 260 int seg; in otx2_free_rcv_seg() local 267 for (seg = 0; seg < sg->segs; seg++, seg_addr++) in otx2_free_rcv_seg() 349 int seg; in otx2_rcv_pkt_handler() local [all …]
|
/linux-6.12.1/fs/ |
D | binfmt_elf_fdpic.c | 740 struct elf_fdpic_loadseg *seg; in elf_fdpic_map_file() local 783 seg = loadmap->segs; in elf_fdpic_map_file() 784 for (loop = loadmap->nsegs; loop > 0; loop--, seg++) { in elf_fdpic_map_file() 785 if (params->hdr.e_entry >= seg->p_vaddr && in elf_fdpic_map_file() 786 params->hdr.e_entry < seg->p_vaddr + seg->p_memsz) { in elf_fdpic_map_file() 788 (params->hdr.e_entry - seg->p_vaddr) + in elf_fdpic_map_file() 789 seg->addr; in elf_fdpic_map_file() 808 seg = loadmap->segs; in elf_fdpic_map_file() 809 for (loop = loadmap->nsegs; loop > 0; loop--, seg++) { in elf_fdpic_map_file() 810 if (phdr->p_vaddr >= seg->p_vaddr && in elf_fdpic_map_file() [all …]
|
/linux-6.12.1/fs/nfsd/ |
D | flexfilelayout.c | 26 struct nfsd4_layout_seg *seg = &args->lg_seg; in nfsd4_ff_proc_layoutget() local 53 if (seg->iomode == IOMODE_READ) { in nfsd4_ff_proc_layoutget() 68 seg->offset = 0; in nfsd4_ff_proc_layoutget() 69 seg->length = NFS4_MAX_UINT64; in nfsd4_ff_proc_layoutget() 71 dprintk("GET: 0x%llx:0x%llx %d\n", seg->offset, seg->length, in nfsd4_ff_proc_layoutget() 72 seg->iomode); in nfsd4_ff_proc_layoutget() 76 seg->length = 0; in nfsd4_ff_proc_layoutget()
|
D | blocklayout.c | 24 struct nfsd4_layout_seg *seg = &args->lg_seg; in nfsd4_block_proc_layoutget() local 32 if (seg->offset & (block_size - 1)) { in nfsd4_block_proc_layoutget() 47 error = sb->s_export_op->map_blocks(inode, seg->offset, seg->length, in nfsd4_block_proc_layoutget() 48 &iomap, seg->iomode != IOMODE_READ, in nfsd4_block_proc_layoutget() 63 if (seg->iomode == IOMODE_READ) in nfsd4_block_proc_layoutget() 70 if (seg->iomode & IOMODE_RW) { in nfsd4_block_proc_layoutget() 85 if (seg->iomode == IOMODE_READ) { in nfsd4_block_proc_layoutget() 102 seg->offset = iomap.offset; in nfsd4_block_proc_layoutget() 103 seg->length = iomap.length; in nfsd4_block_proc_layoutget() 109 seg->length = 0; in nfsd4_block_proc_layoutget() [all …]
|
/linux-6.12.1/drivers/peci/ |
D | internal.h | 47 struct peci_request *peci_xfer_ep_pci_cfg_local_readb(struct peci_device *device, u8 seg, 49 struct peci_request *peci_xfer_ep_pci_cfg_local_readw(struct peci_device *device, u8 seg, 51 struct peci_request *peci_xfer_ep_pci_cfg_local_readl(struct peci_device *device, u8 seg, 54 struct peci_request *peci_xfer_ep_pci_cfg_readb(struct peci_device *device, u8 seg, 56 struct peci_request *peci_xfer_ep_pci_cfg_readw(struct peci_device *device, u8 seg, 58 struct peci_request *peci_xfer_ep_pci_cfg_readl(struct peci_device *device, u8 seg, 61 struct peci_request *peci_xfer_ep_mmio32_readl(struct peci_device *device, u8 bar, u8 seg, 64 struct peci_request *peci_xfer_ep_mmio64_readl(struct peci_device *device, u8 bar, u8 seg,
|
/linux-6.12.1/drivers/infiniband/hw/mthca/ |
D | mthca_mr.c | 88 u32 seg; in mthca_buddy_alloc() local 95 seg = find_first_bit(buddy->bits[o], m); in mthca_buddy_alloc() 96 if (seg < m) in mthca_buddy_alloc() 104 __clear_bit(seg, buddy->bits[o]); in mthca_buddy_alloc() 109 seg <<= 1; in mthca_buddy_alloc() 110 __set_bit(seg ^ 1, buddy->bits[o]); in mthca_buddy_alloc() 116 seg <<= order; in mthca_buddy_alloc() 118 return seg; in mthca_buddy_alloc() 121 static void mthca_buddy_free(struct mthca_buddy *buddy, u32 seg, int order) in mthca_buddy_free() argument 123 seg >>= order; in mthca_buddy_free() [all …]
|