Home
last modified time | relevance | path

Searched refs:page_cnt (Results 1 – 25 of 32) sorted by relevance

12

/linux-6.12.1/drivers/net/ethernet/qlogic/qed/
Dqed_chain.c12 u32 page_cnt) in qed_chain_init() argument
34 chain->page_cnt = page_cnt; in qed_chain_init()
35 chain->capacity = chain->usable_per_page * page_cnt; in qed_chain_init()
36 chain->size = chain->elem_per_page * page_cnt; in qed_chain_init()
80 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_next_ptr()
114 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_pbl()
160 u32 page_cnt) in qed_chain_alloc_sanity_check() argument
165 chain_size *= page_cnt; in qed_chain_alloc_sanity_check()
206 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_alloc_next_ptr()
255 u32 page_cnt, i; in qed_chain_alloc_pbl() local
[all …]
Dqed_sp_commands.c311 u8 page_cnt, i; in qed_sp_pf_start() local
372 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_eq->chain); in qed_sp_pf_start()
373 p_ramrod->event_ring_num_pages = page_cnt; in qed_sp_pf_start()
378 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_consq->chain); in qed_sp_pf_start()
379 p_ramrod->consolid_q_num_pages = page_cnt; in qed_sp_pf_start()
/linux-6.12.1/kernel/bpf/
Darena.c418 static long arena_alloc_pages(struct bpf_arena *arena, long uaddr, long page_cnt, int node_id) in arena_alloc_pages() argument
428 if (page_cnt > page_cnt_max) in arena_alloc_pages()
435 if (pgoff > page_cnt_max - page_cnt) in arena_alloc_pages()
441 pages = kvcalloc(page_cnt, sizeof(struct page *), GFP_KERNEL); in arena_alloc_pages()
448 ret = mtree_insert_range(&arena->mt, pgoff, pgoff + page_cnt - 1, in arena_alloc_pages()
452 page_cnt, 0, page_cnt_max - 1, GFP_KERNEL); in arena_alloc_pages()
457 node_id, page_cnt, pages); in arena_alloc_pages()
470 kern_vm_start + uaddr32 + page_cnt * PAGE_SIZE, pages); in arena_alloc_pages()
472 for (i = 0; i < page_cnt; i++) in arena_alloc_pages()
490 static void zap_pages(struct bpf_arena *arena, long uaddr, long page_cnt) in zap_pages() argument
[all …]
/linux-6.12.1/include/linux/qed/
Dqed_chain.h106 u32 page_cnt; member
262 return chain->page_cnt; in qed_chain_get_page_cnt()
303 if (++(*(u16 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page()
307 if (++(*(u32 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page()
498 u32 reset_val = p_chain->page_cnt - 1; in qed_chain_reset()
556 last_page_idx = p_chain->page_cnt - 1; in qed_chain_get_last_elem()
580 u32 cur_prod, page_mask, page_cnt, page_diff; in qed_chain_set_prod() local
599 page_cnt = qed_chain_get_page_cnt(p_chain); in qed_chain_set_prod()
603 page_diff + page_cnt) % page_cnt; in qed_chain_set_prod()
607 page_diff + page_cnt) % page_cnt; in qed_chain_set_prod()
[all …]
/linux-6.12.1/tools/testing/selftests/bpf/
Dbpf_arena_common.h47 void __arena* bpf_arena_alloc_pages(void *map, void __arena *addr, __u32 page_cnt,
49 void bpf_arena_free_pages(void *map, void __arena *ptr, __u32 page_cnt) __ksym __weak;
63 static inline void __arena* bpf_arena_alloc_pages(void *map, void *addr, __u32 page_cnt, in bpf_arena_alloc_pages() argument
68 static inline void bpf_arena_free_pages(void *map, void __arena *ptr, __u32 page_cnt) in bpf_arena_free_pages() argument
/linux-6.12.1/arch/mips/cavium-octeon/executive/
Dcvmx-helper-util.c133 union cvmx_ipd_portx_bp_page_cnt page_cnt; in cvmx_helper_setup_red() local
141 page_cnt.u64 = 0; in cvmx_helper_setup_red()
142 page_cnt.s.bp_enb = 0; in cvmx_helper_setup_red()
143 page_cnt.s.page_cnt = 100; in cvmx_helper_setup_red()
148 page_cnt.u64); in cvmx_helper_setup_red()
/linux-6.12.1/drivers/base/firmware_loader/
Dsysfs.c246 int page_cnt = min_t(size_t, PAGE_SIZE - page_ofs, count); in firmware_rw() local
250 page_ofs, page_cnt); in firmware_rw()
253 buffer, page_cnt); in firmware_rw()
255 buffer += page_cnt; in firmware_rw()
256 offset += page_cnt; in firmware_rw()
257 count -= page_cnt; in firmware_rw()
/linux-6.12.1/drivers/hv/
Dring_buffer.c186 struct page *pages, u32 page_cnt, u32 max_pkt_size) in hv_ringbuffer_init() argument
197 pages_wraparound = kcalloc(page_cnt * 2 - 1, in hv_ringbuffer_init()
204 for (i = 0; i < 2 * (page_cnt - 1); i++) in hv_ringbuffer_init()
206 &pages[i % (page_cnt - 1) + 1]; in hv_ringbuffer_init()
209 vmap(pages_wraparound, page_cnt * 2 - 1, VM_MAP, in hv_ringbuffer_init()
228 ring_info->ring_size = page_cnt << PAGE_SHIFT; in hv_ringbuffer_init()
Dhv_balloon.c141 __u64 page_cnt:24; member
981 pfn_cnt = dm->ha_wrk.ha_page_range.finfo.page_cnt; in hot_add_req()
984 rg_sz = dm->ha_wrk.ha_region_range.finfo.page_cnt; in hot_add_req()
1183 int num_pages = range_array->finfo.page_cnt; in free_balloon_pages()
1240 bl_resp->range_array[i].finfo.page_cnt = alloc_unit; in alloc_balloon_pages()
/linux-6.12.1/drivers/infiniband/hw/hns/
Dhns_roce_mr.c725 int page_cnt = 0; in cal_mtr_pg_cnt() local
730 page_cnt += region->count; in cal_mtr_pg_cnt()
733 return page_cnt; in cal_mtr_pg_cnt()
797 dma_addr_t *pages, unsigned int page_cnt) in hns_roce_mtr_map() argument
814 mapped_cnt < page_cnt; i++) { in hns_roce_mtr_map()
822 if (r->offset + r->count > page_cnt) { in hns_roce_mtr_map()
826 i, r->offset, r->count, page_cnt); in hns_roce_mtr_map()
831 page_cnt - mapped_cnt); in hns_roce_mtr_map()
842 if (mapped_cnt < page_cnt) { in hns_roce_mtr_map()
845 mapped_cnt, page_cnt); in hns_roce_mtr_map()
[all …]
/linux-6.12.1/arch/mips/include/asm/octeon/
Dcvmx-ipd-defs.h324 uint64_t page_cnt:17; member
326 uint64_t page_cnt:17;
999 uint64_t page_cnt:17; member
1001 uint64_t page_cnt:17;
1014 uint64_t page_cnt:17; member
1016 uint64_t page_cnt:17;
1029 uint64_t page_cnt:17; member
1031 uint64_t page_cnt:17;
1387 uint64_t page_cnt:25; member
1389 uint64_t page_cnt:25;
/linux-6.12.1/drivers/staging/rts5208/
Dxd.c1499 u8 reg_val, page_cnt; in xd_read_multiple_pages() local
1505 page_cnt = end_page - start_page; in xd_read_multiple_pages()
1531 rtsx_add_cmd(chip, WRITE_REG_CMD, XD_PAGE_CNT, 0xFF, page_cnt); in xd_read_multiple_pages()
1536 page_cnt * 512, DMA_512); in xd_read_multiple_pages()
1546 retval = rtsx_transfer_data_partial(chip, XD_CARD, buf, page_cnt * 512, in xd_read_multiple_pages()
1701 u8 page_cnt, reg_val; in xd_write_multiple_pages() local
1709 page_cnt = end_page - start_page; in xd_write_multiple_pages()
1731 rtsx_add_cmd(chip, WRITE_REG_CMD, XD_PAGE_CNT, 0xFF, page_cnt); in xd_write_multiple_pages()
1735 page_cnt * 512, DMA_512); in xd_write_multiple_pages()
1744 retval = rtsx_transfer_data_partial(chip, XD_CARD, buf, page_cnt * 512, in xd_write_multiple_pages()
[all …]
/linux-6.12.1/fs/ntfs3/
Dfslog.c1604 u32 page_cnt = 1, page_pos = 1; in last_log_lsn() local
1727 page_cnt = le16_to_cpu(best_page->page_count); in last_log_lsn()
1728 if (page_cnt > 1) in last_log_lsn()
1736 (page_pos != page_cnt || best_page_pos == page_pos || in last_log_lsn()
1738 (page_pos >= page_cnt || best_page_pos == page_pos)) { in last_log_lsn()
1750 page_cnt = bppc; in last_log_lsn()
1753 if (page_cnt > 1) in last_log_lsn()
1756 page_pos = page_cnt = 1; in last_log_lsn()
1786 page_cnt = page_pos = 1; in last_log_lsn()
1828 ((page_pos == page_cnt && in last_log_lsn()
[all …]
/linux-6.12.1/arch/x86/kernel/cpu/sgx/
Dencl.h55 unsigned int page_cnt; member
Dioctl.c28 if (!(encl->page_cnt % SGX_VA_SLOT_COUNT)) { in sgx_encl_grow()
40 WARN_ON_ONCE(encl->page_cnt % SGX_VA_SLOT_COUNT); in sgx_encl_grow()
42 encl->page_cnt++; in sgx_encl_grow()
48 encl->page_cnt--; in sgx_encl_shrink()
/linux-6.12.1/drivers/infiniband/hw/erdma/
Derdma_verbs.c140 mr->mem.page_cnt > ERDMA_MAX_INLINE_MTT_ENTRIES) { in regmr_cmd()
150 MTT_SIZE(mr->mem.page_cnt)); in regmr_cmd()
162 FIELD_PREP(ERDMA_CMD_REGMR_MTT_CNT_MASK, mr->mem.page_cnt); in regmr_cmd()
177 req.mtt_cnt_h = mr->mem.page_cnt >> 20; in regmr_cmd()
784 mem->page_cnt = mem->mtt_nents; in get_mtt_entries()
785 mem->mtt = erdma_create_mtt(dev, MTT_SIZE(mem->page_cnt), in get_mtt_entries()
1103 mr->mem.page_cnt = max_num_sg; in erdma_ib_alloc_mr()
1133 if (mr->mem.mtt_nents >= mr->mem.page_cnt) in erdma_set_page()
Derdma_verbs.h117 u32 page_cnt; member
/linux-6.12.1/drivers/infiniband/hw/efa/
Defa_verbs.c1257 static struct scatterlist *efa_vmalloc_buf_to_sg(u64 *buf, int page_cnt) in efa_vmalloc_buf_to_sg() argument
1263 sglist = kmalloc_array(page_cnt, sizeof(*sglist), GFP_KERNEL); in efa_vmalloc_buf_to_sg()
1266 sg_init_table(sglist, page_cnt); in efa_vmalloc_buf_to_sg()
1267 for (i = 0; i < page_cnt; i++) { in efa_vmalloc_buf_to_sg()
1288 int page_cnt = pbl->phys.indirect.pbl_buf_size_in_pages; in pbl_chunk_list_create() local
1299 chunk_list_size = DIV_ROUND_UP(page_cnt, EFA_PTRS_PER_CHUNK); in pbl_chunk_list_create()
1310 page_cnt); in pbl_chunk_list_create()
1321 ((page_cnt % EFA_PTRS_PER_CHUNK) * EFA_CHUNK_PAYLOAD_PTR_SIZE) + in pbl_chunk_list_create()
/linux-6.12.1/fs/ubifs/
Dfile.c717 int err, page_idx, page_cnt, ret = 0, n = 0; in ubifs_do_bulk_read() local
732 page_cnt = bu->blk_cnt >> UBIFS_BLOCKS_PER_PAGE_SHIFT; in ubifs_do_bulk_read()
733 if (!page_cnt) { in ubifs_do_bulk_read()
776 for (page_idx = 1; page_idx < page_cnt; page_idx++) { in ubifs_do_bulk_read()
/linux-6.12.1/drivers/target/
Dtarget_core_user.c513 int page_cnt = DIV_ROUND_UP(length, PAGE_SIZE); in tcmu_get_empty_block() local
523 for (cnt = 0; xas_next(&xas) && cnt < page_cnt;) in tcmu_get_empty_block()
527 for (i = cnt; i < page_cnt; i++) { in tcmu_get_empty_block()
551 return i == page_cnt ? dbi : -1; in tcmu_get_empty_block()
710 int page_cnt, page_inx, dpi; in tcmu_copy_data() local
729 page_cnt = DIV_ROUND_UP(data_len, PAGE_SIZE); in tcmu_copy_data()
730 if (page_cnt > udev->data_pages_per_blk) in tcmu_copy_data()
731 page_cnt = udev->data_pages_per_blk; in tcmu_copy_data()
734 for (page_inx = 0; page_inx < page_cnt && data_len; in tcmu_copy_data()
/linux-6.12.1/drivers/net/wireless/realtek/rtw88/
Dfw.c2011 u8 page_cnt, pages; in _rtw_hw_scan_update_probe_req() local
2016 page_cnt = RTW_OLD_PROBE_PG_CNT; in _rtw_hw_scan_update_probe_req()
2018 page_cnt = RTW_PROBE_PG_CNT; in _rtw_hw_scan_update_probe_req()
2020 pages = page_offset + num_probes * page_cnt; in _rtw_hw_scan_update_probe_req()
2030 if (skb->len > page_size * page_cnt) { in _rtw_hw_scan_update_probe_req()
2040 buf_offset += page_cnt * page_size; in _rtw_hw_scan_update_probe_req()
2041 page_offset += page_cnt; in _rtw_hw_scan_update_probe_req()
/linux-6.12.1/drivers/infiniband/hw/irdma/
Dverbs.h106 u32 page_cnt; member
/linux-6.12.1/drivers/infiniband/hw/qedr/
Dmain.c788 u32 page_cnt; in qedr_init_hw() local
802 page_cnt = qed_chain_get_page_cnt(&dev->cnq_array[i].pbl); in qedr_init_hw()
803 cur_pbl->num_pbl_pages = page_cnt; in qedr_init_hw()
Dverbs.c843 int chain_entries, int page_cnt, in qedr_init_cq_params() argument
854 params->pbl_num_pages = page_cnt; in qedr_init_cq_params()
925 int page_cnt; in qedr_create_cq() local
974 page_cnt = cq->q.pbl_info.num_pbes; in qedr_create_cq()
986 page_cnt = qed_chain_get_page_cnt(&cq->pbl); in qedr_create_cq()
991 qedr_init_cq_params(cq, ctx, dev, vector, chain_entries, page_cnt, in qedr_create_cq()
1553 u32 page_cnt, page_size; in qedr_create_srq() local
1589 page_cnt = srq->usrq.pbl_info.num_pbes; in qedr_create_srq()
1601 page_cnt = qed_chain_get_page_cnt(pbl); in qedr_create_srq()
1610 in_params.num_pages = page_cnt; in qedr_create_srq()
/linux-6.12.1/tools/lib/bpf/
Dlibbpf.h1554 perf_buffer__new(int map_fd, size_t page_cnt,
1588 perf_buffer__new_raw(int map_fd, size_t page_cnt, struct perf_event_attr *attr,

12