Home
last modified time | relevance | path

Searched refs:cached (Results 1 – 25 of 225) sorted by relevance

123456789

/linux-6.12.1/tools/lib/api/fs/
Dcgroup.c17 static struct cgroupfs_cache_entry *cached; variable
27 if (cached && !strcmp(cached->subsys, subsys)) { in cgroupfs_find_mountpoint()
28 if (strlen(cached->mountpoint) < maxlen) { in cgroupfs_find_mountpoint()
29 strcpy(buf, cached->mountpoint); in cgroupfs_find_mountpoint()
94 if (!cached) in cgroupfs_find_mountpoint()
95 cached = calloc(1, sizeof(*cached)); in cgroupfs_find_mountpoint()
97 if (cached) { in cgroupfs_find_mountpoint()
98 strncpy(cached->subsys, subsys, sizeof(cached->subsys) - 1); in cgroupfs_find_mountpoint()
99 strcpy(cached->mountpoint, mountpoint); in cgroupfs_find_mountpoint()
/linux-6.12.1/fs/btrfs/
Dextent-io-tree.h141 struct extent_state **cached);
143 struct extent_state **cached);
146 struct extent_state **cached) in lock_extent() argument
148 return __lock_extent(tree, start, end, EXTENT_LOCKED, cached); in lock_extent()
152 u64 end, struct extent_state **cached) in try_lock_extent() argument
154 return __try_lock_extent(tree, start, end, EXTENT_LOCKED, cached); in try_lock_extent()
172 u32 bits, struct extent_state **cached,
177 struct extent_state **cached) in clear_extent_bit() argument
179 return __clear_extent_bit(tree, start, end, bits, cached, NULL); in clear_extent_bit()
183 struct extent_state **cached) in unlock_extent() argument
[all …]
Dextent-io-tree.c628 struct extent_state *cached; in __clear_extent_bit() local
664 cached = *cached_state; in __clear_extent_bit()
671 if (cached && extent_state_in_tree(cached) && in __clear_extent_bit()
672 cached->start <= start && cached->end > start) { in __clear_extent_bit()
674 refcount_dec(&cached->refs); in __clear_extent_bit()
675 state = cached; in __clear_extent_bit()
679 free_extent_state(cached); in __clear_extent_bit()
1645 struct extent_state *cached; in count_range_bits() local
1659 cached = *cached_state; in count_range_bits()
1661 if (!extent_state_in_tree(cached)) in count_range_bits()
[all …]
Dfiemap.c72 bool cached; member
117 if (!cache->cached) in emit_fiemap_extent()
252 cache->cached = false; in emit_fiemap_extent()
266 cache->cached = false; in emit_fiemap_extent()
270 cache->cached = true; in emit_fiemap_extent()
295 if (!cache->cached) in emit_last_fiemap_cache()
300 cache->cached = false; in emit_last_fiemap_cache()
822 if (cache.cached && cache.offset + cache.len >= last_extent_end) { in extent_fiemap()
Dblock-group.h165 int cached; member
375 return cache->cached == BTRFS_CACHE_FINISHED || in btrfs_block_group_done()
376 cache->cached == BTRFS_CACHE_ERROR; in btrfs_block_group_done()
/linux-6.12.1/fs/lockd/
Dmon.c322 struct nsm_handle *cached, *new = NULL; in nsm_get_handle() local
338 cached = nsm_lookup_hostname(&ln->nsm_handles, in nsm_get_handle()
341 cached = nsm_lookup_addr(&ln->nsm_handles, sap); in nsm_get_handle()
343 if (cached != NULL) { in nsm_get_handle()
344 refcount_inc(&cached->sm_count); in nsm_get_handle()
348 "cnt %d\n", cached->sm_name, in nsm_get_handle()
349 cached->sm_addrbuf, in nsm_get_handle()
350 refcount_read(&cached->sm_count)); in nsm_get_handle()
351 return cached; in nsm_get_handle()
382 struct nsm_handle *cached; in nsm_reboot_lookup() local
[all …]
/linux-6.12.1/lib/zstd/compress/
Dhist.c91 { U32 cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() local
93 U32 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp()
98 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp()
103 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp()
108 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp()
/linux-6.12.1/drivers/mtd/maps/
Dpxa2xx-flash.c26 unsigned long start = (unsigned long)map->cached + from; in pxa2xx_map_inval_cache()
70 info->map.cached = ioremap_cache(info->map.phys, info->map.size); in pxa2xx_flash_probe()
71 if (!info->map.cached) in pxa2xx_flash_probe()
87 if (info->map.cached) in pxa2xx_flash_probe()
88 iounmap(info->map.cached); in pxa2xx_flash_probe()
109 if (info->map.cached) in pxa2xx_flash_remove()
110 iounmap(info->map.cached); in pxa2xx_flash_remove()
/linux-6.12.1/tools/perf/util/
Dsmt.c9 static bool cached; in smt_on() local
13 if (cached) in smt_on()
21 cached = true; in smt_on()
/linux-6.12.1/tools/perf/arch/x86/util/
Dtopdown.c14 static bool cached; in topdown_sys_has_perf_metrics() local
17 if (cached) in topdown_sys_has_perf_metrics()
30 cached = true; in topdown_sys_has_perf_metrics()
Dtsc.c61 static bool cached; in arch_get_tsc_freq() local
65 if (cached) in arch_get_tsc_freq()
68 cached = true; in arch_get_tsc_freq()
/linux-6.12.1/fs/proc/
Dmeminfo.c38 long cached; in meminfo_proc_show() local
48 cached = global_node_page_state(NR_FILE_PAGES) - in meminfo_proc_show()
50 if (cached < 0) in meminfo_proc_show()
51 cached = 0; in meminfo_proc_show()
64 show_val_kb(m, "Cached: ", cached); in meminfo_proc_show()
/linux-6.12.1/drivers/gpu/drm/msm/dsi/phy/
Ddsi_phy_10nm.c474 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_10nm_pll_save_state() local
478 cached->pll_out_div = readl(pll_10nm->phy->pll_base + in dsi_10nm_pll_save_state()
480 cached->pll_out_div &= 0x3; in dsi_10nm_pll_save_state()
483 cached->bit_clk_div = cmn_clk_cfg0 & 0xf; in dsi_10nm_pll_save_state()
484 cached->pix_clk_div = (cmn_clk_cfg0 & 0xf0) >> 4; in dsi_10nm_pll_save_state()
487 cached->pll_mux = cmn_clk_cfg1 & 0x3; in dsi_10nm_pll_save_state()
490 pll_10nm->phy->id, cached->pll_out_div, cached->bit_clk_div, in dsi_10nm_pll_save_state()
491 cached->pix_clk_div, cached->pll_mux); in dsi_10nm_pll_save_state()
497 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_10nm_pll_restore_state() local
504 val |= cached->pll_out_div; in dsi_10nm_pll_restore_state()
[all …]
Ddsi_phy_7nm.c544 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_save_state() local
548 cached->pll_out_div = readl(pll_7nm->phy->pll_base + in dsi_7nm_pll_save_state()
550 cached->pll_out_div &= 0x3; in dsi_7nm_pll_save_state()
553 cached->bit_clk_div = cmn_clk_cfg0 & 0xf; in dsi_7nm_pll_save_state()
554 cached->pix_clk_div = (cmn_clk_cfg0 & 0xf0) >> 4; in dsi_7nm_pll_save_state()
557 cached->pll_mux = cmn_clk_cfg1 & 0x3; in dsi_7nm_pll_save_state()
560 pll_7nm->phy->id, cached->pll_out_div, cached->bit_clk_div, in dsi_7nm_pll_save_state()
561 cached->pix_clk_div, cached->pll_mux); in dsi_7nm_pll_save_state()
567 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_restore_state() local
574 val |= cached->pll_out_div; in dsi_7nm_pll_restore_state()
[all …]
/linux-6.12.1/drivers/mtd/devices/
Dphram.c37 bool cached; member
88 if (phram->cached) in phram_map()
104 if (phram->cached) { in phram_unmap()
128 bool cached = np ? !of_property_read_bool(np, "no-map") : false; in register_device() local
136 new->cached = cached; in register_device()
/linux-6.12.1/tools/testing/selftests/drivers/net/lib/py/
Denv.py215 cached = self._required_cmd.get(comm, {})
216 if cached.get(key) is None:
217 cached[key] = cmd("command -v -- " + comm, fail=False,
219 self._required_cmd[comm] = cached
220 return cached[key]
/linux-6.12.1/Documentation/filesystems/
Dfuse-io.rst10 - cached
21 In cached mode reads may be satisfied from the page cache, and data may be
25 The cached mode has two sub modes controlling how writes are handled. The
31 WRITE requests, as well as updating any cached pages (and caching previously
/linux-6.12.1/Documentation/admin-guide/device-mapper/
Dwritecache.rst6 doesn't cache reads because reads are supposed to be cached in page cache
17 2. the underlying device that will be cached
58 new writes (however, writes to already cached blocks are
60 writes) and it will gradually writeback any cached
62 process with "dmsetup status". When the number of cached
/linux-6.12.1/arch/arm/mach-omap2/
Dsram.c184 int cached = 1; in omap2_map_sram() local
194 cached = 0; in omap2_map_sram()
201 omap_sram_base = __arm_ioremap_exec(omap_sram_start, omap_sram_size, cached); in omap2_map_sram()
/linux-6.12.1/Documentation/ABI/testing/
Dsysfs-class-iommu-intel-iommu15 The cached hardware capability register value
23 The cached hardware extended capability register
/linux-6.12.1/fs/bcachefs/
Dextents.c135 if (!ret && !p.ptr.cached) in bch2_bkey_pick_read_device()
140 if (p.ptr.cached && (!ca || dev_ptr_stale_rcu(ca, &p.ptr))) in bch2_bkey_pick_read_device()
614 ret += !p.ptr.cached && !crc_is_compressed(p.crc); in bch2_bkey_nr_ptrs_fully_allocated()
628 if (!p.ptr.cached && crc_is_compressed(p.crc)) in bch2_bkey_sectors_compressed()
654 if (p.ptr.cached) in bch2_bkey_replicas()
669 if (p->ptr.cached) in __extent_ptr_durability()
884 (!ptr->cached || in bch2_bkey_has_target()
1018 ptr->cached = true; in bch2_extent_ptr_set_cached()
1041 ptr->cached && in bch2_extent_normalize()
1068 if (ptr->cached) { in bch2_extent_normalize_by_opts()
[all …]
Dalloc_foreground.h184 bool cached) in bch2_alloc_sectors_append_ptrs_inlined() argument
197 ptr.cached = cached || in bch2_alloc_sectors_append_ptrs_inlined()
/linux-6.12.1/drivers/gpu/drm/ttm/
Dttm_agp_backend.c55 int ret, cached = ttm->caching == ttm_cached; in ttm_agp_bind() local
77 mem->type = (cached) ? AGP_USER_CACHED_MEMORY : AGP_USER_MEMORY; in ttm_agp_bind()
/linux-6.12.1/drivers/block/
Dps3vram.c428 unsigned int cached, count; in ps3vram_read() local
449 cached = CACHE_OFFSET + entry * priv->cache.page_size + offset; in ps3vram_read()
453 (unsigned int)from, cached, offset, avail, count); in ps3vram_read()
457 memcpy(buf, priv->xdr_buf + cached, avail); in ps3vram_read()
472 unsigned int cached, count; in ps3vram_write() local
490 cached = CACHE_OFFSET + entry * priv->cache.page_size + offset; in ps3vram_write()
494 cached, offset, avail, count); in ps3vram_write()
498 memcpy(priv->xdr_buf + cached, buf, avail); in ps3vram_write()
/linux-6.12.1/arch/s390/appldata/
Dappldata_mem.c54 u64 cached; /* size of (used) cache, w/o buffers */ member
102 mem_data->cached = P2K(global_node_page_state(NR_FILE_PAGES) in appldata_get_mem_data()

123456789