/linux-6.12.1/tools/lib/api/fs/ |
D | cgroup.c | 17 static struct cgroupfs_cache_entry *cached; variable 27 if (cached && !strcmp(cached->subsys, subsys)) { in cgroupfs_find_mountpoint() 28 if (strlen(cached->mountpoint) < maxlen) { in cgroupfs_find_mountpoint() 29 strcpy(buf, cached->mountpoint); in cgroupfs_find_mountpoint() 94 if (!cached) in cgroupfs_find_mountpoint() 95 cached = calloc(1, sizeof(*cached)); in cgroupfs_find_mountpoint() 97 if (cached) { in cgroupfs_find_mountpoint() 98 strncpy(cached->subsys, subsys, sizeof(cached->subsys) - 1); in cgroupfs_find_mountpoint() 99 strcpy(cached->mountpoint, mountpoint); in cgroupfs_find_mountpoint()
|
/linux-6.12.1/fs/btrfs/ |
D | extent-io-tree.h | 141 struct extent_state **cached); 143 struct extent_state **cached); 146 struct extent_state **cached) in lock_extent() argument 148 return __lock_extent(tree, start, end, EXTENT_LOCKED, cached); in lock_extent() 152 u64 end, struct extent_state **cached) in try_lock_extent() argument 154 return __try_lock_extent(tree, start, end, EXTENT_LOCKED, cached); in try_lock_extent() 172 u32 bits, struct extent_state **cached, 177 struct extent_state **cached) in clear_extent_bit() argument 179 return __clear_extent_bit(tree, start, end, bits, cached, NULL); in clear_extent_bit() 183 struct extent_state **cached) in unlock_extent() argument [all …]
|
D | extent-io-tree.c | 628 struct extent_state *cached; in __clear_extent_bit() local 664 cached = *cached_state; in __clear_extent_bit() 671 if (cached && extent_state_in_tree(cached) && in __clear_extent_bit() 672 cached->start <= start && cached->end > start) { in __clear_extent_bit() 674 refcount_dec(&cached->refs); in __clear_extent_bit() 675 state = cached; in __clear_extent_bit() 679 free_extent_state(cached); in __clear_extent_bit() 1645 struct extent_state *cached; in count_range_bits() local 1659 cached = *cached_state; in count_range_bits() 1661 if (!extent_state_in_tree(cached)) in count_range_bits() [all …]
|
D | fiemap.c | 72 bool cached; member 117 if (!cache->cached) in emit_fiemap_extent() 252 cache->cached = false; in emit_fiemap_extent() 266 cache->cached = false; in emit_fiemap_extent() 270 cache->cached = true; in emit_fiemap_extent() 295 if (!cache->cached) in emit_last_fiemap_cache() 300 cache->cached = false; in emit_last_fiemap_cache() 822 if (cache.cached && cache.offset + cache.len >= last_extent_end) { in extent_fiemap()
|
D | block-group.h | 165 int cached; member 375 return cache->cached == BTRFS_CACHE_FINISHED || in btrfs_block_group_done() 376 cache->cached == BTRFS_CACHE_ERROR; in btrfs_block_group_done()
|
/linux-6.12.1/fs/lockd/ |
D | mon.c | 322 struct nsm_handle *cached, *new = NULL; in nsm_get_handle() local 338 cached = nsm_lookup_hostname(&ln->nsm_handles, in nsm_get_handle() 341 cached = nsm_lookup_addr(&ln->nsm_handles, sap); in nsm_get_handle() 343 if (cached != NULL) { in nsm_get_handle() 344 refcount_inc(&cached->sm_count); in nsm_get_handle() 348 "cnt %d\n", cached->sm_name, in nsm_get_handle() 349 cached->sm_addrbuf, in nsm_get_handle() 350 refcount_read(&cached->sm_count)); in nsm_get_handle() 351 return cached; in nsm_get_handle() 382 struct nsm_handle *cached; in nsm_reboot_lookup() local [all …]
|
/linux-6.12.1/lib/zstd/compress/ |
D | hist.c | 91 { U32 cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() local 93 U32 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() 98 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() 103 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() 108 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp()
|
/linux-6.12.1/drivers/mtd/maps/ |
D | pxa2xx-flash.c | 26 unsigned long start = (unsigned long)map->cached + from; in pxa2xx_map_inval_cache() 70 info->map.cached = ioremap_cache(info->map.phys, info->map.size); in pxa2xx_flash_probe() 71 if (!info->map.cached) in pxa2xx_flash_probe() 87 if (info->map.cached) in pxa2xx_flash_probe() 88 iounmap(info->map.cached); in pxa2xx_flash_probe() 109 if (info->map.cached) in pxa2xx_flash_remove() 110 iounmap(info->map.cached); in pxa2xx_flash_remove()
|
/linux-6.12.1/tools/perf/util/ |
D | smt.c | 9 static bool cached; in smt_on() local 13 if (cached) in smt_on() 21 cached = true; in smt_on()
|
/linux-6.12.1/tools/perf/arch/x86/util/ |
D | topdown.c | 14 static bool cached; in topdown_sys_has_perf_metrics() local 17 if (cached) in topdown_sys_has_perf_metrics() 30 cached = true; in topdown_sys_has_perf_metrics()
|
D | tsc.c | 61 static bool cached; in arch_get_tsc_freq() local 65 if (cached) in arch_get_tsc_freq() 68 cached = true; in arch_get_tsc_freq()
|
/linux-6.12.1/fs/proc/ |
D | meminfo.c | 38 long cached; in meminfo_proc_show() local 48 cached = global_node_page_state(NR_FILE_PAGES) - in meminfo_proc_show() 50 if (cached < 0) in meminfo_proc_show() 51 cached = 0; in meminfo_proc_show() 64 show_val_kb(m, "Cached: ", cached); in meminfo_proc_show()
|
/linux-6.12.1/drivers/gpu/drm/msm/dsi/phy/ |
D | dsi_phy_10nm.c | 474 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_10nm_pll_save_state() local 478 cached->pll_out_div = readl(pll_10nm->phy->pll_base + in dsi_10nm_pll_save_state() 480 cached->pll_out_div &= 0x3; in dsi_10nm_pll_save_state() 483 cached->bit_clk_div = cmn_clk_cfg0 & 0xf; in dsi_10nm_pll_save_state() 484 cached->pix_clk_div = (cmn_clk_cfg0 & 0xf0) >> 4; in dsi_10nm_pll_save_state() 487 cached->pll_mux = cmn_clk_cfg1 & 0x3; in dsi_10nm_pll_save_state() 490 pll_10nm->phy->id, cached->pll_out_div, cached->bit_clk_div, in dsi_10nm_pll_save_state() 491 cached->pix_clk_div, cached->pll_mux); in dsi_10nm_pll_save_state() 497 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_10nm_pll_restore_state() local 504 val |= cached->pll_out_div; in dsi_10nm_pll_restore_state() [all …]
|
D | dsi_phy_7nm.c | 544 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_save_state() local 548 cached->pll_out_div = readl(pll_7nm->phy->pll_base + in dsi_7nm_pll_save_state() 550 cached->pll_out_div &= 0x3; in dsi_7nm_pll_save_state() 553 cached->bit_clk_div = cmn_clk_cfg0 & 0xf; in dsi_7nm_pll_save_state() 554 cached->pix_clk_div = (cmn_clk_cfg0 & 0xf0) >> 4; in dsi_7nm_pll_save_state() 557 cached->pll_mux = cmn_clk_cfg1 & 0x3; in dsi_7nm_pll_save_state() 560 pll_7nm->phy->id, cached->pll_out_div, cached->bit_clk_div, in dsi_7nm_pll_save_state() 561 cached->pix_clk_div, cached->pll_mux); in dsi_7nm_pll_save_state() 567 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_restore_state() local 574 val |= cached->pll_out_div; in dsi_7nm_pll_restore_state() [all …]
|
/linux-6.12.1/drivers/mtd/devices/ |
D | phram.c | 37 bool cached; member 88 if (phram->cached) in phram_map() 104 if (phram->cached) { in phram_unmap() 128 bool cached = np ? !of_property_read_bool(np, "no-map") : false; in register_device() local 136 new->cached = cached; in register_device()
|
/linux-6.12.1/tools/testing/selftests/drivers/net/lib/py/ |
D | env.py | 215 cached = self._required_cmd.get(comm, {}) 216 if cached.get(key) is None: 217 cached[key] = cmd("command -v -- " + comm, fail=False, 219 self._required_cmd[comm] = cached 220 return cached[key]
|
/linux-6.12.1/Documentation/filesystems/ |
D | fuse-io.rst | 10 - cached 21 In cached mode reads may be satisfied from the page cache, and data may be 25 The cached mode has two sub modes controlling how writes are handled. The 31 WRITE requests, as well as updating any cached pages (and caching previously
|
/linux-6.12.1/Documentation/admin-guide/device-mapper/ |
D | writecache.rst | 6 doesn't cache reads because reads are supposed to be cached in page cache 17 2. the underlying device that will be cached 58 new writes (however, writes to already cached blocks are 60 writes) and it will gradually writeback any cached 62 process with "dmsetup status". When the number of cached
|
/linux-6.12.1/arch/arm/mach-omap2/ |
D | sram.c | 184 int cached = 1; in omap2_map_sram() local 194 cached = 0; in omap2_map_sram() 201 omap_sram_base = __arm_ioremap_exec(omap_sram_start, omap_sram_size, cached); in omap2_map_sram()
|
/linux-6.12.1/Documentation/ABI/testing/ |
D | sysfs-class-iommu-intel-iommu | 15 The cached hardware capability register value 23 The cached hardware extended capability register
|
/linux-6.12.1/fs/bcachefs/ |
D | extents.c | 135 if (!ret && !p.ptr.cached) in bch2_bkey_pick_read_device() 140 if (p.ptr.cached && (!ca || dev_ptr_stale_rcu(ca, &p.ptr))) in bch2_bkey_pick_read_device() 614 ret += !p.ptr.cached && !crc_is_compressed(p.crc); in bch2_bkey_nr_ptrs_fully_allocated() 628 if (!p.ptr.cached && crc_is_compressed(p.crc)) in bch2_bkey_sectors_compressed() 654 if (p.ptr.cached) in bch2_bkey_replicas() 669 if (p->ptr.cached) in __extent_ptr_durability() 884 (!ptr->cached || in bch2_bkey_has_target() 1018 ptr->cached = true; in bch2_extent_ptr_set_cached() 1041 ptr->cached && in bch2_extent_normalize() 1068 if (ptr->cached) { in bch2_extent_normalize_by_opts() [all …]
|
D | alloc_foreground.h | 184 bool cached) in bch2_alloc_sectors_append_ptrs_inlined() argument 197 ptr.cached = cached || in bch2_alloc_sectors_append_ptrs_inlined()
|
/linux-6.12.1/drivers/gpu/drm/ttm/ |
D | ttm_agp_backend.c | 55 int ret, cached = ttm->caching == ttm_cached; in ttm_agp_bind() local 77 mem->type = (cached) ? AGP_USER_CACHED_MEMORY : AGP_USER_MEMORY; in ttm_agp_bind()
|
/linux-6.12.1/drivers/block/ |
D | ps3vram.c | 428 unsigned int cached, count; in ps3vram_read() local 449 cached = CACHE_OFFSET + entry * priv->cache.page_size + offset; in ps3vram_read() 453 (unsigned int)from, cached, offset, avail, count); in ps3vram_read() 457 memcpy(buf, priv->xdr_buf + cached, avail); in ps3vram_read() 472 unsigned int cached, count; in ps3vram_write() local 490 cached = CACHE_OFFSET + entry * priv->cache.page_size + offset; in ps3vram_write() 494 cached, offset, avail, count); in ps3vram_write() 498 memcpy(priv->xdr_buf + cached, buf, avail); in ps3vram_write()
|
/linux-6.12.1/arch/s390/appldata/ |
D | appldata_mem.c | 54 u64 cached; /* size of (used) cache, w/o buffers */ member 102 mem_data->cached = P2K(global_node_page_state(NR_FILE_PAGES) in appldata_get_mem_data()
|