/linux-6.12.1/fs/btrfs/ |
D | extent-io-tree.c | 624 u32 bits, struct extent_state **cached_state, in __clear_extent_bit() argument 663 if (cached_state) { in __clear_extent_bit() 664 cached = *cached_state; in __clear_extent_bit() 667 *cached_state = NULL; in __clear_extent_bit() 668 cached_state = NULL; in __clear_extent_bit() 783 u32 bits, struct extent_state **cached_state) in wait_extent_bit() argument 795 if (cached_state && *cached_state) { in wait_extent_bit() 796 state = *cached_state; in wait_extent_bit() 838 if (cached_state && *cached_state) { in wait_extent_bit() 839 state = *cached_state; in wait_extent_bit() [all …]
|
D | extent-io-tree.h | 163 struct extent_state **cached_state); 167 struct extent_state *cached_state); 197 u32 bits, struct extent_state **cached_state); 200 u64 end, struct extent_state **cached_state) in clear_extent_uptodate() argument 203 cached_state, NULL); in clear_extent_uptodate() 216 struct extent_state **cached_state); 220 struct extent_state **cached_state); 227 struct extent_state **cached_state);
|
D | file.c | 994 struct extent_state **cached_state) in lock_and_cleanup_extent_if_need() argument 1010 cached_state)) { in lock_and_cleanup_extent_if_need() 1020 lock_extent(&inode->io_tree, start_pos, last_pos, cached_state); in lock_and_cleanup_extent_if_need() 1029 cached_state); in lock_and_cleanup_extent_if_need() 1080 struct extent_state *cached_state = NULL; in btrfs_check_nocow_lock() local 1098 &cached_state)) { in btrfs_check_nocow_lock() 1104 &cached_state); in btrfs_check_nocow_lock() 1113 unlock_extent(&inode->io_tree, lockstart, lockend, &cached_state); in btrfs_check_nocow_lock() 1237 struct extent_state *cached_state = NULL; in btrfs_buffered_write() local 1336 &lockend, nowait, &cached_state); in btrfs_buffered_write() [all …]
|
D | direct-io.c | 34 struct extent_state **cached_state, in lock_extent_direct() argument 45 if (!try_lock_dio_extent(io_tree, lockstart, lockend, cached_state)) in lock_extent_direct() 48 lock_dio_extent(io_tree, lockstart, lockend, cached_state); in lock_extent_direct() 54 cached_state)) { in lock_extent_direct() 59 lock_extent(io_tree, lockstart, lockend, cached_state); in lock_extent_direct() 81 unlock_extent(io_tree, lockstart, lockend, cached_state); in lock_extent_direct() 134 unlock_dio_extent(io_tree, lockstart, lockend, cached_state); in lock_extent_direct() 361 struct extent_state *cached_state = NULL; in btrfs_dio_iomap_begin() local 453 ret = lock_extent_direct(inode, lockstart, lockend, &cached_state, flags); in btrfs_dio_iomap_begin() 579 unlock_bits, &cached_state); in btrfs_dio_iomap_begin() [all …]
|
D | defrag.c | 862 struct extent_state *cached_state = NULL; in defrag_prepare_one_folio() local 897 lock_extent(&inode->io_tree, page_start, page_end, &cached_state); in defrag_prepare_one_folio() 900 &cached_state); in defrag_prepare_one_folio() 1159 struct extent_state **cached_state) in defrag_one_locked_target() argument 1178 EXTENT_DEFRAG, cached_state); in defrag_one_locked_target() 1180 EXTENT_DELALLOC | EXTENT_DEFRAG, cached_state); in defrag_one_locked_target() 1197 struct extent_state *cached_state = NULL; in defrag_one_range() local 1231 &cached_state); in defrag_one_range() 1247 &cached_state); in defrag_one_range() 1259 &cached_state); in defrag_one_range()
|
D | ordered-data.h | 212 struct extent_state **cached_state); 214 struct extent_state **cached_state);
|
D | inode.c | 1794 struct extent_state *cached_state = NULL; in fallback_to_cow() local 1831 lock_extent(io_tree, start, end, &cached_state); in fallback_to_cow() 1850 unlock_extent(io_tree, start, end, &cached_state); in fallback_to_cow() 2047 struct extent_state *cached_state = NULL; in run_delalloc_nocow() local 2185 lock_extent(&inode->io_tree, cur_offset, nocow_end, &cached_state); in run_delalloc_nocow() 2196 nocow_end, &cached_state); in run_delalloc_nocow() 2216 nocow_end, &cached_state); in run_delalloc_nocow() 2231 locked_folio, &cached_state, in run_delalloc_nocow() 2650 struct extent_state **cached_state) in btrfs_find_new_delalloc_bytes() argument 2676 EXTENT_DELALLOC_NEW, cached_state); in btrfs_find_new_delalloc_bytes() [all …]
|
D | reflink.c | 638 struct extent_state *cached_state = NULL; in btrfs_extent_same_range() local 649 lock_extent(&BTRFS_I(dst)->io_tree, dst_loff, end, &cached_state); in btrfs_extent_same_range() 651 unlock_extent(&BTRFS_I(dst)->io_tree, dst_loff, end, &cached_state); in btrfs_extent_same_range() 703 struct extent_state *cached_state = NULL; in btrfs_clone_files() local 750 lock_extent(&BTRFS_I(inode)->io_tree, destoff, end, &cached_state); in btrfs_clone_files() 752 unlock_extent(&BTRFS_I(inode)->io_tree, destoff, end, &cached_state); in btrfs_clone_files()
|
D | ordered-data.c | 1153 struct extent_state **cached_state) in btrfs_lock_and_flush_ordered_range() argument 1159 if (cached_state) in btrfs_lock_and_flush_ordered_range() 1160 cachedp = cached_state; in btrfs_lock_and_flush_ordered_range() 1172 if (!cached_state) in btrfs_lock_and_flush_ordered_range() 1190 struct extent_state **cached_state) in btrfs_try_lock_ordered_range() argument 1194 if (!try_lock_extent(&inode->io_tree, start, end, cached_state)) in btrfs_try_lock_ordered_range() 1202 unlock_extent(&inode->io_tree, start, end, cached_state); in btrfs_try_lock_ordered_range()
|
D | file.h | 45 struct extent_state **cached_state,
|
D | transaction.c | 1122 struct extent_state *cached_state = NULL; in btrfs_write_marked_extents() local 1127 mark, &cached_state)) { in btrfs_write_marked_extents() 1132 mark, &cached_state); in btrfs_write_marked_extents() 1154 free_extent_state(cached_state); in btrfs_write_marked_extents() 1157 cached_state = NULL; in btrfs_write_marked_extents() 1174 struct extent_state *cached_state = NULL; in __btrfs_wait_marked_extents() local 1180 EXTENT_NEED_WAIT, &cached_state)) { in __btrfs_wait_marked_extents() 1190 EXTENT_NEED_WAIT, &cached_state); in __btrfs_wait_marked_extents() 1195 free_extent_state(cached_state); in __btrfs_wait_marked_extents() 1198 cached_state = NULL; in __btrfs_wait_marked_extents()
|
D | relocation.c | 997 struct extent_state *cached_state = NULL; in replace_file_extents() local 1009 end, &cached_state); in replace_file_extents() 1017 &cached_state); in replace_file_extents() 1428 struct extent_state *cached_state = NULL; in invalidate_extent_cache() local 1481 lock_extent(&inode->io_tree, start, end, &cached_state); in invalidate_extent_cache() 1483 unlock_extent(&inode->io_tree, start, end, &cached_state); in invalidate_extent_cache() 2810 struct extent_state *cached_state = NULL; in prealloc_file_extent_cluster() local 2818 lock_extent(&inode->io_tree, start, end, &cached_state); in prealloc_file_extent_cluster() 2824 unlock_extent(&inode->io_tree, start, end, &cached_state); in prealloc_file_extent_cluster() 2840 struct extent_state *cached_state = NULL; in setup_relocation_extent_mapping() local [all …]
|
D | fiemap.c | 635 struct extent_state *cached_state = NULL; in extent_fiemap() local 664 lock_extent(&inode->io_tree, range_start, range_end, &cached_state); in extent_fiemap() 844 unlock_extent(&inode->io_tree, range_start, range_end, &cached_state); in extent_fiemap()
|
D | extent_io.c | 325 struct extent_state *cached_state = NULL; in find_lock_delalloc_range() local 340 max_bytes, &cached_state); in find_lock_delalloc_range() 346 free_extent_state(cached_state); in find_lock_delalloc_range() 372 free_extent_state(cached_state); in find_lock_delalloc_range() 373 cached_state = NULL; in find_lock_delalloc_range() 385 lock_extent(tree, delalloc_start, delalloc_end, &cached_state); in find_lock_delalloc_range() 389 EXTENT_DELALLOC, cached_state); in find_lock_delalloc_range() 391 unlock_extent(tree, delalloc_start, delalloc_end, &cached_state); in find_lock_delalloc_range() 909 struct extent_state *cached_state = NULL; in __get_extent_map() local 925 btrfs_lock_and_flush_ordered_range(BTRFS_I(inode), start, start + len - 1, &cached_state); in __get_extent_map() [all …]
|
D | free-space-cache.c | 310 struct extent_state *cached_state = NULL; in btrfs_truncate_free_space_cache() local 343 lock_extent(&inode->io_tree, 0, (u64)-1, &cached_state); in btrfs_truncate_free_space_cache() 355 unlock_extent(&inode->io_tree, 0, (u64)-1, &cached_state); in btrfs_truncate_free_space_cache() 1291 struct extent_state **cached_state) in cleanup_write_cache_enospc() argument 1295 cached_state); in cleanup_write_cache_enospc() 1384 struct extent_state *cached_state = NULL; in __btrfs_write_out_cache() local 1420 &cached_state); in __btrfs_write_out_cache() 1462 &cached_state, false); in __btrfs_write_out_cache() 1476 &cached_state); in __btrfs_write_out_cache() 1498 cleanup_write_cache_enospc(inode, io_ctl, &cached_state); in __btrfs_write_out_cache()
|
D | dev-replace.c | 800 struct extent_state *cached_state = NULL; in btrfs_set_target_alloc_state() local 810 CHUNK_ALLOCATED, &cached_state)) { in btrfs_set_target_alloc_state() 818 free_extent_state(cached_state); in btrfs_set_target_alloc_state()
|
/linux-6.12.1/drivers/gpu/drm/msm/dsi/phy/ |
D | dsi_phy_28nm_8960.c | 65 struct pll_28nm_cached_state cached_state; member 341 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_28nm_pll_save_state() local 344 cached_state->postdiv3 = in dsi_28nm_pll_save_state() 346 cached_state->postdiv2 = in dsi_28nm_pll_save_state() 348 cached_state->postdiv1 = in dsi_28nm_pll_save_state() 351 cached_state->vco_rate = clk_hw_get_rate(phy->vco_hw); in dsi_28nm_pll_save_state() 357 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_28nm_pll_restore_state() local 362 cached_state->vco_rate, 0); in dsi_28nm_pll_restore_state() 369 writel(cached_state->postdiv3, base + REG_DSI_28nm_8960_PHY_PLL_CTRL_10); in dsi_28nm_pll_restore_state() 370 writel(cached_state->postdiv2, base + REG_DSI_28nm_8960_PHY_PLL_CTRL_9); in dsi_28nm_pll_restore_state() [all …]
|
D | dsi_phy_28nm.c | 74 struct pll_28nm_cached_state cached_state; member 582 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_28nm_pll_save_state() local 585 cached_state->postdiv3 = in dsi_28nm_pll_save_state() 587 cached_state->postdiv1 = in dsi_28nm_pll_save_state() 589 cached_state->byte_mux = readl(base + REG_DSI_28nm_PHY_PLL_VREG_CFG); in dsi_28nm_pll_save_state() 591 cached_state->vco_rate = clk_hw_get_rate(phy->vco_hw); in dsi_28nm_pll_save_state() 593 cached_state->vco_rate = 0; in dsi_28nm_pll_save_state() 599 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_28nm_pll_restore_state() local 604 cached_state->vco_rate, 0); in dsi_28nm_pll_restore_state() 611 writel(cached_state->postdiv3, base + REG_DSI_28nm_PHY_PLL_POSTDIV3_CFG); in dsi_28nm_pll_restore_state() [all …]
|
D | dsi_phy_14nm.c | 79 struct pll_14nm_cached_state cached_state; member 693 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_14nm_pll_save_state() local 699 cached_state->n1postdiv = data & 0xf; in dsi_14nm_pll_save_state() 700 cached_state->n2postdiv = (data >> 4) & 0xf; in dsi_14nm_pll_save_state() 703 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_14nm_pll_save_state() 705 cached_state->vco_rate = clk_hw_get_rate(phy->vco_hw); in dsi_14nm_pll_save_state() 711 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_14nm_pll_restore_state() local 717 cached_state->vco_rate, 0); in dsi_14nm_pll_restore_state() 724 data = cached_state->n1postdiv | (cached_state->n2postdiv << 4); in dsi_14nm_pll_restore_state() 727 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_14nm_pll_restore_state()
|
D | dsi_phy_10nm.c | 79 struct pll_10nm_cached_state cached_state; member 474 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_10nm_pll_save_state() 497 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_10nm_pll_restore_state()
|
D | dsi_phy_7nm.c | 86 struct pll_7nm_cached_state cached_state; member 544 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_save_state() 567 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_restore_state()
|
/linux-6.12.1/drivers/iio/multiplexer/ |
D | iio-mux.c | 30 int cached_state; member 49 mux->cached_state = -1; in iio_mux_select() 53 if (mux->cached_state == chan->channel) in iio_mux_select() 72 mux->cached_state = -1; in iio_mux_select() 77 mux->cached_state = chan->channel; in iio_mux_select() 391 mux->cached_state = -1; in mux_probe()
|
/linux-6.12.1/drivers/mux/ |
D | core.c | 130 mux->cached_state = MUX_CACHE_UNKNOWN; in mux_chip_alloc() 145 mux->cached_state = ret < 0 ? MUX_CACHE_UNKNOWN : state; in mux_control_set() 171 if (mux->idle_state == mux->cached_state) in mux_chip_register() 319 if (mux->cached_state == state) in __mux_control_select() 483 mux->idle_state != mux->cached_state) in mux_control_deselect()
|
/linux-6.12.1/include/linux/mux/ |
D | driver.h | 48 int cached_state; member
|
/linux-6.12.1/drivers/gpu/drm/amd/display/amdgpu_dm/ |
D | amdgpu_dm.h | 520 struct drm_atomic_state *cached_state; member
|