Home
last modified time | relevance | path

Searched refs:cached_state (Results 1 – 25 of 30) sorted by relevance

12

/linux-6.12.1/fs/btrfs/
Dextent-io-tree.c624 u32 bits, struct extent_state **cached_state, in __clear_extent_bit() argument
663 if (cached_state) { in __clear_extent_bit()
664 cached = *cached_state; in __clear_extent_bit()
667 *cached_state = NULL; in __clear_extent_bit()
668 cached_state = NULL; in __clear_extent_bit()
783 u32 bits, struct extent_state **cached_state) in wait_extent_bit() argument
795 if (cached_state && *cached_state) { in wait_extent_bit()
796 state = *cached_state; in wait_extent_bit()
838 if (cached_state && *cached_state) { in wait_extent_bit()
839 state = *cached_state; in wait_extent_bit()
[all …]
Dextent-io-tree.h163 struct extent_state **cached_state);
167 struct extent_state *cached_state);
197 u32 bits, struct extent_state **cached_state);
200 u64 end, struct extent_state **cached_state) in clear_extent_uptodate() argument
203 cached_state, NULL); in clear_extent_uptodate()
216 struct extent_state **cached_state);
220 struct extent_state **cached_state);
227 struct extent_state **cached_state);
Dfile.c994 struct extent_state **cached_state) in lock_and_cleanup_extent_if_need() argument
1010 cached_state)) { in lock_and_cleanup_extent_if_need()
1020 lock_extent(&inode->io_tree, start_pos, last_pos, cached_state); in lock_and_cleanup_extent_if_need()
1029 cached_state); in lock_and_cleanup_extent_if_need()
1080 struct extent_state *cached_state = NULL; in btrfs_check_nocow_lock() local
1098 &cached_state)) { in btrfs_check_nocow_lock()
1104 &cached_state); in btrfs_check_nocow_lock()
1113 unlock_extent(&inode->io_tree, lockstart, lockend, &cached_state); in btrfs_check_nocow_lock()
1237 struct extent_state *cached_state = NULL; in btrfs_buffered_write() local
1336 &lockend, nowait, &cached_state); in btrfs_buffered_write()
[all …]
Ddirect-io.c34 struct extent_state **cached_state, in lock_extent_direct() argument
45 if (!try_lock_dio_extent(io_tree, lockstart, lockend, cached_state)) in lock_extent_direct()
48 lock_dio_extent(io_tree, lockstart, lockend, cached_state); in lock_extent_direct()
54 cached_state)) { in lock_extent_direct()
59 lock_extent(io_tree, lockstart, lockend, cached_state); in lock_extent_direct()
81 unlock_extent(io_tree, lockstart, lockend, cached_state); in lock_extent_direct()
134 unlock_dio_extent(io_tree, lockstart, lockend, cached_state); in lock_extent_direct()
361 struct extent_state *cached_state = NULL; in btrfs_dio_iomap_begin() local
453 ret = lock_extent_direct(inode, lockstart, lockend, &cached_state, flags); in btrfs_dio_iomap_begin()
579 unlock_bits, &cached_state); in btrfs_dio_iomap_begin()
[all …]
Ddefrag.c862 struct extent_state *cached_state = NULL; in defrag_prepare_one_folio() local
897 lock_extent(&inode->io_tree, page_start, page_end, &cached_state); in defrag_prepare_one_folio()
900 &cached_state); in defrag_prepare_one_folio()
1159 struct extent_state **cached_state) in defrag_one_locked_target() argument
1178 EXTENT_DEFRAG, cached_state); in defrag_one_locked_target()
1180 EXTENT_DELALLOC | EXTENT_DEFRAG, cached_state); in defrag_one_locked_target()
1197 struct extent_state *cached_state = NULL; in defrag_one_range() local
1231 &cached_state); in defrag_one_range()
1247 &cached_state); in defrag_one_range()
1259 &cached_state); in defrag_one_range()
Dordered-data.h212 struct extent_state **cached_state);
214 struct extent_state **cached_state);
Dinode.c1794 struct extent_state *cached_state = NULL; in fallback_to_cow() local
1831 lock_extent(io_tree, start, end, &cached_state); in fallback_to_cow()
1850 unlock_extent(io_tree, start, end, &cached_state); in fallback_to_cow()
2047 struct extent_state *cached_state = NULL; in run_delalloc_nocow() local
2185 lock_extent(&inode->io_tree, cur_offset, nocow_end, &cached_state); in run_delalloc_nocow()
2196 nocow_end, &cached_state); in run_delalloc_nocow()
2216 nocow_end, &cached_state); in run_delalloc_nocow()
2231 locked_folio, &cached_state, in run_delalloc_nocow()
2650 struct extent_state **cached_state) in btrfs_find_new_delalloc_bytes() argument
2676 EXTENT_DELALLOC_NEW, cached_state); in btrfs_find_new_delalloc_bytes()
[all …]
Dreflink.c638 struct extent_state *cached_state = NULL; in btrfs_extent_same_range() local
649 lock_extent(&BTRFS_I(dst)->io_tree, dst_loff, end, &cached_state); in btrfs_extent_same_range()
651 unlock_extent(&BTRFS_I(dst)->io_tree, dst_loff, end, &cached_state); in btrfs_extent_same_range()
703 struct extent_state *cached_state = NULL; in btrfs_clone_files() local
750 lock_extent(&BTRFS_I(inode)->io_tree, destoff, end, &cached_state); in btrfs_clone_files()
752 unlock_extent(&BTRFS_I(inode)->io_tree, destoff, end, &cached_state); in btrfs_clone_files()
Dordered-data.c1153 struct extent_state **cached_state) in btrfs_lock_and_flush_ordered_range() argument
1159 if (cached_state) in btrfs_lock_and_flush_ordered_range()
1160 cachedp = cached_state; in btrfs_lock_and_flush_ordered_range()
1172 if (!cached_state) in btrfs_lock_and_flush_ordered_range()
1190 struct extent_state **cached_state) in btrfs_try_lock_ordered_range() argument
1194 if (!try_lock_extent(&inode->io_tree, start, end, cached_state)) in btrfs_try_lock_ordered_range()
1202 unlock_extent(&inode->io_tree, start, end, cached_state); in btrfs_try_lock_ordered_range()
Dfile.h45 struct extent_state **cached_state,
Dtransaction.c1122 struct extent_state *cached_state = NULL; in btrfs_write_marked_extents() local
1127 mark, &cached_state)) { in btrfs_write_marked_extents()
1132 mark, &cached_state); in btrfs_write_marked_extents()
1154 free_extent_state(cached_state); in btrfs_write_marked_extents()
1157 cached_state = NULL; in btrfs_write_marked_extents()
1174 struct extent_state *cached_state = NULL; in __btrfs_wait_marked_extents() local
1180 EXTENT_NEED_WAIT, &cached_state)) { in __btrfs_wait_marked_extents()
1190 EXTENT_NEED_WAIT, &cached_state); in __btrfs_wait_marked_extents()
1195 free_extent_state(cached_state); in __btrfs_wait_marked_extents()
1198 cached_state = NULL; in __btrfs_wait_marked_extents()
Drelocation.c997 struct extent_state *cached_state = NULL; in replace_file_extents() local
1009 end, &cached_state); in replace_file_extents()
1017 &cached_state); in replace_file_extents()
1428 struct extent_state *cached_state = NULL; in invalidate_extent_cache() local
1481 lock_extent(&inode->io_tree, start, end, &cached_state); in invalidate_extent_cache()
1483 unlock_extent(&inode->io_tree, start, end, &cached_state); in invalidate_extent_cache()
2810 struct extent_state *cached_state = NULL; in prealloc_file_extent_cluster() local
2818 lock_extent(&inode->io_tree, start, end, &cached_state); in prealloc_file_extent_cluster()
2824 unlock_extent(&inode->io_tree, start, end, &cached_state); in prealloc_file_extent_cluster()
2840 struct extent_state *cached_state = NULL; in setup_relocation_extent_mapping() local
[all …]
Dfiemap.c635 struct extent_state *cached_state = NULL; in extent_fiemap() local
664 lock_extent(&inode->io_tree, range_start, range_end, &cached_state); in extent_fiemap()
844 unlock_extent(&inode->io_tree, range_start, range_end, &cached_state); in extent_fiemap()
Dextent_io.c325 struct extent_state *cached_state = NULL; in find_lock_delalloc_range() local
340 max_bytes, &cached_state); in find_lock_delalloc_range()
346 free_extent_state(cached_state); in find_lock_delalloc_range()
372 free_extent_state(cached_state); in find_lock_delalloc_range()
373 cached_state = NULL; in find_lock_delalloc_range()
385 lock_extent(tree, delalloc_start, delalloc_end, &cached_state); in find_lock_delalloc_range()
389 EXTENT_DELALLOC, cached_state); in find_lock_delalloc_range()
391 unlock_extent(tree, delalloc_start, delalloc_end, &cached_state); in find_lock_delalloc_range()
909 struct extent_state *cached_state = NULL; in __get_extent_map() local
925 btrfs_lock_and_flush_ordered_range(BTRFS_I(inode), start, start + len - 1, &cached_state); in __get_extent_map()
[all …]
Dfree-space-cache.c310 struct extent_state *cached_state = NULL; in btrfs_truncate_free_space_cache() local
343 lock_extent(&inode->io_tree, 0, (u64)-1, &cached_state); in btrfs_truncate_free_space_cache()
355 unlock_extent(&inode->io_tree, 0, (u64)-1, &cached_state); in btrfs_truncate_free_space_cache()
1291 struct extent_state **cached_state) in cleanup_write_cache_enospc() argument
1295 cached_state); in cleanup_write_cache_enospc()
1384 struct extent_state *cached_state = NULL; in __btrfs_write_out_cache() local
1420 &cached_state); in __btrfs_write_out_cache()
1462 &cached_state, false); in __btrfs_write_out_cache()
1476 &cached_state); in __btrfs_write_out_cache()
1498 cleanup_write_cache_enospc(inode, io_ctl, &cached_state); in __btrfs_write_out_cache()
Ddev-replace.c800 struct extent_state *cached_state = NULL; in btrfs_set_target_alloc_state() local
810 CHUNK_ALLOCATED, &cached_state)) { in btrfs_set_target_alloc_state()
818 free_extent_state(cached_state); in btrfs_set_target_alloc_state()
/linux-6.12.1/drivers/gpu/drm/msm/dsi/phy/
Ddsi_phy_28nm_8960.c65 struct pll_28nm_cached_state cached_state; member
341 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_28nm_pll_save_state() local
344 cached_state->postdiv3 = in dsi_28nm_pll_save_state()
346 cached_state->postdiv2 = in dsi_28nm_pll_save_state()
348 cached_state->postdiv1 = in dsi_28nm_pll_save_state()
351 cached_state->vco_rate = clk_hw_get_rate(phy->vco_hw); in dsi_28nm_pll_save_state()
357 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_28nm_pll_restore_state() local
362 cached_state->vco_rate, 0); in dsi_28nm_pll_restore_state()
369 writel(cached_state->postdiv3, base + REG_DSI_28nm_8960_PHY_PLL_CTRL_10); in dsi_28nm_pll_restore_state()
370 writel(cached_state->postdiv2, base + REG_DSI_28nm_8960_PHY_PLL_CTRL_9); in dsi_28nm_pll_restore_state()
[all …]
Ddsi_phy_28nm.c74 struct pll_28nm_cached_state cached_state; member
582 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_28nm_pll_save_state() local
585 cached_state->postdiv3 = in dsi_28nm_pll_save_state()
587 cached_state->postdiv1 = in dsi_28nm_pll_save_state()
589 cached_state->byte_mux = readl(base + REG_DSI_28nm_PHY_PLL_VREG_CFG); in dsi_28nm_pll_save_state()
591 cached_state->vco_rate = clk_hw_get_rate(phy->vco_hw); in dsi_28nm_pll_save_state()
593 cached_state->vco_rate = 0; in dsi_28nm_pll_save_state()
599 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_28nm_pll_restore_state() local
604 cached_state->vco_rate, 0); in dsi_28nm_pll_restore_state()
611 writel(cached_state->postdiv3, base + REG_DSI_28nm_PHY_PLL_POSTDIV3_CFG); in dsi_28nm_pll_restore_state()
[all …]
Ddsi_phy_14nm.c79 struct pll_14nm_cached_state cached_state; member
693 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_14nm_pll_save_state() local
699 cached_state->n1postdiv = data & 0xf; in dsi_14nm_pll_save_state()
700 cached_state->n2postdiv = (data >> 4) & 0xf; in dsi_14nm_pll_save_state()
703 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_14nm_pll_save_state()
705 cached_state->vco_rate = clk_hw_get_rate(phy->vco_hw); in dsi_14nm_pll_save_state()
711 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_14nm_pll_restore_state() local
717 cached_state->vco_rate, 0); in dsi_14nm_pll_restore_state()
724 data = cached_state->n1postdiv | (cached_state->n2postdiv << 4); in dsi_14nm_pll_restore_state()
727 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_14nm_pll_restore_state()
Ddsi_phy_10nm.c79 struct pll_10nm_cached_state cached_state; member
474 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_10nm_pll_save_state()
497 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_10nm_pll_restore_state()
Ddsi_phy_7nm.c86 struct pll_7nm_cached_state cached_state; member
544 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_save_state()
567 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_restore_state()
/linux-6.12.1/drivers/iio/multiplexer/
Diio-mux.c30 int cached_state; member
49 mux->cached_state = -1; in iio_mux_select()
53 if (mux->cached_state == chan->channel) in iio_mux_select()
72 mux->cached_state = -1; in iio_mux_select()
77 mux->cached_state = chan->channel; in iio_mux_select()
391 mux->cached_state = -1; in mux_probe()
/linux-6.12.1/drivers/mux/
Dcore.c130 mux->cached_state = MUX_CACHE_UNKNOWN; in mux_chip_alloc()
145 mux->cached_state = ret < 0 ? MUX_CACHE_UNKNOWN : state; in mux_control_set()
171 if (mux->idle_state == mux->cached_state) in mux_chip_register()
319 if (mux->cached_state == state) in __mux_control_select()
483 mux->idle_state != mux->cached_state) in mux_control_deselect()
/linux-6.12.1/include/linux/mux/
Ddriver.h48 int cached_state; member
/linux-6.12.1/drivers/gpu/drm/amd/display/amdgpu_dm/
Damdgpu_dm.h520 struct drm_atomic_state *cached_state; member

12