Home
last modified time | relevance | path

Searched refs:vdo (Results 1 – 25 of 67) sorted by relevance

123

/linux-6.12.1/drivers/md/dm-vdo/
Dvdo.c89 static bool vdo_is_equal(struct vdo *vdo, const void *context) in vdo_is_equal() argument
91 return (vdo == context); in vdo_is_equal()
103 static struct vdo * __must_check filter_vdos_locked(vdo_filter_fn filter, in filter_vdos_locked()
106 struct vdo *vdo; in filter_vdos_locked() local
108 list_for_each_entry(vdo, &registry.links, registration) { in filter_vdos_locked()
109 if (filter(vdo, context)) in filter_vdos_locked()
110 return vdo; in filter_vdos_locked()
121 struct vdo *vdo_find_matching(vdo_filter_fn filter, const void *context) in vdo_find_matching()
123 struct vdo *vdo; in vdo_find_matching() local
126 vdo = filter_vdos_locked(filter, context); in vdo_find_matching()
[all …]
Ddm-vdo-target.c890 static struct vdo *get_vdo_for_target(struct dm_target *ti) in get_vdo_for_target()
892 return ((struct device_config *) ti->private)->vdo; in get_vdo_for_target()
898 struct vdo *vdo = get_vdo_for_target(ti); in vdo_map_bio() local
900 const struct admin_state_code *code = vdo_get_admin_state_code(&vdo->admin.state); in vdo_map_bio()
906 vdo_count_bios(&vdo->stats.bios_in, bio); in vdo_map_bio()
911 vdo_launch_flush(vdo, bio); in vdo_map_bio()
918 (vdo == vdo_get_work_queue_owner(current_work_queue)->vdo)); in vdo_map_bio()
919 vdo_launch_bio(vdo->data_vio_pool, bio); in vdo_map_bio()
925 struct vdo *vdo = get_vdo_for_target(ti); in vdo_io_hints() local
927 limits->logical_block_size = vdo->device_config->logical_block_size; in vdo_io_hints()
[all …]
Dvdo.h57 struct vdo *vdo; member
166 struct vdo { struct
275 static inline bool vdo_uses_bio_ack_queue(struct vdo *vdo) in vdo_uses_bio_ack_queue() argument
277 return vdo->device_config->thread_counts.bio_ack_threads > 0; in vdo_uses_bio_ack_queue()
285 typedef bool (*vdo_filter_fn)(struct vdo *vdo, const void *context);
288 struct vdo * __must_check vdo_find_matching(vdo_filter_fn filter, const void *context);
290 int __must_check vdo_make_thread(struct vdo *vdo, thread_id_t thread_id,
294 static inline int __must_check vdo_make_default_thread(struct vdo *vdo, in vdo_make_default_thread() argument
297 return vdo_make_thread(vdo, thread_id, NULL, 1, NULL); in vdo_make_default_thread()
301 char **reason, struct vdo **vdo_ptr);
[all …]
Dflush.c26 struct vdo *vdo; member
117 vdo_initialize_completion(&flush->completion, flusher->vdo, in allocate_flush()
135 int vdo_make_flusher(struct vdo *vdo) in vdo_make_flusher() argument
137 int result = vdo_allocate(1, struct flusher, __func__, &vdo->flusher); in vdo_make_flusher()
142 vdo->flusher->vdo = vdo; in vdo_make_flusher()
143 vdo->flusher->thread_id = vdo->thread_config.packer_thread; in vdo_make_flusher()
144 vdo_set_admin_state_code(&vdo->flusher->state, VDO_ADMIN_STATE_NORMAL_OPERATION); in vdo_make_flusher()
145 vdo_initialize_completion(&vdo->flusher->completion, vdo, in vdo_make_flusher()
148 spin_lock_init(&vdo->flusher->lock); in vdo_make_flusher()
149 bio_list_init(&vdo->flusher->waiting_flush_bios); in vdo_make_flusher()
[all …]
Drepair.c207 const struct thread_config *thread_config = &completion->vdo->thread_config; in prepare_repair_completion()
242 repair->completion.vdo->block_map->zones[0].page_cache.rebuilding = false; in free_repair_completion()
253 struct vdo *vdo = completion->vdo; in finish_repair() local
256 vdo_assert_on_admin_thread(vdo, __func__); in finish_repair()
258 if (vdo->load_state != VDO_REBUILD_FOR_UPGRADE) in finish_repair()
259 vdo->states.vdo.complete_recoveries++; in finish_repair()
261 vdo_initialize_recovery_journal_post_repair(vdo->recovery_journal, in finish_repair()
262 vdo->states.vdo.complete_recoveries, in finish_repair()
268 if (vdo_state_requires_read_only_rebuild(vdo->load_state)) { in finish_repair()
281 vdo_continue_completion(parent, vdo_allocate_reference_counters(vdo->depot)); in finish_repair()
[all …]
Ddump.c55 static void do_dump(struct vdo *vdo, unsigned int dump_options_requested, in do_dump() argument
62 active = get_data_vio_pool_active_requests(vdo->data_vio_pool); in do_dump()
63 maximum = get_data_vio_pool_maximum_requests(vdo->data_vio_pool); in do_dump()
64 outstanding = (atomic64_read(&vdo->stats.bios_submitted) - in do_dump()
65 atomic64_read(&vdo->stats.bios_completed)); in do_dump()
68 vdo_get_device_name(vdo->device_config->owning_target)); in do_dump()
69 if (((dump_options_requested & FLAG_SHOW_QUEUES) != 0) && (vdo->threads != NULL)) { in do_dump()
72 for (id = 0; id < vdo->thread_config.thread_count; id++) in do_dump()
73 vdo_dump_work_queue(vdo->threads[id].queue); in do_dump()
76 vdo_dump_hash_zones(vdo->hash_zones); in do_dump()
[all …]
Dcompletion.c52 struct vdo *vdo, in vdo_initialize_completion() argument
56 completion->vdo = vdo; in vdo_initialize_completion()
111 struct vdo *vdo = completion->vdo; in vdo_enqueue_completion() local
114 if (VDO_ASSERT(thread_id < vdo->thread_config.thread_count, in vdo_enqueue_completion()
117 vdo->thread_config.thread_count) != VDO_SUCCESS) in vdo_enqueue_completion()
123 vdo_enqueue_work_queue(vdo->threads[thread_id].queue, completion); in vdo_enqueue_completion()
Dlogical-zone.c54 struct vdo *vdo = zones->vdo; in initialize_zone() local
62 if (zone_number < vdo->thread_config.logical_zone_count - 1) in initialize_zone()
65 vdo_initialize_completion(&zone->completion, vdo, in initialize_zone()
69 zone->thread_id = vdo->thread_config.logical_threads[zone_number]; in initialize_zone()
70 zone->block_map_zone = &vdo->block_map->zones[zone_number]; in initialize_zone()
74 allocation_zone_number = zone->thread_id % vdo->thread_config.physical_zone_count; in initialize_zone()
75 zone->allocation_zone = &vdo->physical_zones->zones[allocation_zone_number]; in initialize_zone()
77 return vdo_make_default_thread(vdo, zone->thread_id); in initialize_zone()
87 int vdo_make_logical_zones(struct vdo *vdo, struct logical_zones **zones_ptr) in vdo_make_logical_zones() argument
92 zone_count_t zone_count = vdo->thread_config.logical_zone_count; in vdo_make_logical_zones()
[all …]
Dvio.h56 return vio->completion.vdo->thread_config.bio_threads[vio->bio_zone]; in get_vio_bio_zone_thread_id()
78 int allocate_vio_components(struct vdo *vdo, enum vio_type vio_type,
81 int __must_check create_multi_block_metadata_vio(struct vdo *vdo, enum vio_type vio_type,
86 static inline int __must_check create_metadata_vio(struct vdo *vdo, enum vio_type vio_type, in create_metadata_vio() argument
91 return create_multi_block_metadata_vio(vdo, vio_type, priority, parent, 1, data, in create_metadata_vio()
109 enum vio_priority priority, struct vdo *vdo) in initialize_vio() argument
118 vdo_initialize_completion(&vio->completion, vdo, VIO_COMPLETION); in initialize_vio()
191 int __must_check make_vio_pool(struct vdo *vdo, size_t pool_size, thread_id_t thread_id,
Dvio.c44 struct vdo *vdo = vio->completion.vdo; in pbn_from_vio_bio() local
47 return ((pbn == VDO_GEOMETRY_BLOCK_LOCATION) ? pbn : pbn + vdo->geometry.bio_offset); in pbn_from_vio_bio()
78 int allocate_vio_components(struct vdo *vdo, enum vio_type vio_type, in allocate_vio_components() argument
100 initialize_vio(vio, bio, block_count, vio_type, priority, vdo); in allocate_vio_components()
118 int create_multi_block_metadata_vio(struct vdo *vdo, enum vio_type vio_type, in create_multi_block_metadata_vio() argument
138 result = allocate_vio_components(vdo, vio_type, priority, parent, block_count, in create_multi_block_metadata_vio()
176 struct vdo *vdo = vio->completion.vdo; in vdo_set_bio_properties() local
177 struct device_config *config = vdo->device_config; in vdo_set_bio_properties()
179 pbn -= vdo->geometry.bio_offset; in vdo_set_bio_properties()
252 struct vdo *vdo = vio->completion.vdo; in update_vio_error_stats() local
[all …]
Dio-submitter.c76 struct atomic_statistics *stats = &vio->completion.vdo->stats; in count_all_bios()
109 struct vdo *vdo = vio->completion.vdo; in send_bio_to_device() local
112 atomic64_inc(&vdo->stats.bios_submitted); in send_bio_to_device()
114 bio_set_dev(bio, vdo_get_backing_device(vdo)); in send_bio_to_device()
141 struct io_submitter *submitter = vio->completion.vdo->io_submitter; in get_bio_list()
273 struct vdo *vdo = vio->completion.vdo; in try_bio_map_merge() local
275 &vdo->io_submitter->bio_queue_data[vio->bio_zone]; in try_bio_map_merge()
345 const struct admin_state_code *code = vdo_get_admin_state(completion->vdo); in __submit_metadata_vio()
375 unsigned int max_requests_active, struct vdo *vdo, in vdo_make_io_submitter() argument
417 result = vdo_make_thread(vdo, vdo->thread_config.bio_threads[i], in vdo_make_io_submitter()
[all …]
Dmessage-stats.h11 int vdo_write_config(struct vdo *vdo, char **buf, unsigned int *maxlen);
12 int vdo_write_stats(struct vdo *vdo, char *buf, unsigned int maxlen);
Ddump.h11 int vdo_dump(struct vdo *vdo, unsigned int argc, char *const *argv, const char *why);
13 void vdo_dump_all(struct vdo *vdo, const char *why);
Ddata-vio.c244 struct vdo *vdo = vdo_from_data_vio(data_vio); in initialize_lbn_lock() local
252 lock->zone = &vdo->logical_zones->zones[zone_number]; in initialize_lbn_lock()
259 struct vdo *vdo = vdo_from_data_vio(data_vio); in launch_locked_request() local
261 if (vdo_is_read_only(vdo)) { in launch_locked_request()
273 struct vdo *vdo = vdo_from_data_vio(data_vio); in acknowledge_data_vio() local
285 vdo_count_bios(&vdo->stats.bios_acknowledged, bio); in acknowledge_data_vio()
287 vdo_count_bios(&vdo->stats.bios_acknowledged_partial, bio); in acknowledge_data_vio()
422 struct vdo *vdo = vdo_from_data_vio(data_vio); in attempt_logical_block_lock() local
428 if (data_vio->logical.lbn >= vdo->states.vdo.config.logical_blocks) { in attempt_logical_block_lock()
534 static void launch_bio(struct vdo *vdo, struct data_vio *data_vio, struct bio *bio) in launch_bio() argument
[all …]
Dflush.h28 int __must_check vdo_make_flusher(struct vdo *vdo);
38 void vdo_launch_flush(struct vdo *vdo, struct bio *bio);
Dlogical-zone.h57 struct vdo *vdo; member
66 int __must_check vdo_make_logical_zones(struct vdo *vdo,
Dphysical-zone.c325 static int initialize_zone(struct vdo *vdo, struct physical_zones *zones) in initialize_zone() argument
342 zone->thread_id = vdo->thread_config.physical_threads[zone_number]; in initialize_zone()
343 zone->allocator = &vdo->depot->allocators[zone_number]; in initialize_zone()
344 zone->next = &zones->zones[(zone_number + 1) % vdo->thread_config.physical_zone_count]; in initialize_zone()
345 result = vdo_make_default_thread(vdo, zone->thread_id); in initialize_zone()
361 int vdo_make_physical_zones(struct vdo *vdo, struct physical_zones **zones_ptr) in vdo_make_physical_zones() argument
365 zone_count_t zone_count = vdo->thread_config.physical_zone_count; in vdo_make_physical_zones()
376 result = initialize_zone(vdo, zones); in vdo_make_physical_zones()
Dblock-map.h51 struct vdo *vdo; member
255 struct vdo *vdo; member
334 block_count_t logical_blocks, struct vdo *vdo,
Dtypes.h211 struct vdo *vdo; member
314 struct vdo *vdo; member
333 struct vdo;
DMakefile5 obj-$(CONFIG_DM_VDO) += dm-vdo.o
7 dm-vdo-objs := \
14 dm-vdo-target.o \
40 vdo.o \
/linux-6.12.1/include/linux/usb/
Dpd_vdo.h78 #define PD_VDO_VID(vdo) ((vdo) >> 16) argument
79 #define PD_VDO_SVDM(vdo) (((vdo) >> 15) & 1) argument
80 #define PD_VDO_SVDM_VER(vdo) (((vdo) >> 13) & 0x3) argument
81 #define PD_VDO_OPOS(vdo) (((vdo) >> 8) & 0x7) argument
82 #define PD_VDO_CMD(vdo) ((vdo) & 0x1f) argument
83 #define PD_VDO_CMDT(vdo) (((vdo) >> 6) & 0x3) argument
154 #define PD_IDH_PTYPE(vdo) (((vdo) >> 27) & 0x7) argument
155 #define PD_IDH_VID(vdo) ((vdo) & 0xffff) argument
156 #define PD_IDH_MODAL_SUPP(vdo) ((vdo) & (1 << 26)) argument
157 #define PD_IDH_DFP_PTYPE(vdo) (((vdo) >> 23) & 0x7) argument
[all …]
Dtypec_altmode.h29 u32 vdo; member
60 int (*enter)(struct typec_altmode *altmode, u32 *vdo);
62 void (*attention)(struct typec_altmode *altmode, u32 vdo);
64 const u32 *vdo, int cnt);
70 int typec_altmode_enter(struct typec_altmode *altmode, u32 *vdo);
72 int typec_altmode_attention(struct typec_altmode *altmode, u32 vdo);
74 const u32 header, const u32 *vdo, int count);
87 int (*enter)(struct typec_altmode *altmode, enum typec_plug_index sop, u32 *vdo);
90 const u32 hdr, const u32 *vdo, int cnt);
93 int typec_cable_altmode_enter(struct typec_altmode *altmode, enum typec_plug_index sop, u32 *vdo);
[all …]
/linux-6.12.1/Documentation/admin-guide/device-mapper/
Dvdo.rst3 dm-vdo
6 The dm-vdo (virtual data optimizer) device mapper target provides
9 with any file system. The vdo target does not protect against data
11 it. It is strongly recommended that lvm be used to manage vdo volumes. See
17 Formatting a vdo volume requires the use of the 'vdoformat' tool, available
20 https://github.com/dm-vdo/vdo/
22 In most cases, a vdo target will recover from a crash automatically the
26 data-loss, a positive action must be taken to bring vdo out of read-only
28 prepare a read-only vdo to exit read-only mode. After running this tool,
29 the vdo target will rebuild its metadata the next time it is
[all …]
/linux-6.12.1/drivers/usb/typec/altmodes/
Ddisplayport.c105 pin_assign = DP_CAP_UFP_D_PIN_ASSIGN(dp->alt->vdo) & in dp_altmode_configure()
106 DP_CAP_DFP_D_PIN_ASSIGN(dp->port->vdo); in dp_altmode_configure()
109 pin_assign &= DP_CAP_DFP_D_PIN_ASSIGN(dp->plug_prime->vdo); in dp_altmode_configure()
114 pin_assign = DP_CAP_PIN_ASSIGN_UFP_D(dp->alt->vdo) & in dp_altmode_configure()
115 DP_CAP_PIN_ASSIGN_DFP_D(dp->port->vdo); in dp_altmode_configure()
118 pin_assign &= DP_CAP_UFP_D_PIN_ASSIGN(dp->plug_prime->vdo); in dp_altmode_configure()
246 u32 vdo; in dp_altmode_work() local
275 vdo = 1; in dp_altmode_work()
276 ret = typec_altmode_vdm(dp->alt, header, &vdo, 2); in dp_altmode_work()
322 static void dp_altmode_attention(struct typec_altmode *alt, const u32 vdo) in dp_altmode_attention() argument
[all …]
/linux-6.12.1/drivers/usb/typec/
Dbus.c126 int typec_altmode_enter(struct typec_altmode *adev, u32 *vdo) in typec_altmode_enter() argument
147 return pdev->ops->enter(pdev, vdo); in typec_altmode_enter()
186 int typec_altmode_attention(struct typec_altmode *adev, u32 vdo) in typec_altmode_attention() argument
197 pdev->ops->attention(pdev, vdo); in typec_altmode_attention()
215 const u32 header, const u32 *vdo, int count) in typec_altmode_vdm() argument
233 return pdev->ops->vdm(pdev, header, vdo, count); in typec_altmode_vdm()
259 int typec_cable_altmode_enter(struct typec_altmode *adev, enum typec_plug_index sop, u32 *vdo) in typec_cable_altmode_enter() argument
278 return pdev->cable_ops->enter(pdev, sop, vdo); in typec_cable_altmode_enter()
322 const u32 header, const u32 *vdo, int count) in typec_cable_altmode_vdm() argument
345 return pdev->cable_ops->vdm(pdev, sop, header, vdo, count); in typec_cable_altmode_vdm()

123