/linux-6.12.1/tools/tracing/rtla/src/ |
D | osnoise_hist.c | 43 int bucket_size; member 63 int bucket_size; member 92 *osnoise_alloc_histogram(int nr_cpus, int entries, int bucket_size) in osnoise_alloc_histogram() argument 102 data->bucket_size = bucket_size; in osnoise_alloc_histogram() 139 bucket = duration / data->bucket_size; in osnoise_hist_update_multiple() 173 int bucket_size; in osnoise_init_trace_hist() local 180 bucket_size = params->output_divisor * params->bucket_size; in osnoise_init_trace_hist() 181 snprintf(buff, sizeof(buff), "duration.buckets=%d", bucket_size); in osnoise_init_trace_hist() 388 bucket * data->bucket_size); in osnoise_print_stats() 524 params->bucket_size = 1; in osnoise_hist_parse_args() [all …]
|
D | timerlat_hist.c | 54 int bucket_size; member 85 int bucket_size; member 121 *timerlat_alloc_histogram(int nr_cpus, int entries, int bucket_size) in timerlat_alloc_histogram() argument 131 data->bucket_size = bucket_size; in timerlat_alloc_histogram() 185 bucket = latency / data->bucket_size; in timerlat_hist_update() 576 bucket * data->bucket_size); in timerlat_print_stats() 737 params->bucket_size = 1; in timerlat_hist_parse_args() 820 params->bucket_size = get_llong_from_str(optarg); in timerlat_hist_parse_args() 821 if ((params->bucket_size == 0) || (params->bucket_size >= 1000000)) in timerlat_hist_parse_args() 1108 tool->data = timerlat_alloc_histogram(nr_cpus, params->entries, params->bucket_size); in timerlat_init_hist()
|
/linux-6.12.1/fs/bcachefs/ |
D | sb-members.c | 152 if (le16_to_cpu(m.bucket_size) < in validate_member() 155 i, le16_to_cpu(m.bucket_size), le16_to_cpu(sb->block_size)); in validate_member() 159 if (le16_to_cpu(m.bucket_size) < in validate_member() 162 i, le16_to_cpu(m.bucket_size), BCH_SB_BTREE_NODE_SIZE(sb)); in validate_member() 181 u64 bucket_size = le16_to_cpu(m.bucket_size); in member_to_text() local 182 u64 device_size = le64_to_cpu(m.nbuckets) * bucket_size; in member_to_text() 220 prt_units_u64(out, bucket_size << 9); in member_to_text()
|
D | journal_reclaim.c | 87 .next_entry = ca->mi.bucket_size, in journal_dev_space_available() 88 .total = ca->mi.bucket_size * ja->nr, in journal_dev_space_available() 107 if (unwritten > ca->mi.bucket_size) in journal_dev_space_available() 117 sectors = ca->mi.bucket_size; in journal_dev_space_available() 123 if (sectors < ca->mi.bucket_size && buckets) { in journal_dev_space_available() 125 sectors = ca->mi.bucket_size; in journal_dev_space_available() 130 .total = sectors + buckets * ca->mi.bucket_size, in journal_dev_space_available() 200 max_entry_size = min_t(unsigned, max_entry_size, ca->mi.bucket_size); in bch2_journal_space_available() 286 ca->mi.bucket_size, GFP_NOFS); in bch2_journal_do_discards()
|
D | alloc_background.h | 115 return d ? max(0, ca->mi.bucket_size - d) : 0; in bch2_bucket_sectors_fragmented() 122 return d ? max(0, ca->mi.bucket_size - d) : 0; in bch2_gc_bucket_sectors_fragmented() 183 ca->mi.bucket_size); in alloc_lru_idx_fragmentation() 185 return div_u64(d * (1ULL << 31), ca->mi.bucket_size); in alloc_lru_idx_fragmentation()
|
D | buckets.h | 17 return div_u64(s, ca->mi.bucket_size); in sector_to_bucket() 22 return ((sector_t) b) * ca->mi.bucket_size; in bucket_to_sector() 29 div_u64_rem(s, ca->mi.bucket_size, &remainder); in bucket_remainder() 35 return div_u64_rem(s, ca->mi.bucket_size, offset); in sector_to_bucket_and_offset()
|
D | sb-members_types.h | 9 u16 bucket_size; /* sectors */ member
|
D | bcachefs_ioctl.h | 306 __u32 bucket_size; member 326 __u32 bucket_size; member
|
D | btree_node_scan.c | 225 bucket_offset + btree_sectors(c) <= ca->mi.bucket_size; in read_btree_nodes_worker() 228 u64 cur_sector = bucket * ca->mi.bucket_size + bucket_offset; in read_btree_nodes_worker() 229 u64 end_sector = ca->mi.nbuckets * ca->mi.bucket_size; in read_btree_nodes_worker() 236 u64 sector = bucket * ca->mi.bucket_size + bucket_offset; in read_btree_nodes_worker()
|
D | sb-members_format.h | 47 __le16 bucket_size; /* sectors */ member
|
D | sb-members.h | 327 .bucket_size = le16_to_cpu(mi->bucket_size), in bch2_mi_to_cpu()
|
D | journal_io.c | 28 m->last_journal_bucket_offset = cpu_to_le32(ca->mi.bucket_size - ca->journal.sectors_free); in bch2_journal_pos_from_member_info_set() 42 if (offset <= ca->mi.bucket_size) in bch2_journal_pos_from_member_info_resume() 43 ca->journal.sectors_free = ca->mi.bucket_size - offset; in bch2_journal_pos_from_member_info_resume() 1001 end = offset + ca->mi.bucket_size; in journal_read_bucket() 1075 ja->sectors_free = ca->mi.bucket_size - in journal_read_bucket() 1453 ca->mi.bucket_size - in __journal_write_alloc() 1511 sectors <= ca->mi.bucket_size && in journal_write_alloc() 1515 ja->sectors_free = ca->mi.bucket_size; in journal_write_alloc()
|
D | chardev.c | 603 arg.bucket_size = ca->mi.bucket_size; in bch2_ioctl_dev_usage() 644 arg.bucket_size = ca->mi.bucket_size; in bch2_ioctl_dev_usage_v2()
|
D | move.c | 664 unsigned dirty_sectors, bucket_size; in bch2_evacuate_bucket() local 694 bucket_size = ca->mi.bucket_size; in bch2_evacuate_bucket() 804 trace_evacuate_bucket(c, &bucket, dirty_sectors, bucket_size, fragmentation, ret); in bch2_evacuate_bucket()
|
D | alloc_foreground.c | 262 ob->sectors_free = ca->mi.bucket_size; in __try_alloc_bucket() 424 bucket_to_sector(ca, bucket), ca->mi.bucket_size)) { in bch2_bucket_alloc_early() 508 bucket_to_sector(ca, bucket), ca->mi.bucket_size)) { in bch2_bucket_alloc_freelist() 1529 ca->mi.bucket_size - in bch2_ob_ptr() 1608 ca->mi.bucket_size - ob->sectors_free, ca->mi.bucket_size); in bch2_open_bucket_to_text()
|
/linux-6.12.1/tools/testing/selftests/kvm/ |
D | kvm_binary_stats_test.c | 128 TEST_ASSERT(pdesc->bucket_size, in stats_test() 133 TEST_ASSERT(!pdesc->bucket_size, in stats_test()
|
/linux-6.12.1/drivers/md/bcache/ |
D | alloc.c | 90 unsigned long next = c->nbuckets * c->cache->sb.bucket_size / 1024; in bch_rescale_priorities() 367 ca->sb.bucket_size, GFP_KERNEL); in bch_allocator_thread() 476 SET_GC_SECTORS_USED(b, ca->sb.bucket_size); in bch_bucket_alloc() 608 ret->sectors_free = c->cache->sb.bucket_size; in pick_data_bucket()
|
D | super.c | 66 unsigned int bucket_size = le16_to_cpu(s->bucket_size); in get_bucket_size() local 73 order = le16_to_cpu(s->bucket_size); in get_bucket_size() 81 bucket_size = 1 << order; in get_bucket_size() 83 bucket_size += in get_bucket_size() 88 return bucket_size; in get_bucket_size() 99 sb->bucket_size = get_bucket_size(sb, s); in read_super_common() 125 if (!is_power_of_2(sb->bucket_size)) in read_super_common() 129 if (sb->bucket_size < PAGE_SECTORS) in read_super_common() 134 sb->bucket_size * sb->nbuckets) in read_super_common() 157 if (sb->first_bucket * sb->bucket_size < 16) in read_super_common() [all …]
|
D | movinggc.c | 224 reserve_sectors = ca->sb.bucket_size * in bch_moving_gc() 232 GC_SECTORS_USED(b) == ca->sb.bucket_size || in bch_moving_gc()
|
D | bcache_ondisk.h | 192 __le16 bucket_size; /* sectors */ member 254 __u32 bucket_size; /* sectors */ member
|
D | bcache.h | 768 #define bucket_bytes(ca) ((ca)->sb.bucket_size << 9) 779 n = sb->bucket_size / PAGE_SECTORS; in meta_bucket_pages() 810 return s & (c->cache->sb.bucket_size - 1); in bucket_remainder()
|
D | sysfs.c | 66 read_attribute(bucket_size); 733 sysfs_hprint(bucket_size, bucket_bytes(c->cache)); in SHOW() 1038 sysfs_hprint(bucket_size, bucket_bytes(ca)); in SHOW() 1123 n * ca->sb.bucket_size / (ARRAY_SIZE(q) + 1)); in SHOW()
|
D | journal.c | 52 while (offset < ca->sb.bucket_size) { in journal_read_bucket() 53 reread: left = ca->sb.bucket_size - offset; in journal_read_bucket() 687 c->journal.blocks_free = ca->sb.bucket_size >> c->block_bits; in journal_reclaim()
|
D | extents.c | 58 if (KEY_SIZE(k) + r > c->cache->sb.bucket_size || in __ptr_invalid() 79 if (KEY_SIZE(k) + r > c->cache->sb.bucket_size) in bch_ptr_status()
|
/linux-6.12.1/drivers/net/ethernet/marvell/ |
D | mv643xx_eth.c | 1137 int bucket_size; in tx_set_rate() local 1147 bucket_size = (burst + 255) >> 8; in tx_set_rate() 1148 if (bucket_size > 65535) in tx_set_rate() 1149 bucket_size = 65535; in tx_set_rate() 1155 wrlp(mp, TX_BW_BURST, bucket_size); in tx_set_rate() 1160 wrlp(mp, TX_BW_BURST_MOVED, bucket_size); in tx_set_rate() 1169 int bucket_size; in txq_set_rate() local 1175 bucket_size = (burst + 255) >> 8; in txq_set_rate() 1176 if (bucket_size > 65535) in txq_set_rate() 1177 bucket_size = 65535; in txq_set_rate() [all …]
|