/linux-6.12.1/drivers/gpu/drm/i915/display/ |
D | intel_tc.c | 37 enum intel_display_power_domain (*cold_off_domain)(struct intel_tc_port *tc); 38 u32 (*hpd_live_status)(struct intel_tc_port *tc); 39 bool (*is_ready)(struct intel_tc_port *tc); 40 bool (*is_owned)(struct intel_tc_port *tc); 41 void (*get_hw_state)(struct intel_tc_port *tc); 42 bool (*connect)(struct intel_tc_port *tc, int required_lanes); 43 void (*disconnect)(struct intel_tc_port *tc); 44 void (*init)(struct intel_tc_port *tc); 70 static u32 tc_phy_hpd_live_status(struct intel_tc_port *tc); 71 static bool tc_phy_is_ready(struct intel_tc_port *tc); [all …]
|
D | icl_dsi_regs.h | 12 #define _MMIO_DSI(tc, dsi0, dsi1) _MMIO_TRANS((tc) - TRANSCODER_DSI_0, \ argument 195 #define DSI_TRANS_FUNC_CONF(tc) _MMIO_DSI(tc, \ argument 241 #define DSI_CMD_RXCTL(tc) _MMIO_DSI(tc, \ argument 256 #define DSI_CMD_TXCTL(tc) _MMIO_DSI(tc, \ argument 269 #define DSI_CMD_TXHDR(tc) _MMIO_DSI(tc, \ argument 285 #define DSI_CMD_TXPYLD(tc) _MMIO_DSI(tc, \ argument 291 #define DSI_LP_MSG(tc) _MMIO_DSI(tc, \ argument 302 #define DSI_HSTX_TO(tc) _MMIO_DSI(tc, \ argument 312 #define DSI_LPRX_HOST_TO(tc) _MMIO_DSI(tc, \ argument 322 #define DSI_PWAIT_TO(tc) _MMIO_DSI(tc, \ argument [all …]
|
/linux-6.12.1/drivers/ntb/test/ |
D | ntb_tool.c | 208 struct tool_ctx *tc; member 233 struct tool_ctx *tc; member 239 struct tool_ctx *tc; member 244 struct tool_ctx *tc; member 291 struct tool_ctx *tc = ctx; in tool_link_event() local 296 up = ntb_link_is_up(tc->ntb, &speed, &width); in tool_link_event() 298 dev_dbg(&tc->ntb->dev, "link is %s speed %d width %d\n", in tool_link_event() 301 wake_up(&tc->link_wq); in tool_link_event() 306 struct tool_ctx *tc = ctx; in tool_db_event() local 309 db_mask = ntb_db_vector_mask(tc->ntb, vec); in tool_db_event() [all …]
|
/linux-6.12.1/drivers/gpu/drm/bridge/ |
D | tc358767.c | 409 static inline int tc_poll_timeout(struct tc_data *tc, unsigned int addr, in tc_poll_timeout() argument 416 return regmap_read_poll_timeout(tc->regmap, addr, val, in tc_poll_timeout() 421 static int tc_aux_wait_busy(struct tc_data *tc) in tc_aux_wait_busy() argument 423 return tc_poll_timeout(tc, DP0_AUXSTATUS, AUX_BUSY, 0, 100, 100000); in tc_aux_wait_busy() 426 static int tc_aux_write_data(struct tc_data *tc, const void *data, in tc_aux_write_data() argument 434 ret = regmap_raw_write(tc->regmap, DP0_AUXWDATA(0), auxwdata, count); in tc_aux_write_data() 441 static int tc_aux_read_data(struct tc_data *tc, void *data, size_t size) in tc_aux_read_data() argument 446 ret = regmap_raw_read(tc->regmap, DP0_AUXRDATA(0), auxrdata, count); in tc_aux_read_data() 470 struct tc_data *tc = aux_to_tc(aux); in tc_aux_transfer() local 476 ret = tc_aux_wait_busy(tc); in tc_aux_transfer() [all …]
|
D | tc358775.c | 292 struct tc_data *tc = bridge_to_tc(bridge); in tc_bridge_pre_enable() local 293 struct device *dev = &tc->dsi->dev; in tc_bridge_pre_enable() 296 ret = regulator_enable(tc->vddio); in tc_bridge_pre_enable() 301 ret = regulator_enable(tc->vdd); in tc_bridge_pre_enable() 306 gpiod_set_value(tc->stby_gpio, 0); in tc_bridge_pre_enable() 309 gpiod_set_value(tc->reset_gpio, 0); in tc_bridge_pre_enable() 315 struct tc_data *tc = bridge_to_tc(bridge); in tc_bridge_post_disable() local 316 struct device *dev = &tc->dsi->dev; in tc_bridge_post_disable() 319 gpiod_set_value(tc->reset_gpio, 1); in tc_bridge_post_disable() 322 gpiod_set_value(tc->stby_gpio, 1); in tc_bridge_post_disable() [all …]
|
/linux-6.12.1/tools/perf/util/ |
D | tsc.c | 18 u64 perf_time_to_tsc(u64 ns, struct perf_tsc_conversion *tc) in perf_time_to_tsc() argument 22 t = ns - tc->time_zero; in perf_time_to_tsc() 23 quot = t / tc->time_mult; in perf_time_to_tsc() 24 rem = t % tc->time_mult; in perf_time_to_tsc() 25 return (quot << tc->time_shift) + in perf_time_to_tsc() 26 (rem << tc->time_shift) / tc->time_mult; in perf_time_to_tsc() 29 u64 tsc_to_perf_time(u64 cyc, struct perf_tsc_conversion *tc) in tsc_to_perf_time() argument 33 if (tc->cap_user_time_short) in tsc_to_perf_time() 34 cyc = tc->time_cycles + in tsc_to_perf_time() 35 ((cyc - tc->time_cycles) & tc->time_mask); in tsc_to_perf_time() [all …]
|
/linux-6.12.1/kernel/time/ |
D | timecounter.c | 8 void timecounter_init(struct timecounter *tc, in timecounter_init() argument 12 tc->cc = cc; in timecounter_init() 13 tc->cycle_last = cc->read(cc); in timecounter_init() 14 tc->nsec = start_tstamp; in timecounter_init() 15 tc->mask = (1ULL << cc->shift) - 1; in timecounter_init() 16 tc->frac = 0; in timecounter_init() 31 static u64 timecounter_read_delta(struct timecounter *tc) in timecounter_read_delta() argument 37 cycle_now = tc->cc->read(tc->cc); in timecounter_read_delta() 40 cycle_delta = (cycle_now - tc->cycle_last) & tc->cc->mask; in timecounter_read_delta() 43 ns_offset = cyclecounter_cyc2ns(tc->cc, cycle_delta, in timecounter_read_delta() [all …]
|
/linux-6.12.1/net/rds/ |
D | tcp.c | 91 u32 rds_tcp_write_seq(struct rds_tcp_connection *tc) in rds_tcp_write_seq() argument 94 return tcp_sk(tc->t_sock->sk)->write_seq; in rds_tcp_write_seq() 97 u32 rds_tcp_snd_una(struct rds_tcp_connection *tc) in rds_tcp_snd_una() argument 99 return tcp_sk(tc->t_sock->sk)->snd_una; in rds_tcp_snd_una() 103 struct rds_tcp_connection *tc) in rds_tcp_restore_callbacks() argument 105 rdsdebug("restoring sock %p callbacks from tc %p\n", sock, tc); in rds_tcp_restore_callbacks() 110 list_del_init(&tc->t_list_item); in rds_tcp_restore_callbacks() 114 if (!tc->t_cpath->cp_conn->c_isv6) in rds_tcp_restore_callbacks() 118 tc->t_sock = NULL; in rds_tcp_restore_callbacks() 120 sock->sk->sk_write_space = tc->t_orig_write_space; in rds_tcp_restore_callbacks() [all …]
|
D | tcp_recv.c | 160 struct rds_tcp_connection *tc = cp->cp_transport_data; in rds_tcp_data_recv() local 161 struct rds_tcp_incoming *tinc = tc->t_tinc; in rds_tcp_data_recv() 165 rdsdebug("tcp data tc %p skb %p offset %u len %zu\n", tc, skb, offset, in rds_tcp_data_recv() 180 tc->t_tinc = tinc; in rds_tcp_data_recv() 194 if (left && tc->t_tinc_hdr_rem) { in rds_tcp_data_recv() 195 to_copy = min(tc->t_tinc_hdr_rem, left); in rds_tcp_data_recv() 201 tc->t_tinc_hdr_rem, in rds_tcp_data_recv() 203 tc->t_tinc_hdr_rem -= to_copy; in rds_tcp_data_recv() 207 if (tc->t_tinc_hdr_rem == 0) { in rds_tcp_data_recv() 209 tc->t_tinc_data_rem = in rds_tcp_data_recv() [all …]
|
D | tcp_send.c | 43 struct rds_tcp_connection *tc = cp->cp_transport_data; in rds_tcp_xmit_path_prepare() local 45 tcp_sock_set_cork(tc->t_sock->sk, true); in rds_tcp_xmit_path_prepare() 50 struct rds_tcp_connection *tc = cp->cp_transport_data; in rds_tcp_xmit_path_complete() local 52 tcp_sock_set_cork(tc->t_sock->sk, false); in rds_tcp_xmit_path_complete() 74 struct rds_tcp_connection *tc = cp->cp_transport_data; in rds_tcp_xmit() local 85 tc->t_last_sent_nxt = rds_tcp_write_seq(tc); in rds_tcp_xmit() 86 rm->m_ack_seq = tc->t_last_sent_nxt + in rds_tcp_xmit() 91 tc->t_last_expected_una = rm->m_ack_seq + 1; in rds_tcp_xmit() 97 rm, rds_tcp_write_seq(tc), in rds_tcp_xmit() 103 set_bit(SOCK_NOSPACE, &tc->t_sock->sk->sk_socket->flags); in rds_tcp_xmit() [all …]
|
D | tcp_connect.c | 44 struct rds_tcp_connection *tc; in rds_tcp_state_change() local 52 tc = cp->cp_transport_data; in rds_tcp_state_change() 53 state_change = tc->t_orig_state_change; in rds_tcp_state_change() 55 rdsdebug("sock %p state_change to %d\n", tc->t_sock, sk->sk_state); in rds_tcp_state_change() 100 struct rds_tcp_connection *tc = cp->cp_transport_data; in rds_tcp_conn_path_connect() local 108 mutex_lock(&tc->t_conn_path_lock); in rds_tcp_conn_path_connect() 111 mutex_unlock(&tc->t_conn_path_lock); in rds_tcp_conn_path_connect() 189 mutex_unlock(&tc->t_conn_path_lock); in rds_tcp_conn_path_connect() 206 struct rds_tcp_connection *tc = cp->cp_transport_data; in rds_tcp_conn_path_shutdown() local 207 struct socket *sock = tc->t_sock; in rds_tcp_conn_path_shutdown() [all …]
|
/linux-6.12.1/tools/testing/selftests/drivers/net/mlxsw/ |
D | tc_restrictions.sh | 44 tc qdisc add dev $swp1 ingress_block 22 clsact 47 tc filter add block 22 protocol ip pref 1 handle 101 flower \ 51 tc qdisc add dev $swp2 ingress_block 22 clsact 54 tc qdisc del dev $swp2 clsact 56 tc qdisc add dev $swp2 egress_block 22 clsact 59 tc filter del block 22 protocol ip pref 1 handle 101 flower 61 tc qdisc add dev $swp2 egress_block 22 clsact 64 tc filter add block 22 protocol ip pref 1 handle 101 flower \ 68 tc qdisc del dev $swp1 clsact 70 tc qdisc add dev $swp1 egress_block 22 clsact [all …]
|
D | tc_sample.sh | 133 tc qdisc add dev $rp1 clsact 137 tc qdisc add dev $rp2 clsact 142 tc qdisc add dev $rp3 clsact 148 tc qdisc add dev $rp4 clsact 155 tc qdisc del dev $rp4 clsact 161 tc qdisc del dev $rp3 clsact 166 tc qdisc del dev $rp2 clsact 170 tc qdisc del dev $rp1 clsact 232 tc filter add dev $rp1 ingress protocol all pref 1 handle 101 matchall \ 250 tc filter del dev $rp1 ingress protocol all pref 1 handle 101 matchall [all …]
|
/linux-6.12.1/drivers/md/ |
D | dm-thin.c | 226 typedef void (*process_bio_fn)(struct thin_c *tc, struct bio *bio); 227 typedef void (*process_cell_fn)(struct thin_c *tc, struct dm_bio_prison_cell *cell); 382 struct thin_c *tc; member 388 static void begin_discard(struct discard_op *op, struct thin_c *tc, struct bio *parent) in begin_discard() argument 392 op->tc = tc; in begin_discard() 400 struct thin_c *tc = op->tc; in issue_discard() local 401 sector_t s = block_to_sectors(tc->pool, data_b); in issue_discard() 402 sector_t len = block_to_sectors(tc->pool, data_e - data_b); in issue_discard() 404 return __blkdev_issue_discard(tc->pool_dev->bdev, s, len, GFP_NOIO, &op->bio); in issue_discard() 587 struct thin_c *tc; member [all …]
|
/linux-6.12.1/arch/mips/kernel/ |
D | smp-mt.c | 46 static unsigned int __init smvp_vpe_init(unsigned int tc, unsigned int mvpconf0, in smvp_vpe_init() argument 49 if (tc >= smp_max_threads || in smvp_vpe_init() 50 (tc > ((mvpconf0 & MVPCONF0_PVPE) >> MVPCONF0_PVPE_SHIFT))) in smvp_vpe_init() 54 if (tc != 0) { in smvp_vpe_init() 64 set_cpu_possible(tc, true); in smvp_vpe_init() 65 set_cpu_present(tc, true); in smvp_vpe_init() 66 __cpu_number_map[tc] = ++ncpu; in smvp_vpe_init() 67 __cpu_logical_map[ncpu] = tc; in smvp_vpe_init() 73 if (tc != 0) in smvp_vpe_init() 76 cpu_set_vpe_id(&cpu_data[ncpu], tc); in smvp_vpe_init() [all …]
|
D | vpe-mt.c | 31 struct tc *t; in vpe_run() 46 if (list_empty(&v->tc)) { in vpe_run() 57 t = list_first_entry(&v->tc, struct tc, tc); in vpe_run() 146 void cleanup_tc(struct tc *tc) in cleanup_tc() argument 158 settc(tc->index); in cleanup_tc() 208 struct tc *t; in vpe_stop() 213 t = list_entry(v->tc.next, struct tc, tc); in vpe_stop() 229 struct tc *t; in vpe_free() 232 t = list_entry(v->tc.next, struct tc, tc); in vpe_free() 330 struct tc *t; in vpe_module_init() [all …]
|
/linux-6.12.1/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_dcb.c | 165 int tc; in ixgbe_dcb_unpack_pfc() local 167 for (*pfc_en = 0, tc = 0; tc < MAX_TRAFFIC_CLASS; tc++) { in ixgbe_dcb_unpack_pfc() 168 if (tc_config[tc].dcb_pfc != pfc_disabled) in ixgbe_dcb_unpack_pfc() 169 *pfc_en |= BIT(tc); in ixgbe_dcb_unpack_pfc() 177 int tc; in ixgbe_dcb_unpack_refill() local 179 for (tc = 0; tc < MAX_TRAFFIC_CLASS; tc++) in ixgbe_dcb_unpack_refill() 180 refill[tc] = tc_config[tc].path[direction].data_credits_refill; in ixgbe_dcb_unpack_refill() 186 int tc; in ixgbe_dcb_unpack_max() local 188 for (tc = 0; tc < MAX_TRAFFIC_CLASS; tc++) in ixgbe_dcb_unpack_max() 189 max[tc] = tc_config[tc].desc_credits_max; in ixgbe_dcb_unpack_max() [all …]
|
/linux-6.12.1/tools/testing/selftests/net/forwarding/ |
D | tc_flower.sh | 29 tc qdisc add dev $h2 clsact 34 tc qdisc del dev $h2 clsact 44 tc filter add dev $h2 ingress protocol ip pref 1 handle 101 flower \ 46 tc filter add dev $h2 ingress protocol ip pref 2 handle 102 flower \ 58 tc filter del dev $h2 ingress protocol ip pref 1 handle 101 flower 59 tc filter del dev $h2 ingress protocol ip pref 2 handle 102 flower 70 tc filter add dev $h2 ingress protocol ip pref 1 handle 101 flower \ 72 tc filter add dev $h2 ingress protocol ip pref 2 handle 102 flower \ 84 tc filter del dev $h2 ingress protocol ip pref 1 handle 101 flower 85 tc filter del dev $h2 ingress protocol ip pref 2 handle 102 flower [all …]
|
D | tc_actions.sh | 20 tc qdisc add dev $h1 clsact 25 tc qdisc del dev $h1 clsact 32 tc qdisc add dev $h2 clsact 37 tc qdisc del dev $h2 clsact 44 tc qdisc add dev $swp1 clsact 53 tc qdisc del dev $swp1 clsact 66 tc filter add dev $h2 ingress protocol ip pref 1 handle 101 flower \ 75 tc filter add dev $swp1 ingress protocol $protocol pref 1 handle 101 \ 85 tc filter del dev $swp1 ingress protocol $protocol pref 1 handle 101 \ 87 tc filter del dev $h2 ingress protocol ip pref 1 handle 101 flower [all …]
|
D | tc_chains.sh | 25 tc qdisc add dev $h2 clsact 30 tc qdisc del dev $h2 clsact 38 tc filter add dev $h2 ingress chain 1 protocol ip pref 1 handle 1101 \ 47 tc filter del dev $h2 ingress chain 1 protocol ip pref 1 handle 1101 \ 57 tc filter add dev $h2 ingress chain 1 protocol ip pref 1 handle 1101 \ 59 tc filter add dev $h2 ingress protocol ip pref 2 handle 102 flower \ 61 tc filter add dev $h2 ingress protocol ip pref 1 handle 101 flower \ 76 tc filter del dev $h2 ingress protocol ip pref 1 handle 101 flower 77 tc filter del dev $h2 ingress protocol ip pref 2 handle 102 flower 78 tc filter del dev $h2 ingress chain 1 protocol ip pref 1 handle 1101 \ [all …]
|
/linux-6.12.1/tools/testing/selftests/drivers/net/mlxsw/spectrum-2/ |
D | tc_flower.sh | 35 tc qdisc add dev $h2 clsact 40 tc qdisc del dev $h2 clsact 101 tc filter add dev $h2 ingress protocol ip pref 1 handle 101 flower \ 110 tc filter add dev $h2 ingress protocol ip pref 2 handle 102 flower \ 125 tc filter del dev $h2 ingress protocol ip pref 1 handle 101 flower 133 tc filter del dev $h2 ingress protocol ip pref 2 handle 102 flower 147 tc filter add dev $h2 ingress protocol ip pref 1 handle 101 flower \ 149 tc filter add dev $h2 ingress protocol ip pref 2 handle 102 flower \ 158 tc filter del dev $h2 ingress protocol ip pref 1 handle 101 flower 166 tc filter add dev $h2 ingress protocol ip pref 3 handle 103 flower \ [all …]
|
/linux-6.12.1/drivers/clocksource/ |
D | timer-atmel-tcb.c | 256 static int __init setup_clkevents(struct atmel_tc *tc, int divisor_idx) in setup_clkevents() argument 259 struct clk *t2_clk = tc->clk[2]; in setup_clkevents() 260 int irq = tc->irq[2]; in setup_clkevents() 261 int bits = tc->tcb_config->counter_width; in setup_clkevents() 268 clkevt.regs = tc->regs; in setup_clkevents() 275 ret = clk_prepare_enable(tc->slow_clk); in setup_clkevents() 281 clkevt.rate = clk_get_rate(tc->slow_clk); in setup_clkevents() 293 clk_disable_unprepare(tc->slow_clk); in setup_clkevents() 304 static int __init setup_clkevents(struct atmel_tc *tc, int divisor_idx) in setup_clkevents() argument 312 static void __init tcb_setup_dual_chan(struct atmel_tc *tc, int mck_divisor_idx) in tcb_setup_dual_chan() argument [all …]
|
/linux-6.12.1/drivers/hwtracing/coresight/ |
D | coresight-cti-core.c | 250 struct cti_trig_con *tc, in cti_add_connection_entry() argument 256 tc->con_dev = csdev; in cti_add_connection_entry() 262 tc->con_dev_name = dev_name(&csdev->dev); in cti_add_connection_entry() 264 tc->con_dev_name = devm_kstrdup(dev, in cti_add_connection_entry() 266 if (!tc->con_dev_name) in cti_add_connection_entry() 269 list_add_tail(&tc->node, &cti_dev->trig_cons); in cti_add_connection_entry() 273 drvdata->config.trig_in_use |= tc->con_in->used_mask; in cti_add_connection_entry() 274 drvdata->config.trig_out_use |= tc->con_out->used_mask; in cti_add_connection_entry() 283 struct cti_trig_con *tc = NULL; in cti_allocate_trig_con() local 286 tc = devm_kzalloc(dev, sizeof(struct cti_trig_con), GFP_KERNEL); in cti_allocate_trig_con() [all …]
|
/linux-6.12.1/drivers/net/ethernet/sfc/ |
D | tc.c | 133 old = rhashtable_lookup_get_insert_fast(&efx->tc->mac_ht, in efx_tc_flower_get_mac() 157 rhashtable_remove_fast(&efx->tc->mac_ht, &ped->linkage, in efx_tc_flower_get_mac() 168 rhashtable_remove_fast(&efx->tc->mac_ht, &ped->linkage, in efx_tc_flower_put_mac() 482 rhashtable_remove_fast(&efx->tc->encap_match_ht, &encap->linkage, in efx_tc_flower_release_encap_match() 595 old = rhashtable_lookup_get_insert_fast(&efx->tc->encap_match_ht, in efx_tc_flower_record_encap_match() 681 rhashtable_remove_fast(&efx->tc->encap_match_ht, &encap->linkage, in efx_tc_flower_record_encap_match() 706 old = rhashtable_lookup_get_insert_fast(&efx->tc->recirc_ht, in efx_tc_get_recirc_id() 719 rc = ida_alloc_range(&efx->tc->recirc_ida, 1, U8_MAX, GFP_USER); in efx_tc_get_recirc_id() 721 rhashtable_remove_fast(&efx->tc->recirc_ht, in efx_tc_get_recirc_id() 737 rhashtable_remove_fast(&efx->tc->recirc_ht, &rid->linkage, in efx_tc_put_recirc_id() [all …]
|
/linux-6.12.1/drivers/net/ethernet/aquantia/atlantic/hw_atl2/ |
D | hw_atl2.c | 116 unsigned int tc, q; in hw_atl2_hw_queue_to_tc_map_set() local 135 for (tc = 0; tc != tcs; tc++) { in hw_atl2_hw_queue_to_tc_map_set() 136 unsigned int tc_q_offset = tc * q_per_tc; in hw_atl2_hw_queue_to_tc_map_set() 139 rx_map |= tc << HW_ATL2_RX_Q_TC_MAP_SHIFT(q); in hw_atl2_hw_queue_to_tc_map_set() 148 tx_map |= tc << HW_ATL2_TX_Q_TC_MAP_SHIFT(q); in hw_atl2_hw_queue_to_tc_map_set() 168 u32 tc = 0U; in hw_atl2_hw_qos_set() local 179 for (tc = 0; tc < cfg->tcs; tc++) { in hw_atl2_hw_qos_set() 183 hw_atl_tpb_tx_pkt_buff_size_per_tc_set(self, tx_buff_size, tc); in hw_atl2_hw_qos_set() 186 hw_atl_tpb_tx_buff_hi_threshold_per_tc_set(self, threshold, tc); in hw_atl2_hw_qos_set() 189 hw_atl_tpb_tx_buff_lo_threshold_per_tc_set(self, threshold, tc); in hw_atl2_hw_qos_set() [all …]
|