/linux-6.12.1/net/ipv4/ |
D | bpf_tcp_ca.c | 88 case offsetof(struct tcp_sock, snd_cwnd): in bpf_tcp_ca_btf_struct_access() 89 end = offsetofend(struct tcp_sock, snd_cwnd); in bpf_tcp_ca_btf_struct_access() 91 case offsetof(struct tcp_sock, snd_cwnd_cnt): in bpf_tcp_ca_btf_struct_access() 92 end = offsetofend(struct tcp_sock, snd_cwnd_cnt); in bpf_tcp_ca_btf_struct_access() 94 case offsetof(struct tcp_sock, snd_cwnd_stamp): in bpf_tcp_ca_btf_struct_access() 95 end = offsetofend(struct tcp_sock, snd_cwnd_stamp); in bpf_tcp_ca_btf_struct_access() 97 case offsetof(struct tcp_sock, snd_ssthresh): in bpf_tcp_ca_btf_struct_access() 98 end = offsetofend(struct tcp_sock, snd_ssthresh); in bpf_tcp_ca_btf_struct_access() 100 case offsetof(struct tcp_sock, ecn_flags): in bpf_tcp_ca_btf_struct_access() 101 end = offsetofend(struct tcp_sock, ecn_flags); in bpf_tcp_ca_btf_struct_access() [all …]
|
D | tcp_recovery.c | 7 const struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_wnd() 32 s32 tcp_rack_skb_timeout(struct tcp_sock *tp, struct sk_buff *skb, u32 reo_wnd) in tcp_rack_skb_timeout() 60 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_detect_loss() 97 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_mark_lost() 118 void tcp_rack_advance(struct tcp_sock *tp, u8 sacked, u32 end_seq, in tcp_rack_advance() 151 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_timeout() 189 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_update_reo_wnd() 220 struct tcp_sock *tp = tcp_sk(sk); in tcp_newreno_mark_lost()
|
D | tcp.c | 404 static u64 tcp_compute_delivery_rate(const struct tcp_sock *tp) in tcp_compute_delivery_rate() 425 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_sock() 514 const struct tcp_sock *tp = tcp_sk(sk); in tcp_poll() 619 struct tcp_sock *tp = tcp_sk(sk); in tcp_ioctl() 664 void tcp_mark_push(struct tcp_sock *tp, struct sk_buff *skb) in tcp_mark_push() 670 static inline bool forced_push(const struct tcp_sock *tp) in forced_push() 677 struct tcp_sock *tp = tcp_sk(sk); in tcp_skb_entail() 692 static inline void tcp_mark_urg(struct tcp_sock *tp, int flags) in tcp_mark_urg() 721 struct tcp_sock *tp = tcp_sk(sk); in tcp_push() 913 struct tcp_sock *tp = tcp_sk(sk); in tcp_xmit_size_goal() [all …]
|
D | tcp_input.c | 334 static void tcp_ecn_queue_cwr(struct tcp_sock *tp) in tcp_ecn_queue_cwr() 354 static void tcp_ecn_withdraw_cwr(struct tcp_sock *tp) in tcp_ecn_withdraw_cwr() 361 struct tcp_sock *tp = tcp_sk(sk); in __tcp_ecn_check_ce() 397 static void tcp_ecn_rcv_synack(struct tcp_sock *tp, const struct tcphdr *th) in tcp_ecn_rcv_synack() 403 static void tcp_ecn_rcv_syn(struct tcp_sock *tp, const struct tcphdr *th) in tcp_ecn_rcv_syn() 409 static bool tcp_ecn_rcv_ecn_echo(const struct tcp_sock *tp, const struct tcphdr *th) in tcp_ecn_rcv_ecn_echo() 423 const struct tcp_sock *tp = tcp_sk(sk); in tcp_sndbuf_expand() 482 const struct tcp_sock *tp = tcp_sk(sk); in __tcp_grow_window() 519 struct tcp_sock *tp = tcp_sk(sk); in tcp_grow_window() 559 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_buffer_space() [all …]
|
D | tcp_dctcp.c | 81 static void dctcp_reset(const struct tcp_sock *tp, struct dctcp *ca) in dctcp_reset() 91 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_init() 121 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ssthresh() 129 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_update_alpha() 177 struct tcp_sock *tp = tcp_sk(sk); in dctcp_react_to_loss() 219 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_get_info() 246 struct tcp_sock *tp = tcp_sk(sk); in dctcp_cwnd_undo()
|
D | tcp_output.c | 55 void tcp_mstamp_refresh(struct tcp_sock *tp) in tcp_mstamp_refresh() 70 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_new_data_sent() 99 const struct tcp_sock *tp = tcp_sk(sk); in tcp_acceptable_seq() 125 struct tcp_sock *tp = tcp_sk(sk); in tcp_advertise_mss() 146 struct tcp_sock *tp = tcp_sk(sk); in tcp_cwnd_restart() 163 static void tcp_event_data_sent(struct tcp_sock *tp, in tcp_event_data_sent() 184 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_ack_sent() 262 struct tcp_sock *tp = tcp_sk(sk); in tcp_select_window() 322 const struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send_synack() 335 struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send_syn() [all …]
|
D | tcp_bbr.c | 268 struct tcp_sock *tp = tcp_sk(sk); in bbr_init_pacing_rate_from_rtt() 288 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_pacing_rate() 306 struct tcp_sock *tp = tcp_sk(sk); in bbr_tso_segs_goal() 323 struct tcp_sock *tp = tcp_sk(sk); in bbr_save_cwnd() 334 struct tcp_sock *tp = tcp_sk(sk); in bbr_cwnd_event() 439 struct tcp_sock *tp = tcp_sk(sk); in bbr_packets_in_net_at_edt() 483 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_cwnd_to_recover_or_restore() 522 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_cwnd() 557 struct tcp_sock *tp = tcp_sk(sk); in bbr_is_next_cycle_phase() 593 struct tcp_sock *tp = tcp_sk(sk); in bbr_advance_cycle_phase() [all …]
|
D | tcp_timer.c | 30 const struct tcp_sock *tp = tcp_sk(sk); in tcp_clamp_rto_to_user_timeout() 107 struct tcp_sock *tp = tcp_sk(sk); in tcp_out_of_resources() 217 struct tcp_sock *tp = tcp_sk(sk); in retransmits_timed_out() 244 struct tcp_sock *tp = tcp_sk(sk); in tcp_write_timeout() 310 struct tcp_sock *tp = tcp_sk(sk); in tcp_delack_timer_handler() 381 struct tcp_sock *tp = tcp_sk(sk); in tcp_probe_timer() 430 struct tcp_sock *tp = tcp_sk(sk); in tcp_update_rto_stats() 447 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_synack_timer() 485 const struct tcp_sock *tp = tcp_sk(sk); in tcp_rtx_probe0_timed_out() 522 struct tcp_sock *tp = tcp_sk(sk); in tcp_retransmit_timer() [all …]
|
D | tcp_cdg.c | 143 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_hystart_update() 244 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_backoff() 265 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cong_avoid() 302 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_acked() 331 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_ssthresh() 348 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cwnd_event() 376 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_init()
|
D | tcp_rate.c | 42 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_skb_sent() 83 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_skb_delivered() 120 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_gen() 196 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_check_app_limited()
|
D | tcp_highspeed.c | 102 struct tcp_sock *tp = tcp_sk(sk); in hstcp_init() 114 struct tcp_sock *tp = tcp_sk(sk); in hstcp_cong_avoid() 153 const struct tcp_sock *tp = tcp_sk(sk); in hstcp_ssthresh()
|
D | tcp_scalable.c | 20 struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_cong_avoid() 36 const struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_ssthresh()
|
D | tcp_westwood.c | 165 const struct tcp_sock *tp = tcp_sk(sk); in westwood_fast_bw() 182 const struct tcp_sock *tp = tcp_sk(sk); in westwood_acked_count() 219 const struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_bw_rttmin() 242 struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_event()
|
D | tcp_illinois.c | 59 struct tcp_sock *tp = tcp_sk(sk); in rtt_reset() 224 struct tcp_sock *tp = tcp_sk(sk); in update_params() 262 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_cong_avoid() 297 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_ssthresh()
|
D | tcp_yeah.c | 43 struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_init() 60 struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_cong_avoid() 189 const struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_ssthresh()
|
D | tcp_vegas.c | 73 const struct tcp_sock *tp = tcp_sk(sk); in vegas_enable() 160 static inline u32 tcp_vegas_ssthresh(struct tcp_sock *tp) in tcp_vegas_ssthresh() 167 struct tcp_sock *tp = tcp_sk(sk); in tcp_vegas_cong_avoid()
|
/linux-6.12.1/include/net/ |
D | tcp.h | 398 void tcp_clear_retrans(struct tcp_sock *tp); 644 void tcp_mark_push(struct tcp_sock *tp, struct sk_buff *skb); 704 static inline int tcp_bound_to_half_wnd(struct tcp_sock *tp, int pktsize) in tcp_bound_to_half_wnd() 748 static inline u32 __tcp_set_rto(const struct tcp_sock *tp) in __tcp_set_rto() 753 static inline void __tcp_fast_path_on(struct tcp_sock *tp, u32 snd_wnd) in __tcp_fast_path_on() 764 static inline void tcp_fast_path_on(struct tcp_sock *tp) in tcp_fast_path_on() 771 struct tcp_sock *tp = tcp_sk(sk); in tcp_fast_path_check() 804 static inline u32 tcp_min_rtt(const struct tcp_sock *tp) in tcp_min_rtt() 813 static inline u32 tcp_receive_window(const struct tcp_sock *tp) in tcp_receive_window() 868 static inline u32 tcp_time_stamp_ms(const struct tcp_sock *tp) in tcp_time_stamp_ms() [all …]
|
/linux-6.12.1/tools/testing/selftests/bpf/progs/ |
D | tcp_ca_write_sk_pacing.c | 13 static unsigned int tcp_left_out(const struct tcp_sock *tp) in tcp_left_out() 18 static unsigned int tcp_packets_in_flight(const struct tcp_sock *tp) in tcp_packets_in_flight() 38 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
|
D | bpf_dctcp.c | 59 static void dctcp_reset(const struct tcp_sock *tp, struct bpf_dctcp *ca) in dctcp_reset() 70 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 118 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 127 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 157 struct tcp_sock *tp = tcp_sk(sk); in dctcp_react_to_loss() 176 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ece_ack_cwr()
|
D | bpf_tracing_net.h | 153 static inline struct tcp_sock *tcp_sk(const struct sock *sk) in tcp_sk() 155 return (struct tcp_sock *)sk; in tcp_sk() 158 static inline bool tcp_in_slow_start(const struct tcp_sock *tp) in tcp_in_slow_start() 165 const struct tcp_sock *tp = tcp_sk(sk); in tcp_is_cwnd_limited()
|
D | bpf_cubic.c | 31 extern __u32 tcp_slow_start(struct tcp_sock *tp, __u32 acked) __ksym; 32 extern void tcp_cong_avoid_ai(struct tcp_sock *tp, __u32 w, __u32 acked) __ksym; 171 struct tcp_sock *tp = tcp_sk(sk); in bictcp_hystart_reset() 388 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 408 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 456 struct tcp_sock *tp = tcp_sk(sk); in hystart_update() 505 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
|
D | cgrp_ls_attach_cgroup.c | 27 struct tcp_sock *tcp_sk; in set_cookie() 56 struct tcp_sock *tcp_sk; in update_cookie_sockops()
|
D | bpf_cc_cubic.c | 49 const struct tcp_sock *tp = tcp_sk(sk); in tcp_update_pacing_rate() 79 struct tcp_sock *tp = tcp_sk(sk); in tcp_cwnd_reduction() 129 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
|
/linux-6.12.1/include/linux/ |
D | tcp.h | 192 struct tcp_sock { struct 528 #define tcp_sk(ptr) container_of_const(ptr, struct tcp_sock, inet_conn.icsk_inet.sk) 533 #define tcp_sk_rw(ptr) container_of(ptr, struct tcp_sock, inet_conn.icsk_inet.sk) 575 static inline void tcp_move_syn(struct tcp_sock *tp, in tcp_move_syn() 582 static inline void tcp_saved_syn_free(struct tcp_sock *tp) in tcp_saved_syn_free() 598 static inline u16 tcp_mss_clamp(const struct tcp_sock *tp, u16 mss) in tcp_mss_clamp()
|
/linux-6.12.1/net/mptcp/ |
D | fastopen.c | 14 struct tcp_sock *tp; in mptcp_fastopen_subflow_synack_set_params() 22 ssk = subflow->tcp_sock; in mptcp_fastopen_subflow_synack_set_params()
|