Lines Matching full:rm
70 int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm, in rds_tcp_xmit() argument
73 struct rds_conn_path *cp = rm->m_inc.i_conn_path; in rds_tcp_xmit()
86 rm->m_ack_seq = tc->t_last_sent_nxt + in rds_tcp_xmit()
88 be32_to_cpu(rm->m_inc.i_hdr.h_len) - 1; in rds_tcp_xmit()
90 set_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags); in rds_tcp_xmit()
91 tc->t_last_expected_una = rm->m_ack_seq + 1; in rds_tcp_xmit()
93 if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags)) in rds_tcp_xmit()
94 rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED; in rds_tcp_xmit()
96 rdsdebug("rm %p tcp nxt %u ack_seq %llu\n", in rds_tcp_xmit()
97 rm, rds_tcp_write_seq(tc), in rds_tcp_xmit()
98 (unsigned long long)rm->m_ack_seq); in rds_tcp_xmit()
106 (void *)&rm->m_inc.i_hdr + hdr_off, in rds_tcp_xmit()
107 sizeof(rm->m_inc.i_hdr) - hdr_off); in rds_tcp_xmit()
115 while (sg < rm->data.op_nents) { in rds_tcp_xmit()
117 if (sg + 1 < rm->data.op_nents) in rds_tcp_xmit()
120 bvec_set_page(&bvec, sg_page(&rm->data.op_sg[sg]), in rds_tcp_xmit()
121 rm->data.op_sg[sg].length - off, in rds_tcp_xmit()
122 rm->data.op_sg[sg].offset + off); in rds_tcp_xmit()
124 rm->data.op_sg[sg].length - off); in rds_tcp_xmit()
126 rdsdebug("tcp sendpage %p:%u:%u ret %d\n", (void *)sg_page(&rm->data.op_sg[sg]), in rds_tcp_xmit()
127 rm->data.op_sg[sg].offset + off, rm->data.op_sg[sg].length - off, in rds_tcp_xmit()
134 if (off == rm->data.op_sg[sg].length) { in rds_tcp_xmit()
166 * rm->m_ack_seq is set to the tcp sequence number that corresponds to the
168 * entire message has been received if rm->m_ack_seq is "before" the next
172 static int rds_tcp_is_acked(struct rds_message *rm, uint64_t ack) in rds_tcp_is_acked() argument
174 if (!test_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags)) in rds_tcp_is_acked()
176 return (__s32)((u32)rm->m_ack_seq - (u32)ack) < 0; in rds_tcp_is_acked()