19 typedef enum _tcp_output_next
28 #define foreach_tcp4_output_next \ 29 _ (DROP, "error-drop") \ 30 _ (IP_LOOKUP, "ip4-lookup") \ 31 _ (IP_REWRITE, "ip4-rewrite") \ 34 #define foreach_tcp6_output_next \ 35 _ (DROP, "error-drop") \ 36 _ (IP_LOOKUP, "ip6-lookup") \ 37 _ (IP_REWRITE, "ip6-rewrite") \ 38 _ (IP_ARP, "ip6-discover-neighbor") 41 #define tcp_error(n,s) s, 68 #ifndef CLIB_MARCH_VARIANT 125 if (tc->state != TCP_STATE_SYN_RCVD ||
tcp_opts_wscale (&tc->rcv_opts))
136 u32 available_space, wnd;
152 observed_wnd = (
i32) tc->rcv_wnd - (tc->rcv_nxt - tc->rcv_las);
158 TCP_EVT_DBG (TCP_EVT_RCV_WND_SHRUNK, tc, observed_wnd, available_space);
162 wnd = available_space;
166 if (wnd && tc->rcv_wscale)
168 wnd &= ~((1 << tc->rcv_wscale) - 1);
170 wnd = 1 << tc->rcv_wscale;
182 if (state < TCP_STATE_ESTABLISHED)
186 return tc->rcv_wnd >> tc->rcv_wscale;
196 u32 buf, seq_len = 4;
202 buf = clib_host_to_net_u16 (opts->
mss);
204 data +=
sizeof (opts->
mss);
227 buf = clib_host_to_net_u32 (opts->
tsval);
229 data +=
sizeof (opts->
tsval);
230 buf = clib_host_to_net_u32 (opts->
tsecr);
232 data +=
sizeof (opts->
tsecr);
246 buf = clib_host_to_net_u32 (opts->
sacks[i].start);
249 buf = clib_host_to_net_u32 (opts->
sacks[i].end);
278 opts->
flags |= TCP_OPTS_FLAG_MSS;
282 opts->
flags |= TCP_OPTS_FLAG_WSCALE;
286 opts->
flags |= TCP_OPTS_FLAG_TSTAMP;
293 opts->
flags |= TCP_OPTS_FLAG_SACK_PERMITTED;
307 opts->
flags |= TCP_OPTS_FLAG_MSS;
313 opts->
flags |= TCP_OPTS_FLAG_WSCALE;
314 opts->
wscale = tc->rcv_wscale;
320 opts->
flags |= TCP_OPTS_FLAG_TSTAMP;
322 opts->
tsecr = tc->tsval_recent;
328 opts->
flags |= TCP_OPTS_FLAG_SACK_PERMITTED;
346 opts->
flags |= TCP_OPTS_FLAG_TSTAMP;
348 opts->
tsecr = tc->tsval_recent;
355 opts->
flags |= TCP_OPTS_FLAG_SACK;
356 if (tc->snd_sack_pos >=
vec_len (tc->snd_sacks))
357 tc->snd_sack_pos = 0;
358 opts->
sacks = &tc->snd_sacks[tc->snd_sack_pos];
378 case TCP_STATE_ESTABLISHED:
379 case TCP_STATE_CLOSE_WAIT:
380 case TCP_STATE_FIN_WAIT_1:
381 case TCP_STATE_LAST_ACK:
382 case TCP_STATE_CLOSING:
383 case TCP_STATE_FIN_WAIT_2:
384 case TCP_STATE_TIME_WAIT:
385 case TCP_STATE_CLOSED:
387 case TCP_STATE_SYN_RCVD:
389 case TCP_STATE_SYN_SENT:
415 TCP_STATE_ESTABLISHED);
418 tc->snd_mss =
clib_min (tc->mss, tc->rcv_opts.mss) - tc->snd_opts_len;
430 u16 default_min_mss = 536;
434 tc->snd_mss =
clib_min (tc->rcv_opts.mss, tc->mss);
436 if (tc->snd_mss < 45)
439 tc->snd_mss = default_min_mss;
440 tc->rcv_opts.mss = default_min_mss;
444 ASSERT (tc->snd_mss > 45);
455 if (b->
flags & VLIB_BUFFER_NEXT_PRESENT)
458 b->
flags &= VLIB_BUFFER_NEXT_PRESENT - 1;
468 #ifndef CLIB_MARCH_VARIANT 472 ASSERT ((b->
flags & VLIB_BUFFER_NEXT_PRESENT) == 0);
473 b->
flags |= VNET_BUFFER_F_LOCALLY_ORIGINATED;
490 u8 tcp_opts_len, tcp_hdr_opts_len;
498 tcp_hdr_opts_len = tcp_opts_len +
sizeof (
tcp_header_t);
501 tc->rcv_nxt, tcp_hdr_opts_len, flags, wnd);
504 vnet_buffer (b)->tcp.connection_index = tc->c_c_index;
515 tc->rcv_las = tc->rcv_nxt;
533 u8 tcp_hdr_opts_len, tcp_opts_len;
543 tcp_hdr_opts_len = tcp_opts_len +
sizeof (
tcp_header_t);
548 vnet_buffer (b)->tcp.connection_index = tc->c_c_index;
559 u8 tcp_opts_len, tcp_hdr_opts_len;
566 tcp_hdr_opts_len = tcp_opts_len +
sizeof (
tcp_header_t);
569 tc->rcv_nxt, tcp_hdr_opts_len,
573 vnet_buffer (b)->tcp.connection_index = tc->c_c_index;
578 u8 is_ip4,
u32 fib_index,
u8 flush)
581 u32 *to_next, next_index;
584 b->
flags |= VNET_BUFFER_F_LOCALLY_ORIGINATED;
621 u8 is_ip4,
u32 fib_index)
632 u32 *to_next, next_index;
635 b->
flags |= VNET_BUFFER_F_LOCALLY_ORIGINATED;
708 src_port = th0->src_port;
709 dst_port = th0->dst_port;
712 if (state == TCP_STATE_CLOSED)
717 tmp = clib_net_to_host_u32 (th0->seq_number);
721 ack = clib_host_to_net_u32 (tmp + 1);
727 seq = th0->ack_number;
754 #ifndef CLIB_MARCH_VARIANT 762 u32 thread_index,
u8 is_ip4)
801 seq = pkt_th->ack_number;
802 ack = (tc->state >= TCP_STATE_SYN_RCVD) ? tc->rcv_nxt : 0;
808 ack = clib_host_to_net_u32 (
vnet_buffer (pkt)->tcp.seq_end);
812 seq, ack, tcp_hdr_len, flags, 0);
848 u16 tcp_hdr_opts_len, advertise_wnd, opts_write_len;
857 tcp_hdr_opts_len = tc->snd_opts_len +
sizeof (
tcp_header_t);
861 tc->rcv_nxt, tcp_hdr_opts_len, flags,
864 ASSERT (opts_write_len == tc->snd_opts_len);
865 vnet_buffer (b)->tcp.connection_index = tc->c_c_index;
870 &tc->c_rmt_ip.ip4, IP_PROTOCOL_TCP, 0);
878 &tc->c_rmt_ip.ip6, IP_PROTOCOL_TCP);
896 &tc->c_rmt_ip4, IP_PROTOCOL_TCP, 1);
905 &tc->c_rmt_ip6, IP_PROTOCOL_TCP);
946 tc->rtt_seq = tc->snd_nxt;
1033 fin_snt = tc->
flags & TCP_CONN_FINSNT;
1045 tc->flags |= TCP_CONN_FINSNT;
1059 tc->flags |= TCP_CONN_FINSNT;
1060 tc->flags &= ~TCP_CONN_FINPNDG;
1061 tc->snd_una_max =
seq_max (tc->snd_una_max, tc->snd_nxt);
1071 u8 compute_opts,
u8 maybe_burst,
u8 update_snd_nxt)
1074 u32 advertise_wnd, data_len;
1083 vnet_buffer (b)->tcp.connection_index = tc->c_c_index;
1088 tcp_hdr_opts_len = tc->snd_opts_len +
sizeof (
tcp_header_t);
1091 advertise_wnd = tc->rcv_wnd >> tc->rcv_wscale;
1097 if (
seq_geq (tc->psh_seq, snd_nxt)
1098 &&
seq_lt (tc->psh_seq, snd_nxt + data_len))
1099 flags |= TCP_FLAG_PSH;
1102 tc->rcv_nxt, tcp_hdr_opts_len, flags,
1108 tm->wrk_ctx[tc->c_thread_index].cached_opts,
1114 ASSERT (len == tc->snd_opts_len);
1122 tc->snd_nxt += data_len;
1123 tc->rcv_las = tc->rcv_nxt;
1134 tc->snd_una_max =
seq_max (tc->snd_nxt, tc->snd_una_max);
1140 tc->rtt_seq = tc->snd_nxt;
1173 if (!(tc->flags & TCP_CONN_SNDACK))
1176 tc->flags |= TCP_CONN_SNDACK;
1183 if (!(tc->flags & TCP_CONN_SNDACK))
1186 tc->flags |= TCP_CONN_SNDACK;
1188 if (tc->pending_dupacks < 255)
1189 tc->pending_dupacks += 1;
1195 u32 thread_index, *pending_acks;
1204 for (i = 0; i <
vec_len (pending_acks); i++)
1207 tc->flags &= ~TCP_CONN_SNDACK;
1208 if (!tc->pending_dupacks)
1216 if (!vec_len (tc->snd_sacks))
1223 tc->snd_sack_pos = 0;
1229 n_acks =
clib_min (n_acks, tc->pending_dupacks);
1231 for (j = 0; j < n_acks; j++)
1234 tc->pending_dupacks = 0;
1235 tc->snd_sack_pos = 0;
1298 ASSERT (n_bytes == max_deq_bytes);
1306 u32 chain_bi = ~0, n_bufs_per_seg, n_bufs;
1307 u16 n_peeked, len_to_deq;
1312 n_bufs_per_seg = ceil ((
double) seg_size / bytes_per_buffer);
1327 TRANSPORT_MAX_HDRS_LEN);
1329 b[0]->
flags |= VLIB_BUFFER_TOTAL_LENGTH_VALID;
1331 max_deq_bytes -= n_bytes;
1334 for (i = 1; i < n_bufs_per_seg; i++)
1337 len_to_deq =
clib_min (max_deq_bytes, bytes_per_buffer);
1345 ASSERT (n_peeked == len_to_deq);
1346 n_bytes += n_peeked;
1352 prev_b->
flags |= VLIB_BUFFER_NEXT_PRESENT;
1354 max_deq_bytes -= n_peeked;
1369 ASSERT (((*b)->current_data + (*b)->current_length) <= bytes_per_buffer);
1385 u32 start, available_bytes;
1388 ASSERT (tc->state >= TCP_STATE_ESTABLISHED);
1389 ASSERT (max_deq_bytes != 0);
1395 ASSERT (available_bytes >= offset);
1396 available_bytes -=
offset;
1397 if (!available_bytes)
1400 max_deq_bytes =
clib_min (tc->snd_mss, max_deq_bytes);
1401 max_deq_bytes =
clib_min (available_bytes, max_deq_bytes);
1404 start = tc->snd_una +
offset;
1405 if (
seq_geq (start, tc->snd_congestion))
1409 if (
seq_gt (start + max_deq_bytes, tc->snd_congestion))
1411 max_deq_bytes = tc->snd_congestion - start;
1412 if (max_deq_bytes == 0)
1421 tc->snd_rxt_bytes += n_bytes;
1424 TCP_EVT_DBG (TCP_EVT_CC_RTX, tc, offset, n_bytes);
1435 tc->prev_ssthresh = tc->ssthresh;
1436 tc->prev_cwnd = tc->cwnd;
1446 tc->rcv_dupacks = 0;
1449 tc->cc_algo->congestion (tc);
1451 tc->snd_congestion = tc->snd_nxt;
1453 tc->cwnd_acc_bytes = 0;
1472 if (
PREDICT_FALSE (tc == 0 || tc->state != TCP_STATE_SYN_SENT))
1480 if (
PREDICT_FALSE (tc == 0 || tc->state == TCP_STATE_SYN_SENT))
1484 if (tc->state == TCP_STATE_CLOSED)
1488 if (tc->state >= TCP_STATE_ESTABLISHED)
1493 if (tc->flags & TCP_CONN_FINSNT)
1503 if ((tc->rto_boff == 0 && tc->snd_una == tc->snd_nxt)
1504 || (tc->rto_boff > 0 &&
seq_geq (tc->snd_una, tc->snd_congestion)
1525 if (tc->rto_boff == 1)
1531 tc->snd_congestion =
seq_max (tc->snd_nxt, tc->snd_congestion);
1533 tc->snd_nxt = tc->snd_una;
1548 if (tc->rto_boff == 1)
1555 else if (tc->state == TCP_STATE_SYN_SENT)
1560 if (tc->flags & TCP_CONN_HALF_OPEN_DONE)
1563 TCP_DBG (
"could not remove half-open connection");
1596 else if (tc->state == TCP_STATE_SYN_RCVD)
1623 ASSERT (tc->state == TCP_STATE_CLOSED);
1649 u32 bi, max_snd_bytes, available_bytes,
offset;
1665 if (tc->state == TCP_STATE_CLOSED || tc->snd_wnd > tc->snd_mss
1666 || (tc->flags & TCP_CONN_FINSNT))
1670 offset = tc->snd_nxt - tc->snd_una;
1674 if (!available_bytes)
1680 if (available_bytes <= offset)
1703 max_snd_bytes =
clib_min (tc->snd_mss,
1709 || tc->snd_nxt == tc->snd_una_max
1710 || tc->rto_boff > 1));
1714 tc->snd_una_max =
seq_max (tc->snd_nxt, tc->snd_una_max);
1752 offset = tc->snd_nxt - tc->snd_una;
1753 while (n_segs < burst_size)
1761 offset += n_written;
1764 tc->snd_nxt += n_written;
1765 tc->snd_una_max =
seq_max (tc->snd_nxt, tc->snd_una_max);
1772 #define scoreboard_rescue_rxt_valid(_sb, _tc) \ 1773 (seq_geq (_sb->rescue_rxt, _tc->snd_una) \ 1774 && seq_leq (_sb->rescue_rxt, _tc->snd_congestion)) 1783 u32 n_written = 0,
offset, max_bytes, n_segs = 0, n_segs_now;
1790 u8 snd_limited = 0, can_rescue = 0;
1795 if (snd_space < tc->snd_mss)
1806 max_deq -= tc->snd_nxt - tc->snd_una;
1808 while (snd_space > 0 && n_segs < burst_size)
1816 snd_space =
clib_min (max_deq, snd_space);
1817 burst_size =
clib_min (burst_size - n_segs,
1818 snd_space / tc->snd_mss);
1820 if (max_deq > n_segs_now * tc->snd_mss)
1822 n_segs += n_segs_now;
1835 tc->snd_congestion - tc->snd_una);
1836 max_bytes =
clib_min (max_bytes, snd_space);
1837 offset = tc->snd_congestion - tc->snd_una - max_bytes;
1838 sb->rescue_rxt = tc->snd_congestion;
1850 max_bytes =
clib_min (hole->end - sb->high_rxt, snd_space);
1851 max_bytes = snd_limited ?
clib_min (max_bytes, tc->snd_mss) : max_bytes;
1855 offset = sb->high_rxt - tc->snd_una;
1858 ASSERT (n_written <= snd_space);
1867 sb->high_rxt += n_written;
1868 snd_space -= n_written;
1886 u32 n_written = 0,
offset = 0, bi, max_deq, n_segs_now;
1888 int snd_space, n_segs = 0;
1899 snd_space = tc->sack_sb.last_bytes_delivered;
1900 while (snd_space > 0 && n_segs < burst_size)
1911 snd_space -= n_written;
1916 if (n_segs == burst_size)
1923 if (snd_space < tc->snd_mss || tc->snd_mss == 0)
1927 max_deq -= tc->snd_nxt - tc->snd_una;
1930 snd_space =
clib_min (max_deq, snd_space);
1931 burst_size =
clib_min (burst_size - n_segs, snd_space / tc->snd_mss);
1933 if (max_deq > n_segs_now * tc->snd_mss)
1935 n_segs += n_segs_now;
1959 u16 * next0,
u32 * error0)
1972 *error0 = TCP_ERROR_LINK_LOCAL_RW;
1984 *error0 = TCP_ERROR_LINK_LOCAL_RW;
1991 u32 * to_next,
u32 n_bufs)
2000 for (i = 0; i <
clib_min (n_trace, n_bufs); i++)
2023 IP_PROTOCOL_TCP, 1);
2024 b0->
flags |= VNET_BUFFER_F_OFFLOAD_TCP_CKSUM;
2032 &tc0->c_rmt_ip6, IP_PROTOCOL_TCP);
2033 b0->
flags |= VNET_BUFFER_F_OFFLOAD_TCP_CKSUM;
2042 u32 * error0,
u16 * next0,
u8 is_ip4)
2047 *error0 = TCP_ERROR_INVALID_CONNECTION;
2084 while (n_left_from >= 4)
2086 u32 error0 = TCP_ERROR_PKTS_SENT, error1 = TCP_ERROR_PKTS_SENT;
2114 while (n_left_from > 0)
2116 u32 error0 = TCP_ERROR_PKTS_SENT;
2119 if (n_left_from > 1)
2156 .name =
"tcp4-output",
2158 .vector_size =
sizeof (
u32),
2164 #define _(s,n) [TCP_OUTPUT_NEXT_##s] = n, 2176 .name =
"tcp6-output",
2178 .vector_size =
sizeof (
u32),
2184 #define _(s,n) [TCP_OUTPUT_NEXT_##s] = n, 2193 typedef enum _tcp_reset_next
2200 #define foreach_tcp4_reset_next \ 2201 _(DROP, "error-drop") \ 2202 _(IP_LOOKUP, "ip4-lookup") 2204 #define foreach_tcp6_reset_next \ 2205 _(DROP, "error-drop") \ 2206 _(IP_LOOKUP, "ip6-lookup") 2212 u32 n_left_from, next_index, *from, *to_next;
2220 while (n_left_from > 0)
2226 while (n_left_from > 0 && n_left_to_next > 0)
2239 n_left_to_next -= 1;
2244 my_thread_index, is_ip4))
2246 error0 = TCP_ERROR_LOOKUP_DROPS;
2257 b0->
flags |= VNET_BUFFER_F_LOCALLY_ORIGINATED;
2271 n_left_to_next, bi0, next0);
2292 .name =
"tcp4-reset",
2293 .vector_size =
sizeof (
u32),
2298 #define _(s,n) [TCP_RESET_NEXT_##s] = n, 2308 .name =
"tcp6-reset",
2309 .vector_size =
sizeof (
u32),
2314 #define _(s,n) [TCP_RESET_NEXT_##s] = n, void tcp_make_fin(tcp_connection_t *tc, vlib_buffer_t *b)
Convert buffer to FIN-ACK.
#define tcp_in_cong_recovery(tc)
u32 flags
buffer flags: VLIB_BUFFER_FREE_LIST_INDEX_MASK: bits used to store free list index, VLIB_BUFFER_IS_TRACED: trace this buffer.
void session_flush_frames_main_thread(vlib_main_t *vm)
static u32 tcp_options_write(u8 *data, tcp_options_t *opts)
Write TCP options to segment.
static void tcp_rxt_timeout_cc(tcp_connection_t *tc)
Reset congestion control, switch cwnd to loss window and try again.
#define TCP_OPTION_LEN_EOL
u32 * pending_acks
vector of pending acks
#define tcp_in_recovery(tc)
void scoreboard_clear(sack_scoreboard_t *sb)
static f64 tcp_time_now_us(u32 thread_index)
static void tcp_retransmit_timer_set(tcp_connection_t *tc)
static u32 transport_rx_fifo_size(transport_connection_t *tc)
#define TCP_OPTION_LEN_SACK_PERMITTED
static u32 vlib_get_trace_count(vlib_main_t *vm, vlib_node_runtime_t *rt)
void tcp_timer_retransmit_handler(u32 index)
static void vlib_buffer_free(vlib_main_t *vm, u32 *buffers, u32 n_buffers)
Free buffers Frees the entire buffer chain for each buffer.
int session_tx_fifo_peek_bytes(transport_connection_t *tc, u8 *buffer, u32 offset, u32 max_bytes)
#define TCP_TO_TIMER_TICK
#define TCP_MIN_RX_FIFO_SIZE
#define tcp_opts_tstamp(_to)
void tcp_make_synack(tcp_connection_t *tc, vlib_buffer_t *b)
Convert buffer to SYN-ACK.
i16 current_data
signed offset in data[], pre_data[] that we are currently processing.
static void tcp_flush_frame_to_ip_lookup(tcp_worker_ctx_t *wrk, u8 is_ip4)
Flush ip lookup tx frames populated by timer pops.
static tcp_connection_t * tcp_connection_get_if_valid(u32 conn_index, u32 thread_index)
static int tcp_make_syn_options(tcp_options_t *opts, u8 wnd_scale)
#define clib_memcpy_fast(a, b, c)
clib_memset(h->entries, 0, sizeof(h->entries[0]) *entries)
struct _sack_scoreboard sack_scoreboard_t
u32 fib_table_get_index_for_sw_if_index(fib_protocol_t proto, u32 sw_if_index)
Get the index of the FIB bound to the interface.
static tcp_connection_t * tcp_half_open_connection_get(u32 conn_index)
void tcp_update_rcv_mss(tcp_connection_t *tc)
Update max segment size we're able to process.
vlib_frame_t * tx_frames[2]
tx frames for tcp 4/6 output nodes
void tcp_send_acks(tcp_worker_ctx_t *wrk)
struct _tcp_main tcp_main_t
This packet is to be rewritten and forwarded to the next processing node.
u16 current_length
Nbytes between current data and the end of this buffer.
#define vec_add1(V, E)
Add 1 element to end of vector (unspecified alignment).
static u32 tcp_initial_wnd_unscaled(tcp_connection_t *tc)
TCP's initial window.
enum _tcp_output_next tcp_output_next_t
struct _tcp_connection tcp_connection_t
static u32 tcp_available_cc_snd_space(const tcp_connection_t *tc)
Estimate of how many bytes we can still push into the network.
#define tcp_opts_sack(_to)
#define VLIB_NODE_FN(node)
static void tcp_push_ip_hdr(tcp_worker_ctx_t *wrk, tcp_connection_t *tc, vlib_buffer_t *b)
#define vec_validate_aligned(V, I, A)
Make sure vector is long enough for given index (no header, specified alignment)
static uword tcp46_send_reset_inline(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *from_frame, u8 is_ip4)
vlib_error_t * errors
Vector of errors for this node.
int tcp_fast_retransmit_no_sack(tcp_worker_ctx_t *wrk, tcp_connection_t *tc, u32 burst_size)
Fast retransmit without SACK info.
u8 n_sack_blocks
Number of SACKs blocks.
static void tcp_enqueue_to_output_i(tcp_worker_ctx_t *wrk, vlib_buffer_t *b, u32 bi, u8 is_ip4, u8 flush)
struct _tcp_header tcp_header_t
int tcp_half_open_connection_cleanup(tcp_connection_t *tc)
Try to cleanup half-open connection.
#define scoreboard_rescue_rxt_valid(_sb, _tc)
struct _sack_scoreboard_hole sack_scoreboard_hole_t
u8 wscale
Window scale advertised.
enum fib_protocol_t_ fib_protocol_t
Protocol Type.
#define TCP_OPTS_MAX_SACK_BLOCKS
#define TCP_MAX_RX_FIFO_SIZE
vlib_node_registration_t ip4_lookup_node
(constructor) VLIB_REGISTER_NODE (ip4_lookup_node)
static void tcp_timer_retransmit_handler_i(u32 index, u8 is_syn)
#define foreach_tcp4_reset_next
static u32 tcp_prepare_retransmit_segment(tcp_worker_ctx_t *wrk, tcp_connection_t *tc, u32 offset, u32 max_deq_bytes, vlib_buffer_t **b)
Build a retransmit segment.
static uword tcp46_output_inline(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame, int is_ip4)
static void * tcp_init_buffer(vlib_main_t *vm, vlib_buffer_t *b)
static ip_adjacency_t * adj_get(adj_index_t adj_index)
Get a pointer to an adjacency object from its index.
void tcp_make_syn(tcp_connection_t *tc, vlib_buffer_t *b)
Convert buffer to SYN.
static int tcp_prepare_segment(tcp_worker_ctx_t *wrk, tcp_connection_t *tc, u32 offset, u32 max_deq_bytes, vlib_buffer_t **b)
Allocate a new buffer and build a new tcp segment.
sack_scoreboard_hole_t * scoreboard_get_hole(sack_scoreboard_t *sb, u32 index)
#define TCP_OPTION_LEN_SACK_BLOCK
static tcp_header_t * tcp_buffer_hdr(vlib_buffer_t *b)
#define vlib_prefetch_buffer_header(b, type)
Prefetch buffer metadata.
vlib_frame_t * vlib_get_frame_to_node(vlib_main_t *vm, u32 to_node_index)
enum _tcp_state tcp_state_t
#define TCP_ALWAYS_ACK
On/off delayed acks.
vhost_vring_state_t state
static void * ip4_next_header(ip4_header_t *i)
static u32 tcp_time_now(void)
sack_block_t * sacks
SACK blocks.
static void tcp46_output_trace_frame(vlib_main_t *vm, vlib_node_runtime_t *node, u32 *to_next, u32 n_bufs)
#define TCP_ESTABLISH_TIME
sack_scoreboard_hole_t * scoreboard_next_rxt_hole(sack_scoreboard_t *sb, sack_scoreboard_hole_t *start, u8 have_sent_1_smss, u8 *can_rescue, u8 *snd_limited)
Figure out the next hole to retransmit.
#define tcp_validate_txf_size(_tc, _a)
static void tcp_enqueue_to_ip_lookup_now(tcp_worker_ctx_t *wrk, vlib_buffer_t *b, u32 bi, u8 is_ip4, u32 fib_index)
static void tcp_push_hdr_i(tcp_connection_t *tc, vlib_buffer_t *b, u32 snd_nxt, u8 compute_opts, u8 maybe_burst, u8 update_snd_nxt)
Push TCP header and update connection variables.
#define TCP_EVT_DBG(_evt, _args...)
static u32 vlib_get_buffer_index(vlib_main_t *vm, void *p)
Translate buffer pointer into buffer index.
u32 tcp_session_push_header(transport_connection_t *tconn, vlib_buffer_t *b)
static void tcp_timer_set(tcp_connection_t *tc, u8 timer_id, u32 interval)
#define TCP_OPTION_LEN_WINDOW_SCALE
vlib_node_registration_t tcp6_reset_node
(constructor) VLIB_REGISTER_NODE (tcp6_reset_node)
#define TCP_RTO_SYN_RETRIES
vlib_error_t error
Error code for buffers to be enqueued to error handler.
int tcp_fast_retransmit_sack(tcp_worker_ctx_t *wrk, tcp_connection_t *tc, u32 burst_size)
Do fast retransmit with SACKs.
#define tcp_trajectory_add_start(b, start)
#define TRANSPORT_MAX_HDRS_LEN
vlib_main_t * vm
convenience pointer to this thread's vlib main
void tcp_send_reset(tcp_connection_t *tc)
Build and set reset packet for connection.
void tcp_send_synack(tcp_connection_t *tc)
#define ADJ_INDEX_INVALID
Invalid ADJ index - used when no adj is known likewise blazoned capitals INVALID speak volumes where ...
static int tcp_make_synack_options(tcp_connection_t *tc, tcp_options_t *opts)
static void * vlib_buffer_make_headroom(vlib_buffer_t *b, u8 size)
Make head room, typically for packet headers.
#define tcp_in_fastrecovery(tc)
void tcp_connection_tx_pacer_reset(tcp_connection_t *tc, u32 window, u32 start_bucket)
static void * vlib_buffer_push_tcp_net_order(vlib_buffer_t *b, u16 sp, u16 dp, u32 seq, u32 ack, u8 tcp_hdr_opts_len, u8 flags, u16 wnd)
Push TCP header to buffer.
#define tcp_opts_mss(_to)
void tcp_flush_frames_to_output(tcp_worker_ctx_t *wrk)
Flush v4 and v6 tcp and ip-lookup tx frames for thread index.
void vlib_put_frame_to_node(vlib_main_t *vm, u32 to_node_index, vlib_frame_t *f)
static void * vlib_buffer_get_current(vlib_buffer_t *b)
Get pointer to current data to process.
#define TCP_TIMER_HANDLE_INVALID
static void tcp_output_handle_link_local(tcp_connection_t *tc0, vlib_buffer_t *b0, u16 *next0, u32 *error0)
#define foreach_tcp6_output_next
static u32 tcp_flight_size(const tcp_connection_t *tc)
Our estimate of the number of bytes in flight (pipe size)
static int tcp_make_reset_in_place(vlib_main_t *vm, vlib_buffer_t *b0, tcp_state_t state, u8 thread_index, u8 is_ip4)
static u8 tcp_window_compute_scale(u32 window)
void tcp_timer_retransmit_syn_handler(u32 index)
#define vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, bi0, next0)
Finish enqueueing one buffer forward in the graph.
#define vlib_get_next_frame(vm, node, next_index, vectors, n_vectors_left)
Get pointer to next frame vector data by (vlib_node_runtime_t, next_index).
#define TCP_OPTION_LEN_TIMESTAMP
#define foreach_tcp4_output_next
static void tcp_enqueue_to_ip_lookup(tcp_worker_ctx_t *wrk, vlib_buffer_t *b, u32 bi, u8 is_ip4, u32 fib_index)
#define TCP_DBG(_fmt, _args...)
#define TCP_MAX_WND_SCALE
static void tcp_timer_reset(tcp_connection_t *tc, u8 timer_id)
This packet matches an "incomplete adjacency" and packets need to be passed to ARP to find rewrite st...
#define VLIB_REGISTER_NODE(x,...)
static void * vlib_buffer_push_tcp(vlib_buffer_t *b, u16 sp_net, u16 dp_net, u32 seq, u32 ack, u8 tcp_hdr_opts_len, u8 flags, u16 wnd)
Push TCP header to buffer.
static_always_inline uword vlib_get_thread_index(void)
#define CLIB_PREFETCH(addr, size, type)
static_always_inline void vlib_buffer_enqueue_to_next(vlib_main_t *vm, vlib_node_runtime_t *node, u32 *buffers, u16 *nexts, uword count)
void tcp_send_reset_w_pkt(tcp_connection_t *tc, vlib_buffer_t *pkt, u32 thread_index, u8 is_ip4)
Send reset without reusing existing buffer.
#define clib_warning(format, args...)
struct _transport_connection transport_connection_t
#define TCP_USE_SACKS
Disable only for testing.
#define tcp_recovery_on(tc)
static u32 tcp_window_to_advertise(tcp_connection_t *tc, tcp_state_t state)
Compute and return window to advertise, scaled as per RFC1323.
#define tcp_fastrecovery_first(tc)
u32 adj_index_t
An index for adjacencies.
void vlib_put_next_frame(vlib_main_t *vm, vlib_node_runtime_t *r, u32 next_index, u32 n_vectors_left)
Release pointer to next frame vector data.
u16 mss
Option flags, see above.
static void tcp_output_handle_packet(tcp_connection_t *tc0, vlib_buffer_t *b0, u32 *error0, u16 *next0, u8 is_ip4)
static void * ip6_next_header(ip6_header_t *i)
static void tcp_make_ack(tcp_connection_t *tc, vlib_buffer_t *b)
Convert buffer to ACK.
void tcp_program_fastretransmit(tcp_worker_ctx_t *wrk, tcp_connection_t *tc)
static u32 transport_max_tx_dequeue(transport_connection_t *tc)
static void tcp_timer_update(tcp_connection_t *tc, u8 timer_id, u32 interval)
u16 ip6_tcp_udp_icmp_compute_checksum(vlib_main_t *vm, vlib_buffer_t *p0, ip6_header_t *ip0, int *bogus_lengthp)
vlib_node_registration_t ip6_lookup_node
(constructor) VLIB_REGISTER_NODE (ip6_lookup_node)
static int tcp_make_established_options(tcp_connection_t *tc, tcp_options_t *opts)
u16 cached_next_index
Next frame index that vector arguments were last enqueued to last time this node ran.
static void tcp_output_push_ip(vlib_main_t *vm, vlib_buffer_t *b0, tcp_connection_t *tc0, u8 is_ip4)
static u8 * format_tcp_tx_trace(u8 *s, va_list *args)
u16 ip4_tcp_udp_compute_checksum(vlib_main_t *vm, vlib_buffer_t *p0, ip4_header_t *ip0)
void tcp_update_burst_snd_vars(tcp_connection_t *tc)
Update burst send vars.
#define seq_geq(_s1, _s2)
void tcp_init_mss(tcp_connection_t *tc)
static uword ip6_address_is_link_local_unicast(const ip6_address_t *a)
#define tcp_fastrecovery_first_off(tc)
static void tcp_update_rcv_wnd(tcp_connection_t *tc)
void tcp_send_fin(tcp_connection_t *tc)
Send FIN.
static u32 tcp_time_now_w_thread(u32 thread_index)
void tcp_send_ack(tcp_connection_t *tc)
static void * vlib_add_trace(vlib_main_t *vm, vlib_node_runtime_t *r, vlib_buffer_t *b, u32 n_data_bytes)
int tcp_retransmit_first_unacked(tcp_worker_ctx_t *wrk, tcp_connection_t *tc)
Retransmit first unacked segment.
template key/value backing page structure
#define tcp_opts_wscale(_to)
u32 tsval
Timestamp value.
u32 tsecr
Echoed/reflected time stamp.
static void * vlib_buffer_push_ip6(vlib_main_t *vm, vlib_buffer_t *b, ip6_address_t *src, ip6_address_t *dst, int proto)
Push IPv6 header to buffer.
#define vec_len(v)
Number of elements in vector (rvalue-only, NULL tolerant)
ip_lookup_next_t lookup_next_index
Next hop after ip4-lookup.
u32 next_buffer
Next buffer for this linked-list of buffers.
#define foreach_tcp6_reset_next
static int tcp_fast_retransmit_unsent(tcp_worker_ctx_t *wrk, tcp_connection_t *tc, u32 burst_size)
#define VLIB_BUFFER_TRACE_TRAJECTORY_INIT(b)
static tcp_worker_ctx_t * tcp_get_worker(u32 thread_index)
static void tcp_retransmit_timer_update(tcp_connection_t *tc)
VLIB buffer representation.
#define seq_max(_s1, _s2)
static void tcp_enqueue_to_ip_lookup_i(tcp_worker_ctx_t *wrk, vlib_buffer_t *b, u32 bi, u8 is_ip4, u32 fib_index, u8 flush)
static void * vlib_frame_vector_args(vlib_frame_t *f)
Get pointer to frame vector data.
static void tcp_make_ack_i(tcp_connection_t *tc, vlib_buffer_t *b, tcp_state_t state, u8 flags)
Prepare ACK.
void tcp_timer_delack_handler(u32 index)
Delayed ack timer handler.
#define TCP_OPTION_LEN_MSS
void tcp_flush_frame_to_output(tcp_worker_ctx_t *wrk, u8 is_ip4)
Flush tx frame populated by retransmits and timer pops.
struct clib_bihash_value offset
template key/value backing page structure
static void tcp_retransmit_timer_force_update(tcp_connection_t *tc)
int tcp_fast_retransmit(tcp_worker_ctx_t *wrk, tcp_connection_t *tc, u32 burst_size)
Do fast retransmit.
u8 * format_tcp_connection(u8 *s, va_list *args)
u32 tcp_initial_window_to_advertise(tcp_connection_t *tc)
Compute initial window and scale factor.
static tcp_connection_t * tcp_connection_get(u32 conn_index, u32 thread_index)
void tcp_update_rto(tcp_connection_t *tc)
static u32 vlib_num_workers()
#define TCP_OPTION_LEN_NOOP
void tcp_send_syn(tcp_connection_t *tc)
Send SYN.
vlib_node_registration_t tcp6_output_node
(constructor) VLIB_REGISTER_NODE (tcp6_output_node)
u16 flags
Copy of main node flags.
void tcp_program_ack(tcp_worker_ctx_t *wrk, tcp_connection_t *tc)
enum _tcp_reset_next tcp_reset_next_t
static u32 transport_max_rx_enqueue(transport_connection_t *tc)
#define tcp_opts_sack_permitted(_to)
static void vlib_buffer_free_one(vlib_main_t *vm, u32 buffer_index)
Free one buffer Shorthand to free a single buffer chain.
tcp_connection_t tcp_connection
static u32 tcp_loss_wnd(const tcp_connection_t *tc)
vlib_frame_t * ip_lookup_tx_frames[2]
tx frames for ip 4/6 lookup nodes
static void * tcp_reuse_buffer(vlib_main_t *vm, vlib_buffer_t *b)
static_always_inline void vlib_get_buffers(vlib_main_t *vm, u32 *bi, vlib_buffer_t **b, int count)
Translate array of buffer indices into buffer pointers.
static void tcp_enqueue_to_output_now(tcp_worker_ctx_t *wrk, vlib_buffer_t *b, u32 bi, u8 is_ip4)
vlib_node_registration_t tcp4_reset_node
(constructor) VLIB_REGISTER_NODE (tcp4_reset_node)
#define VLIB_NODE_FLAG_TRACE
vlib_node_registration_t tcp4_output_node
(constructor) VLIB_REGISTER_NODE (tcp4_output_node)
#define CLIB_CACHE_LINE_BYTES
u32 total_length_not_including_first_buffer
Only valid for first buffer in chain.
static void tcp_enqueue_to_output(tcp_worker_ctx_t *wrk, vlib_buffer_t *b, u32 bi, u8 is_ip4)
static u32 vlib_buffer_alloc(vlib_main_t *vm, u32 *buffers, u32 n_buffers)
Allocate buffers into supplied array.
static void tcp_persist_timer_set(tcp_connection_t *tc)
static tcp_main_t * vnet_get_tcp_main()
static char * tcp_error_strings[]
static void * vlib_buffer_push_ip4(vlib_main_t *vm, vlib_buffer_t *b, ip4_address_t *src, ip4_address_t *dst, int proto, u8 csum_offload)
Push IPv4 header to buffer.
static vlib_buffer_t * vlib_get_buffer(vlib_main_t *vm, u32 buffer_index)
Translate buffer index into buffer pointer.
void tcp_cc_fastrecovery_exit(tcp_connection_t *tc)
static u32 tcp_set_time_now(tcp_worker_ctx_t *wrk)
void tcp_timer_persist_handler(u32 index)
Got 0 snd_wnd from peer, try to do something about it.
static u8 tcp_timer_is_active(tcp_connection_t *tc, tcp_timers_e timer)
u32 * tx_buffers
tx buffer free list
adj_index_t adj_nbr_find(fib_protocol_t nh_proto, vnet_link_t link_type, const ip46_address_t *nh_addr, u32 sw_if_index)
Lookup neighbor adjancency.
static int tcp_make_options(tcp_connection_t *tc, tcp_options_t *opts, tcp_state_t state)
void tcp_program_dupack(tcp_worker_ctx_t *wrk, tcp_connection_t *tc)