40 tag = t->
is_slow_path ?
"NAT64-in2out-slowpath" :
"NAT64-in2out";
65 format (s,
"NAT64-in2out-reass: sw_if_index %d, next index %d, status %s",
67 t->
cached ?
"cached" :
"translated");
77 #define foreach_nat64_in2out_error \ 78 _(UNSUPPORTED_PROTOCOL, "unsupported protocol") \ 79 _(IN2OUT_PACKETS, "good in2out packets processed") \ 80 _(NO_TRANSLATION, "no translation") \ 81 _(UNKNOWN, "unknown") \ 82 _(DROP_FRAGMENT, "Drop fragment") \ 83 _(MAX_REASS, "Maximum reassemblies exceeded") \ 84 _(MAX_FRAG, "Maximum fragments per reassembly exceeded") 89 #define _(sym,str) NAT64_IN2OUT_ERROR_##sym, 96 #define _(sym,string) string, 149 nat64_db_bib_entry_t *bibe;
150 nat64_db_st_entry_t *ste;
151 ip46_address_t saddr, daddr;
152 u32 sw_if_index, fib_index;
193 sport, out_port, fib_index, proto, 0);
213 if (proto == IP_PROTOCOL_TCP)
219 checksum = &tcp->checksum;
233 nat64_db_bib_entry_t *bibe;
234 nat64_db_st_entry_t *ste;
235 ip46_address_t saddr, daddr;
236 u32 sw_if_index, fib_index;
249 if (icmp->type == ICMP4_echo_request || icmp->type == ICMP4_echo_reply)
251 u16 in_id = ((
u16 *) (icmp))[2];
254 IP_PROTOCOL_ICMP, fib_index, 1);
268 IP_PROTOCOL_ICMP, fib_index, 1);
275 (fib_index, SNAT_PROTOCOL_ICMP, &out_addr, &out_id,
281 &out_addr, in_id, out_id,
282 fib_index, IP_PROTOCOL_ICMP, 0);
298 ((
u16 *) (icmp))[2] = bibe->out_port;
320 nat64_db_st_entry_t *ste;
321 nat64_db_bib_entry_t *bibe;
322 ip46_address_t saddr, daddr;
323 u32 sw_if_index, fib_index;
336 if (proto == IP_PROTOCOL_ICMP6)
339 u16 in_id = ((
u16 *) (icmp))[2];
340 proto = IP_PROTOCOL_ICMP;
343 (icmp->type == ICMP4_echo_request
344 || icmp->type == ICMP4_echo_reply))
358 ((
u16 *) (icmp))[2] = bibe->out_port;
385 if (proto == IP_PROTOCOL_TCP)
386 checksum = &tcp->checksum;
412 nat64_db_bib_entry_t *bibe;
413 ip46_address_t saddr, daddr;
425 memset (&saddr, 0,
sizeof (saddr));
426 saddr.ip4.as_u32 = bibe->out_addr.as_u32;
427 memset (&daddr, 0,
sizeof (daddr));
448 nat64_db_bib_entry_t *bibe;
449 nat64_db_st_entry_t *ste;
450 ip46_address_t saddr, daddr,
addr;
451 u32 sw_if_index, fib_index;
486 .out_addr.as_u32 = 0,
487 .fib_index = fib_index,
498 memset (&addr, 0,
sizeof (addr));
518 &ctx.
out_addr, 0, 0, fib_index, proto,
546 nat64_db_bib_entry_t *bibe;
547 nat64_db_st_entry_t *ste;
548 ip46_address_t saddr, daddr;
549 u32 sw_if_index, fib_index;
568 if (proto == IP_PROTOCOL_UDP)
571 checksum = &tcp->checksum;
600 &out_port, thread_index))
605 sport, out_port, fib_index, proto, 0);
620 sport = udp->
src_port = bibe->out_port;
623 memset (&daddr, 0,
sizeof (daddr));
624 daddr.ip4.as_u32 = ste->out_r_addr.as_u32;
660 nat64_db_bib_entry_t *bibe;
661 nat64_db_st_entry_t *ste;
664 ip46_address_t saddr, daddr;
665 u32 sw_if_index, fib_index;
669 u16 *checksum, sport, dport;
673 if (icmp->type == ICMP6_echo_request || icmp->type == ICMP6_echo_reply)
680 if (proto == IP_PROTOCOL_ICMP6)
698 if (proto == IP_PROTOCOL_UDP)
701 checksum = &tcp->checksum;
720 dport = udp->
dst_port = bibe->out_port;
723 memset (&saddr, 0,
sizeof (saddr));
724 memset (&daddr, 0,
sizeof (daddr));
725 saddr.ip4.as_u32 = ste->out_r_addr.as_u32;
726 daddr.ip4.as_u32 = bibe->out_addr.as_u32;
786 nat64_db_bib_entry_t *bibe;
787 nat64_db_st_entry_t *ste;
788 ip46_address_t saddr, daddr,
addr;
789 u32 sw_if_index, fib_index;
824 .out_addr.as_u32 = 0,
825 .fib_index = fib_index,
836 memset (&addr, 0,
sizeof (addr));
856 &ctx.
out_addr, 0, 0, fib_index, proto,
873 memset (&daddr, 0,
sizeof (daddr));
874 daddr.ip4.as_u32 = ste->out_r_addr.as_u32;
900 u32 n_left_from, *from, *to_next;
902 u32 pkts_processed = 0;
903 u32 stats_node_index;
913 while (n_left_from > 0)
919 while (n_left_from > 0 && n_left_to_next > 0)
925 u16 l4_offset0, frag_offset0;
953 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_UNKNOWN];
967 (vm, b0, ip60, thread_index))
971 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
980 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
996 (ip60->
protocol == IP_PROTOCOL_IPV6_FRAGMENTATION))
1002 if (proto0 == SNAT_PROTOCOL_ICMP)
1008 (vm, b0, ip60, thread_index))
1012 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1022 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1026 else if (proto0 == SNAT_PROTOCOL_TCP || proto0 == SNAT_PROTOCOL_UDP)
1032 (vm, b0, ip60, thread_index))
1036 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1045 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1052 && (b0->
flags & VLIB_BUFFER_IS_TRACED)))
1065 n_left_to_next, bi0, next0);
1070 NAT64_IN2OUT_ERROR_IN2OUT_PACKETS,
1085 .name =
"nat64-in2out",
1086 .vector_size =
sizeof (
u32),
1115 .name =
"nat64-in2out-slowpath",
1116 .vector_size =
sizeof (
u32),
1151 nat64_db_st_entry_t *ste;
1152 nat64_db_bib_entry_t *bibe;
1170 if (ctx->
proto == IP_PROTOCOL_TCP)
1202 nat64_db_st_entry_t *ste;
1203 nat64_db_bib_entry_t *bibe;
1210 ip46_address_t daddr;
1215 if (ctx->
proto == IP_PROTOCOL_UDP)
1218 checksum = &tcp->checksum;
1238 sport = bibe->out_port;
1239 dport = ste->r_port;
1243 memset (&daddr, 0,
sizeof (daddr));
1244 daddr.ip4.as_u32 = ste->out_r_addr.as_u32;
1283 u32 n_left_from, *from, *to_next;
1285 u32 pkts_processed = 0;
1286 u32 *fragments_to_drop = 0;
1287 u32 *fragments_to_loopback = 0;
1295 while (n_left_from > 0)
1301 while (n_left_from > 0 && n_left_to_next > 0)
1308 u16 l4_offset0, frag_offset0;
1310 nat_reass_ip6_t *reass0;
1311 ip6_frag_hdr_t *frag0;
1312 nat64_db_bib_entry_t *bibe0;
1313 nat64_db_st_entry_t *ste0;
1316 u32 sw_if_index0, fib_index0;
1317 ip46_address_t saddr0, daddr0;
1327 n_left_to_next -= 1;
1342 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_DROP_FRAGMENT];
1354 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_UNKNOWN];
1359 (!(l4_protocol0 == IP_PROTOCOL_TCP
1360 || l4_protocol0 == IP_PROTOCOL_UDP)))
1363 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_DROP_FRAGMENT];
1368 frag0 = (ip6_frag_hdr_t *)
u8_ptr_add (ip60, frag_offset0);
1373 frag0->identification,
1375 1, &fragments_to_drop);
1380 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_MAX_REASS];
1391 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_MAX_FRAG];
1411 l4_protocol0, fib_index0, 1);
1416 l4_protocol0, fib_index0, 1);
1422 (fib_index0, proto0, &out_addr0, &out_port0,
1427 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1435 out_port0, fib_index0,
1441 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1455 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1465 ctx0.
proto = l4_protocol0;
1475 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1484 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_UNKNOWN];
1492 && (b0->
flags & VLIB_BUFFER_IS_TRACED)))
1512 to_next, n_left_to_next,
1516 if (n_left_from == 0 &&
vec_len (fragments_to_loopback))
1523 sizeof (
u32) * len);
1530 fragments_to_loopback + (len -
1543 NAT64_IN2OUT_ERROR_IN2OUT_PACKETS,
1547 &node->
errors[NAT64_IN2OUT_ERROR_DROP_FRAGMENT],
1558 .name =
"nat64-in2out-reass",
1559 .vector_size =
sizeof (
u32),
1594 m = t->
do_handoff ?
"next worker" :
"same worker";
1606 u32 n_left_from, *from, *to_next = 0, *to_next_drop = 0;
1614 u32 n_left_to_next_worker = 0, *to_next_worker = 0;
1615 u32 next_worker_index = 0;
1616 u32 current_worker_index = ~0;
1636 while (n_left_from > 0)
1657 if (next_worker_index != current_worker_index)
1662 congested_handoff_queue_by_worker_index);
1673 to_next_drop[0] = bi0;
1685 handoff_queue_elt_by_worker_index);
1688 current_worker_index = next_worker_index;
1691 ASSERT (to_next_worker != 0);
1694 to_next_worker[0] = bi0;
1696 n_left_to_next_worker--;
1698 if (n_left_to_next_worker == 0)
1702 current_worker_index = ~0;
1703 handoff_queue_elt_by_worker_index[next_worker_index] = 0;
1725 && (b0->
flags & VLIB_BUFFER_IS_TRACED)))
1744 for (i = 0; i <
vec_len (handoff_queue_elt_by_worker_index); i++)
1746 if (handoff_queue_elt_by_worker_index[i])
1748 hf = handoff_queue_elt_by_worker_index[
i];
1756 handoff_queue_elt_by_worker_index[
i] = 0;
1761 congested_handoff_queue_by_worker_index[
i] =
1765 current_worker_index = ~0;
1772 .name =
"nat64-in2out-handoff",
1773 .vector_size =
sizeof (
u32),
#define vec_validate(V, I)
Make sure vector is long enough for given index (no header, unspecified alignment) ...
int nat_ip6_reass_add_fragment(nat_reass_ip6_t *reass, u32 bi)
Cache fragment.
nat64_db_t * db
BIB and session DB per thread.
static int unk_proto_st_walk(nat64_db_st_entry_t *ste, void *arg)
static int nat64_in2out_frag_set_cb(ip6_header_t *ip6, ip4_header_t *ip4, void *arg)
static int nat64_in2out_frag_hairpinning(vlib_buffer_t *b, ip6_header_t *ip6, nat64_in2out_frag_set_ctx_t *ctx)
snat_address_t * addr_pool
Address pool vector.
static int icmp6_to_icmp(vlib_buffer_t *p, ip6_to_ip4_set_fn_t fn, void *ctx, ip6_to_ip4_set_fn_t inner_fn, void *inner_ctx)
Translate ICMP6 packet to ICMP4.
void nat64_extract_ip4(ip6_address_t *ip6, ip4_address_t *ip4, u32 fib_index)
Extract IPv4 address from the IPv4-embedded IPv6 addresses.
vlib_node_registration_t nat64_in2out_node
(constructor) VLIB_REGISTER_NODE (nat64_in2out_node)
nat64_db_st_entry_t * nat64_db_st_entry_create(nat64_db_t *db, nat64_db_bib_entry_t *bibe, ip6_address_t *in_r_addr, ip4_address_t *out_r_addr, u16 r_port)
Create new NAT64 session table entry.
u32 fib_table_get_index_for_sw_if_index(fib_protocol_t proto, u32 sw_if_index)
Get the index of the FIB bound to the interface.
nat64_db_bib_entry_t * nat64_db_bib_entry_create(nat64_db_t *db, ip6_address_t *in_addr, ip4_address_t *out_addr, u16 in_port, u16 out_port, u32 fib_index, u8 proto, u8 is_static)
Create new NAT64 BIB entry.
void nat_ip6_reass_get_frags(nat_reass_ip6_t *reass, u32 **bi)
Get cached fragments.
nat64_db_bib_entry_t * nat64_db_bib_entry_find(nat64_db_t *db, ip46_address_t *addr, u16 port, u8 proto, u32 fib_index, u8 is_ip6)
Find NAT64 BIB entry.
static ip_csum_t ip_csum_with_carry(ip_csum_t sum, ip_csum_t x)
static u8 * format_nat64_in2out_trace(u8 *s, va_list *args)
nat64_db_st_entry_t * nat64_db_st_entry_by_index(nat64_db_t *db, u8 proto, u32 ste_index)
Get ST entry by index and protocol.
u32 buffer_index[VLIB_FRAME_SIZE]
vlib_error_t * errors
Vector of errors for this node.
vlib_node_registration_t nat64_in2out_handoff_node
(constructor) VLIB_REGISTER_NODE (nat64_in2out_handoff_node)
struct _tcp_header tcp_header_t
static vlib_frame_queue_t * is_vlib_frame_queue_congested(u32 frame_queue_index, u32 index, u32 queue_hi_thresh, vlib_frame_queue_t **handoff_queue_by_worker_index)
#define vec_reset_length(v)
Reset vector length to zero NULL-pointer tolerant.
static int nat64_in2out_tcp_udp_hairpinning(vlib_main_t *vm, vlib_buffer_t *b, ip6_header_t *ip6, u32 thread_index)
u32 nat64_get_worker_in2out(ip6_address_t *addr)
Get worker thread index for NAT64 in2out.
#define static_always_inline
ip_csum_t ip_incremental_checksum(ip_csum_t sum, void *_data, uword n_bytes)
vlib_node_registration_t nat64_in2out_reass_node
(constructor) VLIB_REGISTER_NODE (nat64_in2out_reass_node)
vlib_frame_t * vlib_get_frame_to_node(vlib_main_t *vm, u32 to_node_index)
static_always_inline int ip6_parse(const ip6_header_t *ip6, u32 buff_len, u8 *l4_protocol, u16 *l4_offset, u16 *frag_hdr_offset)
Parse some useful information from IPv6 header.
struct unk_proto_st_walk_ctx_t_ unk_proto_st_walk_ctx_t
static int ip6_to_ip4_tcp_udp(vlib_buffer_t *p, ip6_to_ip4_set_fn_t fn, void *ctx, u8 udp_checksum)
Translate IPv6 UDP/TCP packet to IPv4.
u16 current_length
Nbytes between current data and the end of this buffer.
void nat64_session_reset_timeout(nat64_db_st_entry_t *ste, vlib_main_t *vm)
Reset NAT64 session timeout.
static uword nat64_in2out_slowpath_node_fn(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame)
int nat64_alloc_out_addr_and_port(u32 fib_index, snat_protocol_t proto, ip4_address_t *addr, u16 *port, u32 thread_index)
Alloce IPv4 address and port pair from NAT64 pool.
static u8 * format_nat64_in2out_reass_trace(u8 *s, va_list *args)
void vlib_put_frame_to_node(vlib_main_t *vm, u32 to_node_index, vlib_frame_t *f)
static void * vlib_buffer_get_current(vlib_buffer_t *b)
Get pointer to current data to process.
static vlib_frame_queue_elt_t * vlib_get_worker_handoff_queue_elt(u32 frame_queue_index, u32 vlib_worker_index, vlib_frame_queue_elt_t **handoff_queue_elt_by_worker_index)
static int nat64_in2out_inner_icmp_set_cb(ip6_header_t *ip6, ip4_header_t *ip4, void *arg)
static char * nat64_in2out_error_strings[]
#define vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, bi0, next0)
Finish enqueueing one buffer forward in the graph.
#define vlib_get_next_frame(vm, node, next_index, vectors, n_vectors_left)
Get pointer to next frame vector data by (vlib_node_runtime_t, next_index).
void nat64_compose_ip6(ip6_address_t *ip6, ip4_address_t *ip4, u32 fib_index)
Compose IPv4-embedded IPv6 addresses.
vlib_error_t error
Error code for buffers to be enqueued to error handler.
static void vlib_node_increment_counter(vlib_main_t *vm, u32 node_index, u32 counter_index, u64 increment)
static u8 * format_nat64_in2out_handoff_trace(u8 *s, va_list *args)
#define VLIB_REGISTER_NODE(x,...)
static int nat64_in2out_unk_proto_hairpinning(vlib_main_t *vm, vlib_buffer_t *b, ip6_header_t *ip6, u32 thread_index)
static int nat64_in2out_icmp_set_cb(ip6_header_t *ip6, ip4_header_t *ip4, void *arg)
static_always_inline uword vlib_get_thread_index(void)
u8 nat_reass_is_drop_frag(u8 is_ip6)
Get status of virtual fragmentation reassembly.
static int nat64_in2out_tcp_udp_set_cb(ip6_header_t *ip6, ip4_header_t *ip4, void *arg)
#define vec_free(V)
Free vector's memory (no header).
vlib_node_registration_t nat64_in2out_slowpath_node
(constructor) VLIB_REGISTER_NODE (nat64_in2out_slowpath_node)
#define clib_memcpy(a, b, c)
void vlib_put_next_frame(vlib_main_t *vm, vlib_node_runtime_t *r, u32 next_index, u32 n_vectors_left)
Release pointer to next frame vector data.
static void * ip6_next_header(ip6_header_t *i)
#define ip6_frag_hdr_offset(hdr)
u16 cached_next_index
Next frame index that vector arguments were last enqueued to last time this node ran.
nat64_db_st_entry_t * nat64_db_st_entry_find(nat64_db_t *db, ip46_address_t *l_addr, ip46_address_t *r_addr, u16 l_port, u16 r_port, u8 proto, u32 fib_index, u8 is_ip6)
Find NAT64 session table entry.
void nat64_db_st_walk(nat64_db_t *db, u8 proto, nat64_db_st_walk_fn_t fn, void *ctx)
Walk NAT64 session table.
static int nat64_in2out_icmp_hairpinning(vlib_main_t *vm, vlib_buffer_t *b, ip6_header_t *ip6, u32 thread_index)
u32 fq_in2out_index
Worker handoff.
u32 nat64_db_st_entry_get_index(nat64_db_t *db, nat64_db_st_entry_t *ste)
static u32 ip_proto_to_snat_proto(u8 ip_proto)
static ip_csum_t ip_csum_sub_even(ip_csum_t c, ip_csum_t x)
nat64_db_bib_entry_t * nat64_db_bib_entry_by_index(nat64_db_t *db, u8 proto, u32 bibe_index)
Get BIB entry by index and protocol.
IPv6 to IPv4 translation.
static uword nat64_in2out_reass_node_fn(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame)
#define u8_ptr_add(ptr, index)
static void * vlib_add_trace(vlib_main_t *vm, vlib_node_runtime_t *r, vlib_buffer_t *b, u32 n_data_bytes)
struct _vlib_node_registration vlib_node_registration_t
NAT64 global declarations.
VLIB_NODE_FUNCTION_MULTIARCH(nat64_in2out_node, nat64_in2out_node_fn)
#define ip46_address_is_equal(a1, a2)
#define vec_len(v)
Number of elements in vector (rvalue-only, NULL tolerant)
nat_reass_ip6_t * nat_ip6_reass_find_or_create(ip6_address_t src, ip6_address_t dst, u32 frag_id, u8 proto, u8 reset_timeout, u32 **bi_to_drop)
Find or create reassembly.
static void * vlib_frame_vector_args(vlib_frame_t *f)
Get pointer to frame vector data.
static uword nat64_in2out_handoff_node_fn(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame)
static void vlib_put_frame_queue_elt(vlib_frame_queue_elt_t *hf)
static uword nat64_in2out_node_fn_inline(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame, u8 is_slow_path)
static int nat64_in2out_unk_proto_set_cb(ip6_header_t *ip6, ip4_header_t *ip4, void *arg)
static vlib_thread_main_t * vlib_get_thread_main()
static_always_inline int is_hairpinning(ip6_address_t *dst_addr)
Check whether is a hairpinning.
#define vec_foreach(var, vec)
Vector iterator.
static uword nat64_in2out_node_fn(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame)
static_always_inline void nat_send_all_to_node(vlib_main_t *vm, u32 *bi_vector, vlib_node_runtime_t *node, vlib_error_t *error, u32 next)
u16 flags
Copy of main node flags.
NAT plugin virtual fragmentation reassembly.
#define VLIB_NODE_FLAG_TRACE
#define foreach_nat64_in2out_error
#define vec_validate_init_empty(V, I, INIT)
Make sure vector is long enough for given index and initialize empty space (no header, unspecified alignment)
struct nat64_in2out_set_ctx_t_ nat64_in2out_set_ctx_t
u32 flags
buffer flags: VLIB_BUFFER_FREE_LIST_INDEX_MASK: bits used to store free list index, VLIB_BUFFER_IS_TRACED: trace this buffer.
static vlib_buffer_t * vlib_get_buffer(vlib_main_t *vm, u32 buffer_index)
Translate buffer index into buffer pointer.
static int ip6_to_ip4(vlib_buffer_t *p, ip6_to_ip4_set_fn_t fn, void *ctx)
Translate IPv6 packet to IPv4 (IP header only).
static u16 ip_csum_fold(ip_csum_t c)
struct nat64_in2out_frag_set_ctx_t_ nat64_in2out_frag_set_ctx_t
static ip_csum_t ip_csum_add_even(ip_csum_t c, ip_csum_t x)
static int ip6_to_ip4_fragmented(vlib_buffer_t *p, ip6_to_ip4_set_fn_t fn, void *ctx)
Translate IPv6 fragmented packet to IPv4.