41 tag = t->
is_slow_path ?
"NAT64-in2out-slowpath" :
"NAT64-in2out";
66 format (s,
"NAT64-in2out-reass: sw_if_index %d, next index %d, status %s",
68 t->
cached ?
"cached" :
"translated");
78 #define foreach_nat64_in2out_error \ 79 _(UNSUPPORTED_PROTOCOL, "unsupported protocol") \ 80 _(IN2OUT_PACKETS, "good in2out packets processed") \ 81 _(NO_TRANSLATION, "no translation") \ 82 _(UNKNOWN, "unknown") \ 83 _(DROP_FRAGMENT, "Drop fragment") \ 84 _(MAX_REASS, "Maximum reassemblies exceeded") \ 85 _(MAX_FRAG, "Maximum fragments per reassembly exceeded") 90 #define _(sym,str) NAT64_IN2OUT_ERROR_##sym, 97 #define _(sym,string) string, 150 nat64_db_bib_entry_t *bibe;
151 nat64_db_st_entry_t *ste;
152 ip46_address_t saddr, daddr;
153 u32 sw_if_index, fib_index;
194 sport, out_port, fib_index, proto, 0);
214 if (proto == IP_PROTOCOL_TCP)
220 checksum = &tcp->checksum;
234 nat64_db_bib_entry_t *bibe;
235 nat64_db_st_entry_t *ste;
236 ip46_address_t saddr, daddr;
237 u32 sw_if_index, fib_index;
250 if (icmp->type == ICMP4_echo_request || icmp->type == ICMP4_echo_reply)
252 u16 in_id = ((
u16 *) (icmp))[2];
255 IP_PROTOCOL_ICMP, fib_index, 1);
269 IP_PROTOCOL_ICMP, fib_index, 1);
276 (fib_index, SNAT_PROTOCOL_ICMP, &out_addr, &out_id,
282 &out_addr, in_id, out_id,
283 fib_index, IP_PROTOCOL_ICMP, 0);
299 ((
u16 *) (icmp))[2] = bibe->out_port;
321 nat64_db_st_entry_t *ste;
322 nat64_db_bib_entry_t *bibe;
323 ip46_address_t saddr, daddr;
324 u32 sw_if_index, fib_index;
337 if (proto == IP_PROTOCOL_ICMP6)
340 u16 in_id = ((
u16 *) (icmp))[2];
341 proto = IP_PROTOCOL_ICMP;
344 (icmp->type == ICMP4_echo_request
345 || icmp->type == ICMP4_echo_reply))
359 ((
u16 *) (icmp))[2] = bibe->out_port;
386 if (proto == IP_PROTOCOL_TCP)
387 checksum = &tcp->checksum;
413 nat64_db_bib_entry_t *bibe;
414 ip46_address_t saddr, daddr;
426 memset (&saddr, 0,
sizeof (saddr));
427 saddr.ip4.as_u32 = bibe->out_addr.as_u32;
428 memset (&daddr, 0,
sizeof (daddr));
449 nat64_db_bib_entry_t *bibe;
450 nat64_db_st_entry_t *ste;
451 ip46_address_t saddr, daddr,
addr;
452 u32 sw_if_index, fib_index;
487 .out_addr.as_u32 = 0,
488 .fib_index = fib_index,
499 memset (&addr, 0,
sizeof (addr));
519 &ctx.
out_addr, 0, 0, fib_index, proto,
547 nat64_db_bib_entry_t *bibe;
548 nat64_db_st_entry_t *ste;
549 ip46_address_t saddr, daddr;
550 u32 sw_if_index, fib_index;
569 if (proto == IP_PROTOCOL_UDP)
572 checksum = &tcp->checksum;
601 &out_port, thread_index))
606 sport, out_port, fib_index, proto, 0);
621 sport = udp->
src_port = bibe->out_port;
624 memset (&daddr, 0,
sizeof (daddr));
625 daddr.ip4.as_u32 = ste->out_r_addr.as_u32;
661 nat64_db_bib_entry_t *bibe;
662 nat64_db_st_entry_t *ste;
665 ip46_address_t saddr, daddr;
666 u32 sw_if_index, fib_index;
670 u16 *checksum, sport, dport;
674 if (icmp->type == ICMP6_echo_request || icmp->type == ICMP6_echo_reply)
681 if (proto == IP_PROTOCOL_ICMP6)
699 if (proto == IP_PROTOCOL_UDP)
702 checksum = &tcp->checksum;
721 dport = udp->
dst_port = bibe->out_port;
724 memset (&saddr, 0,
sizeof (saddr));
725 memset (&daddr, 0,
sizeof (daddr));
726 saddr.ip4.as_u32 = ste->out_r_addr.as_u32;
727 daddr.ip4.as_u32 = bibe->out_addr.as_u32;
787 nat64_db_bib_entry_t *bibe;
788 nat64_db_st_entry_t *ste;
789 ip46_address_t saddr, daddr,
addr;
790 u32 sw_if_index, fib_index;
825 .out_addr.as_u32 = 0,
826 .fib_index = fib_index,
837 memset (&addr, 0,
sizeof (addr));
857 &ctx.
out_addr, 0, 0, fib_index, proto,
874 memset (&daddr, 0,
sizeof (daddr));
875 daddr.ip4.as_u32 = ste->out_r_addr.as_u32;
901 u32 n_left_from, *from, *to_next;
903 u32 pkts_processed = 0;
904 u32 stats_node_index;
914 while (n_left_from > 0)
920 while (n_left_from > 0 && n_left_to_next > 0)
926 u16 l4_offset0, frag_offset0;
954 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_UNKNOWN];
968 (vm, b0, ip60, thread_index))
972 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
981 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
997 (ip60->
protocol == IP_PROTOCOL_IPV6_FRAGMENTATION))
1003 if (proto0 == SNAT_PROTOCOL_ICMP)
1009 (vm, b0, ip60, thread_index))
1013 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1023 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1027 else if (proto0 == SNAT_PROTOCOL_TCP || proto0 == SNAT_PROTOCOL_UDP)
1033 (vm, b0, ip60, thread_index))
1037 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1046 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1053 && (b0->
flags & VLIB_BUFFER_IS_TRACED)))
1066 n_left_to_next, bi0, next0);
1071 NAT64_IN2OUT_ERROR_IN2OUT_PACKETS,
1086 .name =
"nat64-in2out",
1087 .vector_size =
sizeof (
u32),
1116 .name =
"nat64-in2out-slowpath",
1117 .vector_size =
sizeof (
u32),
1152 nat64_db_st_entry_t *ste;
1153 nat64_db_bib_entry_t *bibe;
1171 if (ctx->
proto == IP_PROTOCOL_TCP)
1203 nat64_db_st_entry_t *ste;
1204 nat64_db_bib_entry_t *bibe;
1211 ip46_address_t daddr;
1216 if (ctx->
proto == IP_PROTOCOL_UDP)
1219 checksum = &tcp->checksum;
1239 sport = bibe->out_port;
1240 dport = ste->r_port;
1244 memset (&daddr, 0,
sizeof (daddr));
1245 daddr.ip4.as_u32 = ste->out_r_addr.as_u32;
1284 u32 n_left_from, *from, *to_next;
1286 u32 pkts_processed = 0;
1287 u32 *fragments_to_drop = 0;
1288 u32 *fragments_to_loopback = 0;
1296 while (n_left_from > 0)
1302 while (n_left_from > 0 && n_left_to_next > 0)
1309 u16 l4_offset0, frag_offset0;
1311 nat_reass_ip6_t *reass0;
1312 ip6_frag_hdr_t *frag0;
1313 nat64_db_bib_entry_t *bibe0;
1314 nat64_db_st_entry_t *ste0;
1317 u32 sw_if_index0, fib_index0;
1318 ip46_address_t saddr0, daddr0;
1328 n_left_to_next -= 1;
1343 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_DROP_FRAGMENT];
1355 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_UNKNOWN];
1360 (!(l4_protocol0 == IP_PROTOCOL_TCP
1361 || l4_protocol0 == IP_PROTOCOL_UDP)))
1364 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_DROP_FRAGMENT];
1369 frag0 = (ip6_frag_hdr_t *)
u8_ptr_add (ip60, frag_offset0);
1374 frag0->identification,
1376 1, &fragments_to_drop);
1381 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_MAX_REASS];
1392 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_MAX_FRAG];
1412 l4_protocol0, fib_index0, 1);
1417 l4_protocol0, fib_index0, 1);
1423 (fib_index0, proto0, &out_addr0, &out_port0,
1428 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1436 out_port0, fib_index0,
1442 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1456 node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1466 ctx0.
proto = l4_protocol0;
1476 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_NO_TRANSLATION];
1485 b0->
error = node->
errors[NAT64_IN2OUT_ERROR_UNKNOWN];
1493 && (b0->
flags & VLIB_BUFFER_IS_TRACED)))
1513 to_next, n_left_to_next,
1517 if (n_left_from == 0 &&
vec_len (fragments_to_loopback))
1524 sizeof (
u32) * len);
1531 fragments_to_loopback + (len -
1544 NAT64_IN2OUT_ERROR_IN2OUT_PACKETS,
1548 &node->
errors[NAT64_IN2OUT_ERROR_DROP_FRAGMENT],
1559 .name =
"nat64-in2out-reass",
1560 .vector_size =
sizeof (
u32),
1595 m = t->
do_handoff ?
"next worker" :
"same worker";
1607 u32 n_left_from, *from, *to_next = 0, *to_next_drop = 0;
1615 u32 n_left_to_next_worker = 0, *to_next_worker = 0;
1616 u32 next_worker_index = 0;
1617 u32 current_worker_index = ~0;
1637 while (n_left_from > 0)
1658 if (next_worker_index != current_worker_index)
1663 congested_handoff_queue_by_worker_index);
1674 to_next_drop[0] = bi0;
1686 handoff_queue_elt_by_worker_index);
1689 current_worker_index = next_worker_index;
1692 ASSERT (to_next_worker != 0);
1695 to_next_worker[0] = bi0;
1697 n_left_to_next_worker--;
1699 if (n_left_to_next_worker == 0)
1703 current_worker_index = ~0;
1704 handoff_queue_elt_by_worker_index[next_worker_index] = 0;
1726 && (b0->
flags & VLIB_BUFFER_IS_TRACED)))
1745 for (i = 0; i <
vec_len (handoff_queue_elt_by_worker_index); i++)
1747 if (handoff_queue_elt_by_worker_index[i])
1749 hf = handoff_queue_elt_by_worker_index[
i];
1757 handoff_queue_elt_by_worker_index[
i] = 0;
1762 congested_handoff_queue_by_worker_index[
i] =
1766 current_worker_index = ~0;
1773 .name =
"nat64-in2out-handoff",
1774 .vector_size =
sizeof (
u32),
#define vec_validate(V, I)
Make sure vector is long enough for given index (no header, unspecified alignment) ...
int nat_ip6_reass_add_fragment(nat_reass_ip6_t *reass, u32 bi)
Cache fragment.
nat64_db_t * db
BIB and session DB per thread.
static int unk_proto_st_walk(nat64_db_st_entry_t *ste, void *arg)
static int nat64_in2out_frag_set_cb(ip6_header_t *ip6, ip4_header_t *ip4, void *arg)
static int nat64_in2out_frag_hairpinning(vlib_buffer_t *b, ip6_header_t *ip6, nat64_in2out_frag_set_ctx_t *ctx)
snat_address_t * addr_pool
Address pool vector.
static int icmp6_to_icmp(vlib_buffer_t *p, ip6_to_ip4_set_fn_t fn, void *ctx, ip6_to_ip4_set_fn_t inner_fn, void *inner_ctx)
Translate ICMP6 packet to ICMP4.
void nat64_extract_ip4(ip6_address_t *ip6, ip4_address_t *ip4, u32 fib_index)
Extract IPv4 address from the IPv4-embedded IPv6 addresses.
vlib_node_registration_t nat64_in2out_node
(constructor) VLIB_REGISTER_NODE (nat64_in2out_node)
nat64_db_st_entry_t * nat64_db_st_entry_create(nat64_db_t *db, nat64_db_bib_entry_t *bibe, ip6_address_t *in_r_addr, ip4_address_t *out_r_addr, u16 r_port)
Create new NAT64 session table entry.
u32 fib_table_get_index_for_sw_if_index(fib_protocol_t proto, u32 sw_if_index)
Get the index of the FIB bound to the interface.
nat64_db_bib_entry_t * nat64_db_bib_entry_create(nat64_db_t *db, ip6_address_t *in_addr, ip4_address_t *out_addr, u16 in_port, u16 out_port, u32 fib_index, u8 proto, u8 is_static)
Create new NAT64 BIB entry.
void nat_ip6_reass_get_frags(nat_reass_ip6_t *reass, u32 **bi)
Get cached fragments.
nat64_db_bib_entry_t * nat64_db_bib_entry_find(nat64_db_t *db, ip46_address_t *addr, u16 port, u8 proto, u32 fib_index, u8 is_ip6)
Find NAT64 BIB entry.
static ip_csum_t ip_csum_with_carry(ip_csum_t sum, ip_csum_t x)
static u8 * format_nat64_in2out_trace(u8 *s, va_list *args)
nat64_db_st_entry_t * nat64_db_st_entry_by_index(nat64_db_t *db, u8 proto, u32 ste_index)
Get ST entry by index and protocol.
u32 buffer_index[VLIB_FRAME_SIZE]
vlib_error_t * errors
Vector of errors for this node.
vlib_node_registration_t nat64_in2out_handoff_node
(constructor) VLIB_REGISTER_NODE (nat64_in2out_handoff_node)
struct _tcp_header tcp_header_t
static vlib_frame_queue_t * is_vlib_frame_queue_congested(u32 frame_queue_index, u32 index, u32 queue_hi_thresh, vlib_frame_queue_t **handoff_queue_by_worker_index)
#define vec_reset_length(v)
Reset vector length to zero NULL-pointer tolerant.
static int nat64_in2out_tcp_udp_hairpinning(vlib_main_t *vm, vlib_buffer_t *b, ip6_header_t *ip6, u32 thread_index)
u32 nat64_get_worker_in2out(ip6_address_t *addr)
Get worker thread index for NAT64 in2out.
#define static_always_inline
vlib_node_registration_t nat64_in2out_reass_node
(constructor) VLIB_REGISTER_NODE (nat64_in2out_reass_node)
vlib_frame_t * vlib_get_frame_to_node(vlib_main_t *vm, u32 to_node_index)
static_always_inline int ip6_parse(const ip6_header_t *ip6, u32 buff_len, u8 *l4_protocol, u16 *l4_offset, u16 *frag_hdr_offset)
Parse some useful information from IPv6 header.
struct unk_proto_st_walk_ctx_t_ unk_proto_st_walk_ctx_t
static int ip6_to_ip4_tcp_udp(vlib_buffer_t *p, ip6_to_ip4_set_fn_t fn, void *ctx, u8 udp_checksum)
Translate IPv6 UDP/TCP packet to IPv4.
u16 current_length
Nbytes between current data and the end of this buffer.
void nat64_session_reset_timeout(nat64_db_st_entry_t *ste, vlib_main_t *vm)
Reset NAT64 session timeout.
static uword nat64_in2out_slowpath_node_fn(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame)
int nat64_alloc_out_addr_and_port(u32 fib_index, snat_protocol_t proto, ip4_address_t *addr, u16 *port, u32 thread_index)
Alloce IPv4 address and port pair from NAT64 pool.
static u8 * format_nat64_in2out_reass_trace(u8 *s, va_list *args)
void vlib_put_frame_to_node(vlib_main_t *vm, u32 to_node_index, vlib_frame_t *f)
static void * vlib_buffer_get_current(vlib_buffer_t *b)
Get pointer to current data to process.
static vlib_frame_queue_elt_t * vlib_get_worker_handoff_queue_elt(u32 frame_queue_index, u32 vlib_worker_index, vlib_frame_queue_elt_t **handoff_queue_elt_by_worker_index)
static int nat64_in2out_inner_icmp_set_cb(ip6_header_t *ip6, ip4_header_t *ip4, void *arg)
static char * nat64_in2out_error_strings[]
#define vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, bi0, next0)
Finish enqueueing one buffer forward in the graph.
#define vlib_get_next_frame(vm, node, next_index, vectors, n_vectors_left)
Get pointer to next frame vector data by (vlib_node_runtime_t, next_index).
void nat64_compose_ip6(ip6_address_t *ip6, ip4_address_t *ip4, u32 fib_index)
Compose IPv4-embedded IPv6 addresses.
vlib_error_t error
Error code for buffers to be enqueued to error handler.
static void vlib_node_increment_counter(vlib_main_t *vm, u32 node_index, u32 counter_index, u64 increment)
static u8 * format_nat64_in2out_handoff_trace(u8 *s, va_list *args)
#define VLIB_REGISTER_NODE(x,...)
static int nat64_in2out_unk_proto_hairpinning(vlib_main_t *vm, vlib_buffer_t *b, ip6_header_t *ip6, u32 thread_index)
static int nat64_in2out_icmp_set_cb(ip6_header_t *ip6, ip4_header_t *ip4, void *arg)
static_always_inline uword vlib_get_thread_index(void)
u8 nat_reass_is_drop_frag(u8 is_ip6)
Get status of virtual fragmentation reassembly.
static int nat64_in2out_tcp_udp_set_cb(ip6_header_t *ip6, ip4_header_t *ip4, void *arg)
#define vec_free(V)
Free vector's memory (no header).
static ip_csum_t ip_incremental_checksum(ip_csum_t sum, void *_data, uword n_bytes)
vlib_node_registration_t nat64_in2out_slowpath_node
(constructor) VLIB_REGISTER_NODE (nat64_in2out_slowpath_node)
#define clib_memcpy(a, b, c)
void vlib_put_next_frame(vlib_main_t *vm, vlib_node_runtime_t *r, u32 next_index, u32 n_vectors_left)
Release pointer to next frame vector data.
static void * ip6_next_header(ip6_header_t *i)
#define ip6_frag_hdr_offset(hdr)
u16 cached_next_index
Next frame index that vector arguments were last enqueued to last time this node ran.
nat64_db_st_entry_t * nat64_db_st_entry_find(nat64_db_t *db, ip46_address_t *l_addr, ip46_address_t *r_addr, u16 l_port, u16 r_port, u8 proto, u32 fib_index, u8 is_ip6)
Find NAT64 session table entry.
void nat64_db_st_walk(nat64_db_t *db, u8 proto, nat64_db_st_walk_fn_t fn, void *ctx)
Walk NAT64 session table.
static int nat64_in2out_icmp_hairpinning(vlib_main_t *vm, vlib_buffer_t *b, ip6_header_t *ip6, u32 thread_index)
u32 fq_in2out_index
Worker handoff.
u32 nat64_db_st_entry_get_index(nat64_db_t *db, nat64_db_st_entry_t *ste)
static ip_csum_t ip_csum_sub_even(ip_csum_t c, ip_csum_t x)
nat64_db_bib_entry_t * nat64_db_bib_entry_by_index(nat64_db_t *db, u8 proto, u32 bibe_index)
Get BIB entry by index and protocol.
IPv6 to IPv4 translation.
static uword nat64_in2out_reass_node_fn(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame)
#define u8_ptr_add(ptr, index)
static void * vlib_add_trace(vlib_main_t *vm, vlib_node_runtime_t *r, vlib_buffer_t *b, u32 n_data_bytes)
struct _vlib_node_registration vlib_node_registration_t
NAT64 global declarations.
static u32 ip_proto_to_snat_proto(u8 ip_proto)
The NAT inline functions.
VLIB_NODE_FUNCTION_MULTIARCH(nat64_in2out_node, nat64_in2out_node_fn)
#define ip46_address_is_equal(a1, a2)
#define vec_len(v)
Number of elements in vector (rvalue-only, NULL tolerant)
nat_reass_ip6_t * nat_ip6_reass_find_or_create(ip6_address_t src, ip6_address_t dst, u32 frag_id, u8 proto, u8 reset_timeout, u32 **bi_to_drop)
Find or create reassembly.
static void * vlib_frame_vector_args(vlib_frame_t *f)
Get pointer to frame vector data.
static uword nat64_in2out_handoff_node_fn(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame)
static void vlib_put_frame_queue_elt(vlib_frame_queue_elt_t *hf)
static uword nat64_in2out_node_fn_inline(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame, u8 is_slow_path)
static int nat64_in2out_unk_proto_set_cb(ip6_header_t *ip6, ip4_header_t *ip4, void *arg)
static vlib_thread_main_t * vlib_get_thread_main()
static_always_inline int is_hairpinning(ip6_address_t *dst_addr)
Check whether is a hairpinning.
#define vec_foreach(var, vec)
Vector iterator.
static uword nat64_in2out_node_fn(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame)
u16 flags
Copy of main node flags.
static void nat_send_all_to_node(vlib_main_t *vm, u32 *bi_vector, vlib_node_runtime_t *node, vlib_error_t *error, u32 next)
NAT plugin virtual fragmentation reassembly.
#define VLIB_NODE_FLAG_TRACE
#define foreach_nat64_in2out_error
#define vec_validate_init_empty(V, I, INIT)
Make sure vector is long enough for given index and initialize empty space (no header, unspecified alignment)
struct nat64_in2out_set_ctx_t_ nat64_in2out_set_ctx_t
u32 flags
buffer flags: VLIB_BUFFER_FREE_LIST_INDEX_MASK: bits used to store free list index, VLIB_BUFFER_IS_TRACED: trace this buffer.
static vlib_buffer_t * vlib_get_buffer(vlib_main_t *vm, u32 buffer_index)
Translate buffer index into buffer pointer.
static int ip6_to_ip4(vlib_buffer_t *p, ip6_to_ip4_set_fn_t fn, void *ctx)
Translate IPv6 packet to IPv4 (IP header only).
static u16 ip_csum_fold(ip_csum_t c)
struct nat64_in2out_frag_set_ctx_t_ nat64_in2out_frag_set_ctx_t
static ip_csum_t ip_csum_add_even(ip_csum_t c, ip_csum_t x)
static int ip6_to_ip4_fragmented(vlib_buffer_t *p, ip6_to_ip4_set_fn_t fn, void *ctx)
Translate IPv6 fragmented packet to IPv4.