FD.io VPP  v19.08.3-2-gbabecb413
Vector Packet Processing
nat44_handoff.c
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2018 Cisco and/or its affiliates.
3  * Licensed under the Apache License, Version 2.0 (the "License");
4  * you may not use this file except in compliance with the License.
5  * You may obtain a copy of the License at:
6  *
7  * http://www.apache.org/licenses/LICENSE-2.0
8  *
9  * Unless required by applicable law or agreed to in writing, software
10  * distributed under the License is distributed on an "AS IS" BASIS,
11  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12  * See the License for the specific language governing permissions and
13  * limitations under the License.
14  */
15 /**
16  * @file
17  * @brief NAT44 worker handoff
18  */
19 
20 #include <vlib/vlib.h>
21 #include <vnet/vnet.h>
22 #include <vnet/handoff.h>
23 #include <vnet/fib/ip4_fib.h>
24 #include <vppinfra/error.h>
25 #include <nat/nat.h>
26 #include <nat/nat_inlines.h>
27 
28 typedef struct
29 {
35 
36 #define foreach_nat44_handoff_error \
37 _(CONGESTION_DROP, "congestion drop") \
38 _(SAME_WORKER, "same worker") \
39 _(DO_HANDOFF, "do handoff")
40 
41 typedef enum
42 {
43 #define _(sym,str) NAT44_HANDOFF_ERROR_##sym,
45 #undef _
48 
49 static char *nat44_handoff_error_strings[] = {
50 #define _(sym,string) string,
52 #undef _
53 };
54 
55 
56 static u8 *
57 format_nat44_handoff_trace (u8 * s, va_list * args)
58 {
59  CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
60  CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
61  nat44_handoff_trace_t *t = va_arg (*args, nat44_handoff_trace_t *);
62  char *tag, *output;
63 
64  tag = t->in2out ? "IN2OUT" : "OUT2IN";
65  output = t->output ? "OUTPUT-FEATURE" : "";
66  s =
67  format (s, "NAT44_%s_WORKER_HANDOFF %s: next-worker %d trace index %d",
68  tag, output, t->next_worker_index, t->trace_index);
69 
70  return s;
71 }
72 
73 static inline uword
75  vlib_node_runtime_t * node,
76  vlib_frame_t * frame, u8 is_output,
77  u8 is_in2out)
78 {
79  u32 n_enq, n_left_from, *from, do_handoff = 0, same_worker = 0;
80 
81  u16 thread_indices[VLIB_FRAME_SIZE], *ti = thread_indices;
82  vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs;
83  snat_main_t *sm = &snat_main;
84 
85  snat_get_worker_function_t *get_worker;
86  u32 fq_index, thread_index = vm->thread_index;
87 
88  from = vlib_frame_vector_args (frame);
89  n_left_from = frame->n_vectors;
90 
91  vlib_get_buffers (vm, from, b, n_left_from);
92 
93  if (is_in2out)
94  {
95  fq_index = is_output ? sm->fq_in2out_output_index : sm->fq_in2out_index;
96  get_worker = sm->worker_in2out_cb;
97  }
98  else
99  {
100  fq_index = sm->fq_out2in_index;
101  get_worker = sm->worker_out2in_cb;
102  }
103 
104  while (n_left_from >= 4)
105  {
106  u32 sw_if_index0, sw_if_index1, sw_if_index2, sw_if_index3;
107  u32 rx_fib_index0, rx_fib_index1, rx_fib_index2, rx_fib_index3;
108  u32 iph_offset0 = 0, iph_offset1 = 0, iph_offset2 = 0, iph_offset3 = 0;
109  ip4_header_t *ip0, *ip1, *ip2, *ip3;
110 
111  if (PREDICT_TRUE (n_left_from >= 8))
112  {
113  vlib_prefetch_buffer_header (b[4], STORE);
114  vlib_prefetch_buffer_header (b[5], STORE);
115  vlib_prefetch_buffer_header (b[6], STORE);
116  vlib_prefetch_buffer_header (b[7], STORE);
117  CLIB_PREFETCH (&b[4]->data, CLIB_CACHE_LINE_BYTES, STORE);
118  CLIB_PREFETCH (&b[5]->data, CLIB_CACHE_LINE_BYTES, STORE);
119  CLIB_PREFETCH (&b[6]->data, CLIB_CACHE_LINE_BYTES, STORE);
120  CLIB_PREFETCH (&b[7]->data, CLIB_CACHE_LINE_BYTES, STORE);
121  }
122 
123  if (is_output)
124  {
125  iph_offset0 = vnet_buffer (b[0])->ip.save_rewrite_length;
126  iph_offset1 = vnet_buffer (b[1])->ip.save_rewrite_length;
127  iph_offset2 = vnet_buffer (b[2])->ip.save_rewrite_length;
128  iph_offset3 = vnet_buffer (b[3])->ip.save_rewrite_length;
129  }
130 
131  ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b[0]) +
132  iph_offset0);
133  ip1 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b[1]) +
134  iph_offset1);
135  ip2 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b[2]) +
136  iph_offset2);
137  ip3 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b[3]) +
138  iph_offset3);
139 
140  sw_if_index0 = vnet_buffer (b[0])->sw_if_index[VLIB_RX];
141  sw_if_index1 = vnet_buffer (b[1])->sw_if_index[VLIB_RX];
142  sw_if_index2 = vnet_buffer (b[2])->sw_if_index[VLIB_RX];
143  sw_if_index3 = vnet_buffer (b[3])->sw_if_index[VLIB_RX];
144 
145  rx_fib_index0 = ip4_fib_table_get_index_for_sw_if_index (sw_if_index0);
146  rx_fib_index1 = ip4_fib_table_get_index_for_sw_if_index (sw_if_index1);
147  rx_fib_index2 = ip4_fib_table_get_index_for_sw_if_index (sw_if_index2);
148  rx_fib_index3 = ip4_fib_table_get_index_for_sw_if_index (sw_if_index3);
149 
150  ti[0] = get_worker (ip0, rx_fib_index0, is_output);
151  ti[1] = get_worker (ip1, rx_fib_index1, is_output);
152  ti[2] = get_worker (ip2, rx_fib_index2, is_output);
153  ti[3] = get_worker (ip3, rx_fib_index3, is_output);
154 
155  if (ti[0] == thread_index)
156  same_worker++;
157  else
158  do_handoff++;
159 
160  if (ti[1] == thread_index)
161  same_worker++;
162  else
163  do_handoff++;
164 
165  if (ti[2] == thread_index)
166  same_worker++;
167  else
168  do_handoff++;
169 
170  if (ti[3] == thread_index)
171  same_worker++;
172  else
173  do_handoff++;
174 
175  b += 4;
176  ti += 4;
177  n_left_from -= 4;
178  }
179 
180  while (n_left_from > 0)
181  {
182  u32 sw_if_index0;
183  u32 rx_fib_index0;
184  u32 iph_offset0 = 0;
185  ip4_header_t *ip0;
186 
187 
188  if (is_output)
189  iph_offset0 = vnet_buffer (b[0])->ip.save_rewrite_length;
190 
191  ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b[0]) +
192  iph_offset0);
193 
194  sw_if_index0 = vnet_buffer (b[0])->sw_if_index[VLIB_RX];
195  rx_fib_index0 = ip4_fib_table_get_index_for_sw_if_index (sw_if_index0);
196 
197  ti[0] = get_worker (ip0, rx_fib_index0, is_output);
198 
199  if (ti[0] == thread_index)
200  same_worker++;
201  else
202  do_handoff++;
203 
204  b += 1;
205  ti += 1;
206  n_left_from -= 1;
207  }
208 
209  if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)))
210  {
211  u32 i;
212  b = bufs;
213  ti = thread_indices;
214 
215  for (i = 0; i < frame->n_vectors; i++)
216  {
217  if (b[0]->flags & VLIB_BUFFER_IS_TRACED)
218  {
220  vlib_add_trace (vm, node, b[0], sizeof (*t));
221  t->next_worker_index = ti[0];
223  t->in2out = is_in2out;
224  t->output = is_output;
225 
226  b += 1;
227  ti += 1;
228  }
229  else
230  break;
231  }
232  }
233 
234  n_enq = vlib_buffer_enqueue_to_thread (vm, fq_index, from, thread_indices,
235  frame->n_vectors, 1);
236 
237  if (n_enq < frame->n_vectors)
238  {
240  NAT44_HANDOFF_ERROR_CONGESTION_DROP,
241  frame->n_vectors - n_enq);
242  }
243 
245  NAT44_HANDOFF_ERROR_SAME_WORKER, same_worker);
247  NAT44_HANDOFF_ERROR_DO_HANDOFF, do_handoff);
248  return frame->n_vectors;
249 }
250 
251 
252 
254  vlib_node_runtime_t * node,
255  vlib_frame_t * frame)
256 {
257  return nat44_worker_handoff_fn_inline (vm, node, frame, 0, 1);
258 }
259 
260 /* *INDENT-OFF* */
262  .name = "nat44-in2out-worker-handoff",
263  .vector_size = sizeof (u32),
264  .sibling_of = "nat-default",
265  .format_trace = format_nat44_handoff_trace,
268  .error_strings = nat44_handoff_error_strings,
269 };
270 /* *INDENT-ON* */
271 
274  node,
275  vlib_frame_t * frame)
276 {
277  return nat44_worker_handoff_fn_inline (vm, node, frame, 1, 1);
278 }
279 
280 /* *INDENT-OFF* */
282  .name = "nat44-in2out-output-worker-handoff",
283  .vector_size = sizeof (u32),
284  .sibling_of = "nat-default",
285  .format_trace = format_nat44_handoff_trace,
288  .error_strings = nat44_handoff_error_strings,
289 };
290 /* *INDENT-ON* */
291 
293  vlib_node_runtime_t * node,
294  vlib_frame_t * frame)
295 {
296  return nat44_worker_handoff_fn_inline (vm, node, frame, 0, 0);
297 }
298 
299 /* *INDENT-OFF* */
301  .name = "nat44-out2in-worker-handoff",
302  .vector_size = sizeof (u32),
303  .sibling_of = "nat-default",
304  .format_trace = format_nat44_handoff_trace,
307  .error_strings = nat44_handoff_error_strings,
308 };
309 /* *INDENT-ON* */
310 
311 /*
312  * fd.io coding-style-patch-verification: ON
313  *
314  * Local Variables:
315  * eval: (c-set-style "gnu")
316  * End:
317  */
u32 flags
Definition: vhost_user.h:141
#define CLIB_UNUSED(x)
Definition: clib.h:83
static char * nat44_handoff_error_strings[]
Definition: nat44_handoff.c:49
#define foreach_nat44_handoff_error
Definition: nat44_handoff.c:36
static u32 vlib_buffer_get_trace_index(vlib_buffer_t *b)
Extract the trace (pool) index from a trace handle.
Definition: buffer.h:389
u32 fq_in2out_output_index
Definition: nat.h:611
#define PREDICT_TRUE(x)
Definition: clib.h:113
u32 thread_index
Definition: main.h:218
int i
vlib_node_registration_t snat_in2out_worker_handoff_node
(constructor) VLIB_REGISTER_NODE (snat_in2out_worker_handoff_node)
u8 * format(u8 *s, const char *fmt,...)
Definition: format.c:424
u8 data[128]
Definition: ipsec.api:251
#define VLIB_NODE_FN(node)
Definition: node.h:202
unsigned char u8
Definition: types.h:56
u32 ip4_fib_table_get_index_for_sw_if_index(u32 sw_if_index)
Definition: ip4_fib.c:224
#define vlib_prefetch_buffer_header(b, type)
Prefetch buffer metadata.
Definition: buffer.h:203
unsigned int u32
Definition: types.h:88
#define VLIB_FRAME_SIZE
Definition: node.h:378
vl_api_fib_path_type_t type
Definition: fib_types.api:123
unsigned short u16
Definition: types.h:57
nat44_handoff_error_t
Definition: nat44_handoff.c:41
static void * vlib_buffer_get_current(vlib_buffer_t *b)
Get pointer to current data to process.
Definition: buffer.h:229
#define PREDICT_FALSE(x)
Definition: clib.h:112
u32 node_index
Node index.
Definition: node.h:496
static void vlib_node_increment_counter(vlib_main_t *vm, u32 node_index, u32 counter_index, u64 increment)
Definition: node_funcs.h:1150
u32 fq_out2in_index
Definition: nat.h:612
snat_main_t snat_main
Definition: nat.c:39
#define VLIB_REGISTER_NODE(x,...)
Definition: node.h:169
u16 n_vectors
Definition: node.h:397
#define CLIB_PREFETCH(addr, size, type)
Definition: cache.h:80
vlib_main_t * vm
Definition: buffer.c:323
snat_get_worker_function_t * worker_in2out_cb
Definition: nat.h:558
#define ARRAY_LEN(x)
Definition: clib.h:63
snat_get_worker_function_t * worker_out2in_cb
Definition: nat.h:559
u32 fq_in2out_index
Definition: nat.h:610
u32() snat_get_worker_function_t(ip4_header_t *ip, u32 rx_fib_index, u8 is_output)
Definition: nat.h:536
static uword nat44_worker_handoff_fn_inline(vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame, u8 is_output, u8 is_in2out)
Definition: nat44_handoff.c:74
static void * vlib_add_trace(vlib_main_t *vm, vlib_node_runtime_t *r, vlib_buffer_t *b, u32 n_data_bytes)
Definition: trace_funcs.h:55
VLIB buffer representation.
Definition: buffer.h:102
u64 uword
Definition: types.h:112
static void * vlib_frame_vector_args(vlib_frame_t *f)
Get pointer to frame vector data.
Definition: node_funcs.h:244
vlib_node_registration_t snat_out2in_worker_handoff_node
(constructor) VLIB_REGISTER_NODE (snat_out2in_worker_handoff_node)
static_always_inline u32 vlib_buffer_enqueue_to_thread(vlib_main_t *vm, u32 frame_queue_index, u32 *buffer_indices, u16 *thread_indices, u32 n_packets, int drop_on_congestion)
Definition: buffer_node.h:487
#define vnet_buffer(b)
Definition: buffer.h:365
vlib_node_registration_t snat_in2out_output_worker_handoff_node
(constructor) VLIB_REGISTER_NODE (snat_in2out_output_worker_handoff_node)
u16 flags
Copy of main node flags.
Definition: node.h:509
static u8 * format_nat44_handoff_trace(u8 *s, va_list *args)
Definition: nat44_handoff.c:57
static_always_inline void vlib_get_buffers(vlib_main_t *vm, u32 *bi, vlib_buffer_t **b, int count)
Translate array of buffer indices into buffer pointers.
Definition: buffer_funcs.h:244
#define VLIB_NODE_FLAG_TRACE
Definition: node.h:302
#define CLIB_CACHE_LINE_BYTES
Definition: cache.h:59
Definition: defs.h:46