36 str, xd->
port_id, rv, rte_strerror (rv));
46 struct rte_eth_dev_info dev_info;
56 if (xd->
flags & DPDK_DEVICE_FLAG_ADMIN_UP)
63 if (xd->
pmd == VNET_DPDK_PMD_I40E)
65 if ((xd->
flags & DPDK_DEVICE_FLAG_RX_FLOW_OFFLOAD) != 0)
66 xd->
port_conf.fdir_conf.mode = RTE_FDIR_MODE_PERFECT;
68 xd->
port_conf.fdir_conf.mode = RTE_FDIR_MODE_NONE;
71 rte_eth_dev_info_get (xd->
port_id, &dev_info);
73 bitmap = xd->
port_conf.txmode.offloads & ~dev_info.tx_offload_capa;
76 dpdk_log_warn (
"unsupported tx offloads requested on port %u: %U",
81 bitmap = xd->
port_conf.rxmode.offloads & ~dev_info.rx_offload_capa;
84 dpdk_log_warn (
"unsupported rx offloads requested on port %u: %U",
108 rte_eth_tx_queue_setup (xd->
port_id, j,
123 u16 socket_id = rte_lcore_to_socket_id (lcore);
133 rte_eth_rx_queue_setup (xd->
port_id, j,
149 if (xd->
flags & DPDK_DEVICE_FLAG_ADMIN_UP)
158 xd->
flags |= DPDK_DEVICE_FLAG_PMD_INIT_FAIL;
167 if (xd->
flags & DPDK_DEVICE_FLAG_PMD_INIT_FAIL)
170 rv = rte_eth_dev_start (xd->
port_id);
180 rte_eth_dev_default_mac_addr_set (xd->
port_id,
181 (
struct ether_addr *)
187 if (xd->
flags & DPDK_DEVICE_FLAG_PROMISC)
188 rte_eth_promiscuous_enable (xd->
port_id);
190 rte_eth_promiscuous_disable (xd->
port_id);
192 rte_eth_allmulticast_enable (xd->
port_id);
194 if (xd->
pmd == VNET_DPDK_PMD_BOND)
197 int nlink = rte_eth_bond_slaves_get (xd->
port_id, slink, 16);
201 rte_eth_allmulticast_enable (dpdk_port);
212 if (xd->
flags & DPDK_DEVICE_FLAG_PMD_INIT_FAIL)
215 rte_eth_allmulticast_disable (xd->
port_id);
216 rte_eth_dev_stop (xd->
port_id);
220 if (xd->
pmd == VNET_DPDK_PMD_BOND)
223 int nlink = rte_eth_bond_slaves_get (xd->
port_id, slink, 16);
227 rte_eth_dev_stop (dpdk_port);
246 uword event_type, *event_data = 0;
255 for (i = 0; i <
vec_len (event_data); i++)
257 dpdk_port = event_data[
i];
274 .name =
"send-garp-na-process",
286 (vm, send_garp_na_proc_node.index,
SEND_GARP_NA, *dpdk_port);
291 enum rte_eth_event_type type,
void *param)
293 struct rte_eth_link link;
296 RTE_SET_USED (param);
297 if (type != RTE_ETH_EVENT_INTR_LSC)
299 dpdk_log_info (
"Unknown event %d received for port %d", type, port_id);
303 rte_eth_link_get_nowait (port_id, &link);
304 u8 link_up = link.link_status;
306 if (xd->flags & DPDK_DEVICE_FLAG_BOND_SLAVE)
308 uword bd_port = xd->bond_port;
309 int bd_mode = rte_eth_bond_mode_get (bd_port);
311 "slave of port %d BondEthernet%d in mode %d",
312 port_id, (link_up) ?
"UP" :
"DOWN",
313 bd_port, xd->bond_instance_num, bd_mode);
314 if (bd_mode == BONDING_MODE_ACTIVE_BACKUP)
321 xd->flags |= DPDK_DEVICE_FLAG_BOND_SLAVE_UP;
323 xd->flags &= ~DPDK_DEVICE_FLAG_BOND_SLAVE_UP;
329 port_id, (
unsigned) link.link_speed,
330 (link.link_duplex == ETH_LINK_FULL_DUPLEX) ?
331 "full-duplex" :
"half-duplex");
341 enum rte_eth_event_type type,
343 void *ret_param __attribute__ ((unused)))
349 struct rte_pci_device *
352 const struct rte_bus *bus;
354 bus = rte_bus_find_by_device (info->device);
355 if (bus && !strcmp (bus->name,
"pci"))
356 return RTE_DEV_TO_PCI (info->device);
format_function_t format_dpdk_tx_offload_caps
void vl_api_force_rpc_call_main_thread(void *fp, u8 *data, u32 data_length)
static uword * vlib_process_wait_for_event(vlib_main_t *vm)
vnet_main_t * vnet_get_main(void)
clib_memset(h->entries, 0, sizeof(h->entries[0]) *entries)
static f64 vlib_time_now(vlib_main_t *vm)
struct rte_pci_device * dpdk_get_pci_device(const struct rte_eth_dev_info *info)
static vnet_hw_interface_t * vnet_get_hw_interface(vnet_main_t *vnm, u32 hw_if_index)
static vnet_sw_interface_t * vnet_get_sw_interface(vnet_main_t *vnm, u32 sw_if_index)
int dpdk_port_state_callback(dpdk_portid_t port_id, enum rte_eth_event_type type, void *param, void *ret_param)
#define vec_validate_aligned(V, I, A)
Make sure vector is long enough for given index (no header, specified alignment)
void send_ip6_na(vlib_main_t *vm, u32 sw_if_index)
#define vec_reset_length(v)
Reset vector length to zero NULL-pointer tolerant.
static uword vlib_process_suspend(vlib_main_t *vm, f64 dt)
Suspend a vlib cooperative multi-tasking thread for a period of time.
#define dpdk_log_warn(...)
static uword vlib_process_get_events(vlib_main_t *vm, uword **data_vector)
Return the first event type which has occurred and a vector of per-event data of that type...
#define clib_error_return(e, args...)
void dpdk_device_setup(dpdk_device_t *xd)
struct rte_eth_conf port_conf
static uword send_garp_na_process(vlib_main_t *vm, vlib_node_runtime_t *rt, vlib_frame_t *f)
struct rte_eth_txconf tx_conf
vlib_worker_thread_t * vlib_worker_threads
void dpdk_device_start(dpdk_device_t *xd)
static_always_inline uword vnet_get_device_input_thread_index(vnet_main_t *vnm, u32 hw_if_index, u16 queue_id)
static void vlib_process_signal_event(vlib_main_t *vm, uword node_index, uword type_opaque, uword data)
vnet_sw_interface_flags_t flags
static void garp_na_proc_callback(uword *dpdk_port)
#define dpdk_log_info(...)
void send_ip4_garp(vlib_main_t *vm, u32 sw_if_index)
#define VLIB_REGISTER_NODE(x,...)
static_always_inline uword vlib_get_thread_index(void)
#define dpdk_log_err(...)
static vlib_node_registration_t send_garp_na_proc_node
(constructor) VLIB_REGISTER_NODE (send_garp_na_proc_node)
static int dpdk_port_state_callback_inline(dpdk_portid_t port_id, enum rte_eth_event_type type, void *param)
void dpdk_device_stop(dpdk_device_t *xd)
void dpdk_device_error(dpdk_device_t *xd, char *str, int rv)
format_function_t format_dpdk_device_name
struct rte_mempool ** pktmbuf_pools
void dpdk_update_link_state(dpdk_device_t *xd, f64 now)
static vlib_main_t * vlib_get_main(void)
struct _vlib_node_registration vlib_node_registration_t
enum @419 dpdk_send_garp_na_process_event_t
#define vec_len(v)
Number of elements in vector (rvalue-only, NULL tolerant)
clib_error_t * vnet_hw_interface_set_flags(vnet_main_t *vnm, u32 hw_if_index, vnet_hw_interface_flags_t flags)
format_function_t format_dpdk_rx_offload_caps
#define clib_error_free(e)
#define CLIB_CACHE_LINE_BYTES
u8 * buffer_pool_for_queue
CLIB vectors are ubiquitous dynamically resized arrays with by user defined "headers".