|
static_always_inline void | aes_gcm_enc_first_round (u8x16 *r, aes_gcm_counter_t *ctr, u8x16 k, int n_blocks) |
|
static_always_inline void | aes_gcm_enc_round (u8x16 *r, u8x16 k, int n_blocks) |
|
static_always_inline void | aes_gcm_enc_last_round (u8x16 *r, u8x16 *d, u8x16 const *k, int rounds, int n_blocks) |
|
static_always_inline u8x16 | aes_gcm_ghash_blocks (u8x16 T, aes_gcm_key_data_t *kd, u8x16u *in, int n_blocks) |
|
static_always_inline u8x16 | aes_gcm_ghash (u8x16 T, aes_gcm_key_data_t *kd, u8x16u *in, u32 n_left) |
|
static_always_inline u8x16 | aes_gcm_calc (u8x16 T, aes_gcm_key_data_t *kd, u8x16 *d, aes_gcm_counter_t *ctr, u8x16u *inv, u8x16u *outv, int rounds, int n, int last_block_bytes, aes_gcm_flags_t f) |
|
static_always_inline u8x16 | aes_gcm_calc_double (u8x16 T, aes_gcm_key_data_t *kd, u8x16 *d, aes_gcm_counter_t *ctr, u8x16u *inv, u8x16u *outv, int rounds, aes_gcm_flags_t f) |
|
static_always_inline u8x16 | aes_gcm_ghash_last (u8x16 T, aes_gcm_key_data_t *kd, u8x16 *d, int n_blocks, int n_bytes) |
|
static_always_inline u8x16 | aes_gcm_enc (u8x16 T, aes_gcm_key_data_t *kd, aes_gcm_counter_t *ctr, u8x16u *inv, u8x16u *outv, u32 n_left, int rounds) |
|
static_always_inline u8x16 | aes_gcm_dec (u8x16 T, aes_gcm_key_data_t *kd, aes_gcm_counter_t *ctr, u8x16u *inv, u8x16u *outv, u32 n_left, int rounds) |
|
static_always_inline int | aes_gcm (u8x16u *in, u8x16u *out, u8x16u *addt, u8x16u *iv, u8x16u *tag, u32 data_bytes, u32 aad_bytes, u8 tag_len, aes_gcm_key_data_t *kd, int aes_rounds, int is_encrypt) |
|
static_always_inline u32 | aes_ops_enc_aes_gcm (vlib_main_t *vm, vnet_crypto_op_t *ops[], u32 n_ops, aes_key_size_t ks) |
|
static_always_inline u32 | aes_ops_dec_aes_gcm (vlib_main_t *vm, vnet_crypto_op_t *ops[], u32 n_ops, aes_key_size_t ks) |
|
static_always_inline void * | aes_gcm_key_exp (vnet_crypto_key_t *key, aes_key_size_t ks) |
|
clib_error_t * | crypto_native_aes_gcm_init_slm (vlib_main_t *vm) |
|