2 // Copyright (c) 2010-2017 Intel Corporation
4 // Licensed under the Apache License, Version 2.0 (the "License");
5 // you may not use this file except in compliance with the License.
6 // You may obtain a copy of the License at
8 // http://www.apache.org/licenses/LICENSE-2.0
10 // Unless required by applicable law or agreed to in writing, software
11 // distributed under the License is distributed on an "AS IS" BASIS,
12 // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 // See the License for the specific language governing permissions and
14 // limitations under the License.
18 * Non compatible implementation of RFC3686(CTR-AES 128 bit key), RFC4303 (tunnel ipv4 ESP)
20 * 1. Crypto not safe!!!!! (underlying AES-CTR implementation is OK, but ESP implementation is lousy)
21 * 2. Only ESP/tunnel/ipv4/AES-CTR
22 * 3. Not fully implemented
23 * 4. No proper key / SADB
24 * So performance demonstrator only
27 #include "task_init.h"
28 #include "task_base.h"
33 #include "prox_cksum.h"
36 #include <rte_cryptodev.h>
37 #include <rte_bus_vdev.h>
38 #include "prox_port_cfg.h"
39 #include "prox_compat.h"
41 typedef unsigned int u32;
42 typedef unsigned char u8;
44 #define BYTE_LENGTH(x) (x/8)
45 #define DIGEST_BYTE_LENGTH_SHA1 (BYTE_LENGTH(160))
47 //#define CIPHER_KEY_LENGTH_AES_CBC (32)
48 #define CIPHER_KEY_LENGTH_AES_CBC (16)//==TEST
49 #define CIPHER_IV_LENGTH_AES_CBC 16
51 #define MAXIMUM_IV_LENGTH 16
52 #define IV_OFFSET (sizeof(struct rte_crypto_op) + sizeof(struct rte_crypto_sym_op))
54 #define MAX_SESSIONS 1024
55 #define POOL_CACHE_SIZE 128
60 struct task_base base;
64 prox_rte_ether_addr local_mac;
66 prox_rte_ether_addr dst_mac;
67 struct rte_mempool *crypto_op_pool;
68 struct rte_mempool *session_pool;
69 struct rte_cryptodev_sym_session *sess;
70 struct rte_crypto_op *ops_burst[NUM_OPS];
72 uint8_t (*handle_esp_finish)(struct task_esp *task,
73 struct rte_mbuf *mbuf, uint8_t status);
74 uint8_t (*handle_esp_ah)(struct task_esp *task, struct rte_mbuf *mbuf,
75 struct rte_crypto_op *cop);
78 static uint8_t hmac_sha1_key[] = {
79 0xF8, 0x2A, 0xC7, 0x54, 0xDB, 0x96, 0x18, 0xAA,
80 0xC3, 0xA1, 0x53, 0xF6, 0x1F, 0x17, 0x60, 0xBD,
81 0xDE, 0xF4, 0xDE, 0xAD };
83 static uint8_t aes_cbc_key[] = {
84 0xE4, 0x23, 0x33, 0x8A, 0x35, 0x64, 0x61, 0xE2,
85 0x49, 0x03, 0xDD, 0xC6, 0xB8, 0xCA, 0x55, 0x7A,
86 0xE4, 0x23, 0x33, 0x8A, 0x35, 0x64, 0x61, 0xE2,
87 0x49, 0x03, 0xDD, 0xC6, 0xB8, 0xCA, 0x55, 0x7A };
89 static uint8_t aes_cbc_iv[] = {
90 0xE4, 0x23, 0x33, 0x8A, 0x35, 0x64, 0x61, 0xE2,
91 0x49, 0x03, 0xDD, 0xC6, 0xB8, 0xCA, 0x55, 0x7A };
93 static void printf_cdev_info(uint8_t cdev_id)
95 struct rte_cryptodev_info dev_info;
96 rte_cryptodev_info_get(cdev_id, &dev_info);
97 plog_info("!!!numdevs:%d\n", rte_cryptodev_count());
98 //uint16_t rte_cryptodev_queue_pair_count(uint8_t dev_id);
99 plog_info("dev:%d name:%s nb_queue_pairs:%d max_nb_sessions:%d\n",
100 cdev_id, dev_info.driver_name, dev_info.max_nb_queue_pairs, dev_info.sym.max_nb_sessions);
101 const struct rte_cryptodev_capabilities *cap = &dev_info.capabilities[0];
103 while (cap->op != RTE_CRYPTO_OP_TYPE_UNDEFINED) {
104 //plog_info("cap->sym.xform_type:%d,");
105 if (cap->sym.xform_type == RTE_CRYPTO_SYM_XFORM_CIPHER)
106 plog_info("RTE_CRYPTO_SYM_XFORM_CIPHER: %d\n", cap->sym.cipher.algo);
107 cap = &dev_info.capabilities[++i];
111 static uint8_t get_cdev_id(void)
113 static uint8_t last_unused_cdev_id=0;
115 uint8_t cdev_count, cdev_id;
117 cdev_count = rte_cryptodev_count();
118 plog_info("crypto dev count: %d \n", cdev_count);
119 for (cdev_id = last_unused_cdev_id; cdev_id < cdev_count; cdev_id++) {
121 printf_cdev_info(cdev_id);
122 last_unused_cdev_id = cdev_id + 1;
126 sprintf(name, "crypto_aesni_mb%d", cdev_count);
128 #if RTE_VERSION < RTE_VERSION_NUM(18,8,0,0)
129 int ret = rte_vdev_init(name, "max_nb_queue_pairs=8,max_nb_sessions=1024,socket_id=0");
131 int ret = rte_vdev_init(name, "max_nb_queue_pairs=8,socket_id=0");
133 PROX_PANIC(ret != 0, "Failed rte_vdev_init\n");
134 cdev_id = rte_cryptodev_get_dev_id(name);
136 printf_cdev_info(cdev_id);
137 last_unused_cdev_id = cdev_id + 1;
141 static inline uint8_t handle_enc_finish(struct task_esp *task,
142 struct rte_mbuf *mbuf, uint8_t status)
144 prox_rte_ether_hdr *peth = rte_pktmbuf_mtod(mbuf,
145 prox_rte_ether_hdr *);
146 prox_rte_ipv4_hdr* pip4 = (prox_rte_ipv4_hdr *)(peth + 1);
147 pip4->dst_addr = task->remote_ipv4;
148 pip4->src_addr = task->local_ipv4;
149 prox_ip_cksum(mbuf, pip4, sizeof(prox_rte_ether_hdr),
150 sizeof(prox_rte_ipv4_hdr), 1);
154 static inline uint8_t handle_dec_finish(struct task_esp *task,
155 struct rte_mbuf *mbuf, uint8_t status)
157 if (likely(status == RTE_CRYPTO_OP_STATUS_SUCCESS)) {
158 u8* m = rte_pktmbuf_mtod(mbuf, u8*);
159 rte_memcpy(m + sizeof(prox_rte_ipv4_hdr) +
160 sizeof(struct prox_esp_hdr) +
161 CIPHER_IV_LENGTH_AES_CBC, m,
162 sizeof(prox_rte_ether_hdr));
163 m = (u8*)rte_pktmbuf_adj(mbuf, sizeof(prox_rte_ipv4_hdr) +
164 sizeof(struct prox_esp_hdr) +
165 CIPHER_IV_LENGTH_AES_CBC);
166 prox_rte_ipv4_hdr* pip4 = (prox_rte_ipv4_hdr *)(m +
167 sizeof(prox_rte_ether_hdr));
169 if (unlikely((pip4->version_ihl >> 4) != 4)) {
170 // plog_info("non IPv4 packet after esp dec %i\n",
171 // pip4->version_ihl);
172 // plogdx_info(mbuf, "DEC TX: ");
175 if (pip4->time_to_live) {
176 pip4->time_to_live--;
179 plog_info("TTL = 0 => Dropping\n");
182 uint16_t ipv4_length = rte_be_to_cpu_16(pip4->total_length);
183 int len = rte_pktmbuf_pkt_len(mbuf);
184 rte_pktmbuf_trim(mbuf, len - sizeof(prox_rte_ether_hdr) -
188 do_ipv4_swap(task, mbuf);
190 prox_rte_ether_hdr *peth = rte_pktmbuf_mtod(mbuf,
191 prox_rte_ether_hdr *);
192 prox_rte_ether_addr_copy(&task->local_mac, &peth->s_addr);
193 prox_rte_ether_addr_copy(&task->dst_mac, &peth->d_addr);
194 //rte_memcpy(peth, task->dst_mac, sizeof(task->dst_mac));
196 pip4->dst_addr = task->remote_ipv4;
197 pip4->src_addr = task->local_ipv4;
198 prox_ip_cksum(mbuf, pip4, sizeof(prox_rte_ether_hdr),
199 sizeof(prox_rte_ipv4_hdr), 1);
207 static inline uint8_t handle_esp_ah_enc(struct task_esp *task,
208 struct rte_mbuf *mbuf, struct rte_crypto_op *cop)
211 prox_rte_ether_hdr *peth = rte_pktmbuf_mtod(mbuf,
212 prox_rte_ether_hdr *);
213 prox_rte_ipv4_hdr* pip4 = (prox_rte_ipv4_hdr *)(peth + 1);
214 uint16_t ipv4_length = rte_be_to_cpu_16(pip4->total_length);
215 struct rte_crypto_sym_op *sym_cop = cop->sym;
217 if (unlikely((pip4->version_ihl >> 4) != 4)) {
218 plog_info("Received non IPv4 packet at esp enc %i\n",
222 if (pip4->time_to_live) {
223 pip4->time_to_live--;
226 plog_info("TTL = 0 => Dropping\n");
230 // Remove padding if any (we don't want to encapsulate garbage at end of IPv4 packet)
231 int l1 = rte_pktmbuf_pkt_len(mbuf);
232 int padding = l1 - (ipv4_length + sizeof(prox_rte_ether_hdr));
233 if (unlikely(padding > 0)) {
234 rte_pktmbuf_trim(mbuf, padding);
237 l1 = rte_pktmbuf_pkt_len(mbuf);
238 int encrypt_len = l1 - sizeof(prox_rte_ether_hdr) + 2; // According to RFC4303 table 1, encrypt len is ip+tfc_pad(o)+pad+pad len(1) + next header(1)
240 if ((encrypt_len & 0xf) != 0){
241 padding = 16 - (encrypt_len % 16);
242 encrypt_len += padding;
245 const int extra_space = sizeof(prox_rte_ipv4_hdr) +
246 sizeof(struct prox_esp_hdr) + CIPHER_IV_LENGTH_AES_CBC;
248 prox_rte_ether_addr src_mac = peth->s_addr;
249 prox_rte_ether_addr dst_mac = peth->d_addr;
250 uint32_t src_addr = pip4->src_addr;
251 uint32_t dst_addr = pip4->dst_addr;
252 uint8_t ttl = pip4->time_to_live;
253 uint8_t version_ihl = pip4->version_ihl;
255 peth = (prox_rte_ether_hdr *)rte_pktmbuf_prepend(mbuf, extra_space); // encap + prefix
256 peth = (prox_rte_ether_hdr *)rte_pktmbuf_append(mbuf, 0 + 1 + 1 +
257 padding + 4 + DIGEST_BYTE_LENGTH_SHA1); // padding + pad_len + next_head + seqn + ICV pad + ICV
258 peth = rte_pktmbuf_mtod(mbuf, prox_rte_ether_hdr *);
259 l1 = rte_pktmbuf_pkt_len(mbuf);
260 peth->ether_type = ETYPE_IPv4;
263 prox_rte_ether_addr_copy(&dst_mac, &peth->s_addr);
264 prox_rte_ether_addr_copy(&src_mac, &peth->d_addr);
266 prox_rte_ether_addr_copy(&task->local_mac, &peth->s_addr);
267 //prox_rte_ether_addr_copy(&dst_mac, &peth->d_addr);//IS: dstmac should be rewritten by arp
268 prox_rte_ether_addr_copy(&task->dst_mac, &peth->d_addr);
271 pip4 = (prox_rte_ipv4_hdr *)(peth + 1);
272 pip4->src_addr = task->local_ipv4;
273 pip4->dst_addr = task->remote_ipv4;
274 pip4->time_to_live = ttl;
275 pip4->next_proto_id = IPPROTO_ESP; // 50 for ESP, ip in ip next proto trailer
276 pip4->version_ihl = version_ihl; // 20 bytes, ipv4
277 pip4->total_length = rte_cpu_to_be_16(ipv4_length +
278 sizeof(prox_rte_ipv4_hdr) + sizeof(struct prox_esp_hdr)
279 + CIPHER_IV_LENGTH_AES_CBC + padding + 1 + 1 +
280 DIGEST_BYTE_LENGTH_SHA1); // iphdr+SPI+SN+IV+payload+padding+padlen+next header + crc + auth
281 pip4->packet_id = 0x0101;
282 pip4->type_of_service = 0;
283 pip4->time_to_live = 64;
284 prox_ip_cksum(mbuf, pip4, sizeof(prox_rte_ether_hdr),
285 sizeof(prox_rte_ipv4_hdr), 1);
287 data = (u8*)(pip4 + 1);
289 *((u32*) data) = 0x2016; // FIXME SPI
290 *((u32*) data + 1) = 0x2; // FIXME SN
292 struct prox_esp_hdr *pesp = (struct prox_esp_hdr*)(pip4+1);
293 pesp->spi = src_addr;//for simplicity assume 1 tunnel per source ip
296 // pesp->spi=0xAAAAAAAA;//debug
297 // pesp->seq =0xBBBBBBBB;//debug
299 u8 *padl = (u8*)data + (8 + encrypt_len - 2 + CIPHER_IV_LENGTH_AES_CBC); // No ESN yet. (-2 means NH is crypted)
300 //padl += CIPHER_IV_LENGTH_AES_CBC;
302 *(padl + 1) = 4; // ipv4 in 4
304 sym_cop->auth.digest.data = data + 8 + CIPHER_IV_LENGTH_AES_CBC +
306 //sym_cop->auth.digest.phys_addr = rte_pktmbuf_mtophys_offset(mbuf, (sizeof(prox_rte_ether_hdr) + sizeof(prox_rte_ipv4_hdr) + 8 + CIPHER_IV_LENGTH_AES_CBC + encrypt_len));
307 sym_cop->auth.digest.phys_addr = rte_pktmbuf_iova_offset(mbuf,
308 (sizeof(prox_rte_ether_hdr) + sizeof(prox_rte_ipv4_hdr)
309 + 8 + CIPHER_IV_LENGTH_AES_CBC + encrypt_len));
310 //sym_cop->auth.digest.length = DIGEST_BYTE_LENGTH_SHA1;
312 //sym_cop->cipher.iv.data = data + 8;
313 //sym_cop->cipher.iv.phys_addr = rte_pktmbuf_mtophys(mbuf) + sizeof(prox_rte_ether_hdr) + sizeof(prox_rte_ipv4_hdr) + 4 + 4;
314 //sym_cop->cipher.iv.length = CIPHER_IV_LENGTH_AES_CBC;
316 //rte_memcpy(sym_cop->cipher.iv.data, aes_cbc_iv, CIPHER_IV_LENGTH_AES_CBC);
318 uint8_t *iv_ptr = rte_crypto_op_ctod_offset(cop, uint8_t *, IV_OFFSET);
319 rte_memcpy(iv_ptr, aes_cbc_iv, CIPHER_IV_LENGTH_AES_CBC);
322 sym_cop->cipher.data.offset = sizeof(prox_rte_ether_hdr) + sizeof(prox_rte_ipv4_hdr) + 4 + 4 + CIPHER_IV_LENGTH_AES_CBC;
323 sym_cop->cipher.data.length = encrypt_len;
325 uint64_t *iv = (uint64_t *)(pesp + 1);
326 memset(iv, 0, CIPHER_IV_LENGTH_AES_CBC);
328 //uint64_t *iv = (uint64_t *)(pesp + 1);
329 //memset(iv, 0, CIPHER_IV_LENGTH_AES_CBC);
330 sym_cop->cipher.data.offset = sizeof(prox_rte_ether_hdr) +
331 sizeof(prox_rte_ipv4_hdr) + sizeof(struct prox_esp_hdr);
332 sym_cop->cipher.data.length = encrypt_len + CIPHER_IV_LENGTH_AES_CBC;
335 sym_cop->auth.data.offset = sizeof(prox_rte_ether_hdr) +
336 sizeof(prox_rte_ipv4_hdr);
337 sym_cop->auth.data.length = sizeof(struct prox_esp_hdr) +
338 CIPHER_IV_LENGTH_AES_CBC + encrypt_len;// + 4;// FIXME
340 sym_cop->m_src = mbuf;
341 rte_crypto_op_attach_sym_session(cop, task->sess);
343 //cop->type = RTE_CRYPTO_OP_TYPE_SYMMETRIC;
344 //cop->status = RTE_CRYPTO_OP_STATUS_NOT_PROCESSED;
349 static inline uint8_t handle_esp_ah_dec(struct task_esp *task,
350 struct rte_mbuf *mbuf, struct rte_crypto_op *cop)
352 struct rte_crypto_sym_op *sym_cop = cop->sym;
353 prox_rte_ether_hdr *peth = rte_pktmbuf_mtod(mbuf,
354 prox_rte_ether_hdr *);
355 prox_rte_ipv4_hdr* pip4 = (prox_rte_ipv4_hdr *)(peth + 1);
356 uint16_t ipv4_length = rte_be_to_cpu_16(pip4->total_length);
357 u8 *data = (u8*)(pip4 + 1);
359 if (pip4->next_proto_id != IPPROTO_ESP){
360 plog_info("Received non ESP packet on esp dec\n");
361 plogdx_info(mbuf, "DEC RX: ");
365 rte_crypto_op_attach_sym_session(cop, task->sess);
367 sym_cop->auth.digest.data = (unsigned char *)((unsigned char*)pip4 +
368 ipv4_length - DIGEST_BYTE_LENGTH_SHA1);
369 //sym_cop->auth.digest.phys_addr = rte_pktmbuf_mtophys_offset(mbuf, sizeof(prox_rte_ether_hdr) + sizeof(prox_rte_ipv4_hdr) + sizeof(struct prox_esp_hdr)); // FIXME
370 sym_cop->auth.digest.phys_addr = rte_pktmbuf_iova_offset(mbuf,
371 sizeof(prox_rte_ether_hdr) + sizeof(prox_rte_ipv4_hdr)
372 + sizeof(struct prox_esp_hdr));
373 //sym_cop->auth.digest.length = DIGEST_BYTE_LENGTH_SHA1;
375 //sym_cop->cipher.iv.data = (uint8_t *)data + 8;
376 //sym_cop->cipher.iv.phys_addr = rte_pktmbuf_mtophys(mbuf) + sizeof(prox_rte_ether_hdr) + sizeof(prox_rte_ipv4_hdr) + 4 + 4;
377 //sym_cop->cipher.iv.length = CIPHER_IV_LENGTH_AES_CBC;
380 rte_memcpy(rte_crypto_op_ctod_offset(cop, uint8_t *, IV_OFFSET),
382 CIPHER_IV_LENGTH_AES_CBC);
384 uint8_t * iv = (uint8_t *)(pip4 + 1) + sizeof(struct prox_esp_hdr);
385 rte_memcpy(rte_crypto_op_ctod_offset(cop, uint8_t *, IV_OFFSET),
387 CIPHER_IV_LENGTH_AES_CBC);
390 sym_cop->auth.data.offset = sizeof(prox_rte_ether_hdr) +
391 sizeof(prox_rte_ipv4_hdr);
392 sym_cop->auth.data.length = ipv4_length - sizeof(prox_rte_ipv4_hdr) - 4 -
393 CIPHER_IV_LENGTH_AES_CBC;
395 sym_cop->cipher.data.offset = sizeof(prox_rte_ether_hdr) +
396 sizeof(prox_rte_ipv4_hdr) + sizeof(struct prox_esp_hdr) +
397 CIPHER_IV_LENGTH_AES_CBC;
398 sym_cop->cipher.data.length = ipv4_length - sizeof(prox_rte_ipv4_hdr) -
399 CIPHER_IV_LENGTH_AES_CBC - 28; // FIXME
401 sym_cop->m_src = mbuf;
405 static inline void do_ipv4_swap(struct task_esp *task, struct rte_mbuf *mbuf)
407 prox_rte_ether_hdr *peth = rte_pktmbuf_mtod(mbuf,
408 prox_rte_ether_hdr *);
409 prox_rte_ether_addr src_mac = peth->s_addr;
410 prox_rte_ether_addr dst_mac = peth->d_addr;
411 uint32_t src_ip, dst_ip;
413 prox_rte_ipv4_hdr* pip4 = (prox_rte_ipv4_hdr *)(peth + 1);
414 src_ip = pip4->src_addr;
415 dst_ip = pip4->dst_addr;
417 //peth->s_addr = dst_mac;
418 peth->d_addr = src_mac;//should be replaced by arp
419 pip4->src_addr = dst_ip;
420 pip4->dst_addr = src_ip;
421 prox_rte_ether_addr_copy(&task->local_mac, &peth->s_addr);
425 static void init_task_esp_enc(struct task_base *tbase, struct task_args *targ)
427 struct task_esp *task = (struct task_esp *)tbase;
428 unsigned int session_size;
430 tbase->flags |= TBASE_FLAG_NEVER_FLUSH;
432 uint8_t lcore_id = targ->lconf->id;
434 task->handle_esp_finish = handle_enc_finish;
435 task->handle_esp_ah = handle_esp_ah_enc;
437 sprintf(name, "core_%03u_crypto_pool", lcore_id);
438 task->crypto_op_pool = rte_crypto_op_pool_create(name,
439 RTE_CRYPTO_OP_TYPE_SYMMETRIC, targ->nb_mbuf, 128,
440 MAXIMUM_IV_LENGTH, rte_socket_id());
441 plog_info("rte_crypto_op_pool_create nb_elements =%d\n",
443 PROX_PANIC(task->crypto_op_pool == NULL, "Can't create ENC \
446 task->cdev_id = get_cdev_id();
448 struct rte_cryptodev_config cdev_conf;
449 cdev_conf.nb_queue_pairs = 2;
450 cdev_conf.socket_id = rte_socket_id();
451 rte_cryptodev_configure(task->cdev_id, &cdev_conf);
453 session_size = rte_cryptodev_sym_get_private_session_size(
455 plog_info("rte_cryptodev_sym_get_private_session_size=%d\n",
457 sprintf(name, "core_%03u_session_pool", lcore_id);
458 task->session_pool = rte_cryptodev_sym_session_pool_create(name,
463 PROX_PANIC(task->session_pool == NULL, "Failed rte_mempool_create\n");
466 plog_info("enc: task->qp_id=%u\n", task->qp_id);
467 struct prox_rte_cryptodev_qp_conf qp_conf;
468 qp_conf.nb_descriptors = 2048;
469 qp_conf.mp_session = task->session_pool;
470 prox_rte_cryptodev_queue_pair_setup(task->cdev_id, task->qp_id,
471 &qp_conf, rte_cryptodev_socket_id(task->cdev_id));
473 int ret = rte_cryptodev_start(task->cdev_id);
474 PROX_PANIC(ret < 0, "Failed to start device\n");
476 //Setup Cipher Parameters
477 struct rte_crypto_sym_xform cipher_xform = {0};
478 struct rte_crypto_sym_xform auth_xform = {0};
480 cipher_xform.type = RTE_CRYPTO_SYM_XFORM_CIPHER;
481 // cipher_xform.next = &auth_xform;
482 cipher_xform.next = NULL; //CRYPTO_ONLY
484 cipher_xform.cipher.algo = RTE_CRYPTO_CIPHER_AES_CBC;
485 cipher_xform.cipher.op = RTE_CRYPTO_CIPHER_OP_ENCRYPT;
486 cipher_xform.cipher.key.data = aes_cbc_key;
487 cipher_xform.cipher.key.length = CIPHER_KEY_LENGTH_AES_CBC;
489 cipher_xform.cipher.iv.offset = IV_OFFSET;
490 cipher_xform.cipher.iv.length = CIPHER_IV_LENGTH_AES_CBC;
492 //Setup HMAC Parameters
493 auth_xform.type = RTE_CRYPTO_SYM_XFORM_AUTH;
494 auth_xform.next = NULL;
495 auth_xform.auth.op = RTE_CRYPTO_AUTH_OP_GENERATE;
496 auth_xform.auth.algo = RTE_CRYPTO_AUTH_SHA1_HMAC;
497 auth_xform.auth.key.length = DIGEST_BYTE_LENGTH_SHA1;
498 auth_xform.auth.key.data = hmac_sha1_key;
499 auth_xform.auth.digest_length = DIGEST_BYTE_LENGTH_SHA1;
501 auth_xform.auth.iv.offset = 0;
502 auth_xform.auth.iv.length = 0;
504 task->sess = rte_cryptodev_sym_session_create(task->cdev_id,
505 &cipher_xform, task->session_pool);
506 PROX_PANIC(task->sess < 0, "Failed ENC sym_session_create\n");
508 task->local_ipv4 = rte_cpu_to_be_32(targ->local_ipv4);
509 task->remote_ipv4 = rte_cpu_to_be_32(targ->remote_ipv4);
510 //memcpy(&task->src_mac, &prox_port_cfg[task->base.tx_params_hw.tx_port_queue->port].eth_addr, sizeof(prox_rte_ether_addr));
511 struct prox_port_cfg *port = find_reachable_port(targ);
512 memcpy(&task->local_mac, &port->eth_addr, sizeof(prox_rte_ether_addr));
514 if (targ->flags & TASK_ARG_DST_MAC_SET){
515 memcpy(&task->dst_mac, &targ->edaddr, sizeof(task->dst_mac));
516 plog_info("TASK_ARG_DST_MAC_SET ("MAC_BYTES_FMT")\n",
517 MAC_BYTES(task->dst_mac.addr_bytes));
518 //prox_rte_ether_addr_copy(&ptask->dst_mac, &peth->d_addr);
519 //rte_memcpy(hdr, task->src_dst_mac, sizeof(task->src_dst_mac));
523 static void init_task_esp_dec(struct task_base *tbase, struct task_args *targ)
525 struct task_esp *task = (struct task_esp *)tbase;
526 unsigned int session_size;
528 tbase->flags |= TBASE_FLAG_NEVER_FLUSH;
530 uint8_t lcore_id = targ->lconf->id;
532 task->handle_esp_finish = handle_dec_finish;
533 task->handle_esp_ah = handle_esp_ah_dec;
535 sprintf(name, "core_%03u_crypto_pool", lcore_id);
536 task->crypto_op_pool = rte_crypto_op_pool_create(name,
537 RTE_CRYPTO_OP_TYPE_SYMMETRIC, targ->nb_mbuf, 128,
538 MAXIMUM_IV_LENGTH, rte_socket_id());
539 PROX_PANIC(task->crypto_op_pool == NULL, "Can't create DEC \
542 task->cdev_id = get_cdev_id();
543 struct rte_cryptodev_config cdev_conf;
544 cdev_conf.nb_queue_pairs = 2;
545 cdev_conf.socket_id = SOCKET_ID_ANY;
546 cdev_conf.socket_id = rte_socket_id();
547 rte_cryptodev_configure(task->cdev_id, &cdev_conf);
549 session_size = rte_cryptodev_sym_get_private_session_size(
551 plog_info("rte_cryptodev_sym_get_private_session_size=%d\n",
553 sprintf(name, "core_%03u_session_pool", lcore_id);
554 task->session_pool = rte_cryptodev_sym_session_pool_create(name,
559 PROX_PANIC(task->session_pool == NULL, "Failed rte_mempool_create\n");
562 plog_info("dec: task->qp_id=%u\n", task->qp_id);
563 struct prox_rte_cryptodev_qp_conf qp_conf;
564 qp_conf.nb_descriptors = 2048;
565 qp_conf.mp_session = task->session_pool;
566 prox_rte_cryptodev_queue_pair_setup(task->cdev_id, task->qp_id,
567 &qp_conf, rte_cryptodev_socket_id(task->cdev_id));
569 int ret = rte_cryptodev_start(task->cdev_id);
570 PROX_PANIC(ret < 0, "Failed to start device\n");
572 //Setup Cipher Parameters
573 struct rte_crypto_sym_xform cipher_xform = {0};
574 struct rte_crypto_sym_xform auth_xform = {0};
576 cipher_xform.type = RTE_CRYPTO_SYM_XFORM_CIPHER;
577 cipher_xform.next = NULL;
578 cipher_xform.cipher.algo = RTE_CRYPTO_CIPHER_AES_CBC;
579 cipher_xform.cipher.op = RTE_CRYPTO_CIPHER_OP_DECRYPT;
580 cipher_xform.cipher.key.data = aes_cbc_key;
581 cipher_xform.cipher.key.length = CIPHER_KEY_LENGTH_AES_CBC;
583 cipher_xform.cipher.iv.offset = IV_OFFSET;
584 cipher_xform.cipher.iv.length = CIPHER_IV_LENGTH_AES_CBC;
586 //Setup HMAC Parameters
587 auth_xform.type = RTE_CRYPTO_SYM_XFORM_AUTH;
588 auth_xform.next = &cipher_xform;
589 auth_xform.auth.op = RTE_CRYPTO_AUTH_OP_VERIFY;
590 auth_xform.auth.algo = RTE_CRYPTO_AUTH_SHA1_HMAC;
591 auth_xform.auth.key.length = DIGEST_BYTE_LENGTH_SHA1;
592 auth_xform.auth.key.data = hmac_sha1_key;
593 auth_xform.auth.digest_length = DIGEST_BYTE_LENGTH_SHA1;
595 auth_xform.auth.iv.offset = 0;
596 auth_xform.auth.iv.length = 0;
598 task->sess = rte_cryptodev_sym_session_create(task->cdev_id, &cipher_xform,
600 PROX_PANIC(task->sess < 0, "Failed DEC sym_session_create\n");
602 task->local_ipv4 = rte_cpu_to_be_32(targ->local_ipv4);
603 task->remote_ipv4 = rte_cpu_to_be_32(targ->remote_ipv4);
604 //memcpy(&task->src_mac, &prox_port_cfg[task->base.tx_params_hw.tx_port_queue->port].eth_addr, sizeof(prox_rte_ether_addr));
605 struct prox_port_cfg *port = find_reachable_port(targ);
606 memcpy(&task->local_mac, &port->eth_addr, sizeof(prox_rte_ether_addr));
608 if (targ->flags & TASK_ARG_DST_MAC_SET){
609 memcpy(&task->dst_mac, &targ->edaddr, sizeof(task->dst_mac));
610 plog_info("TASK_ARG_DST_MAC_SET ("MAC_BYTES_FMT")\n",
611 MAC_BYTES(task->dst_mac.addr_bytes));
612 //prox_rte_ether_addr_copy(&ptask->dst_mac, &peth->d_addr);
613 //rte_memcpy(hdr, task->src_dst_mac, sizeof(task->src_dst_mac));
617 static int crypto_send_burst(struct task_esp *task, uint16_t n)
619 uint8_t out[MAX_PKT_BURST];
620 struct rte_mbuf *mbufs[MAX_PKT_BURST];
623 ret = rte_cryptodev_enqueue_burst(task->cdev_id,
624 task->qp_id, task->ops_burst, n);
625 if (unlikely(ret < n)) {
626 for (i = 0; i < (n-ret); i++) {
627 mbufs[i] = task->ops_burst[ret + i]->sym->m_src;
628 out[i] = OUT_DISCARD;
629 rte_crypto_op_free(task->ops_burst[ret + i]);
631 return task->base.tx_pkt(&task->base, mbufs, i, out);
636 static int handle_esp_bulk(struct task_base *tbase, struct rte_mbuf **mbufs,
639 struct task_esp *task = (struct task_esp *)tbase;
640 uint8_t out[MAX_PKT_BURST];
642 uint16_t nb_deq = 0, j, idx = 0;
643 struct rte_mbuf *drop_mbufs[MAX_PKT_BURST];
644 struct rte_crypto_op *ops_burst[MAX_PKT_BURST];
647 if (rte_crypto_op_bulk_alloc(task->crypto_op_pool,
648 RTE_CRYPTO_OP_TYPE_SYMMETRIC,
649 ops_burst, n_pkts) != n_pkts) {
650 plog_info("Failed to allocate crypto operations, discarding \
651 %d packets\n", n_pkts);
652 for (j = 0; j < n_pkts; j++) {
653 out[j] = OUT_DISCARD;
655 nbr_tx_pkt += task->base.tx_pkt(&task->base, mbufs, n_pkts,
659 for (j = 0; j < n_pkts; j++) {
660 result = task->handle_esp_ah(task, mbufs[j],
663 task->ops_burst[task->len] = ops_burst[j];
665 /* enough ops to be sent */
666 if (task->len == MAX_PKT_BURST) {
667 nbr_tx_pkt += crypto_send_burst(task,
668 (uint16_t) MAX_PKT_BURST);
673 drop_mbufs[idx] = mbufs[j];
676 rte_crypto_op_free(ops_burst[j]);
677 plog_info("Failed handle_esp_ah for 1 \
681 if (idx) nbr_tx_pkt += task->base.tx_pkt(&task->base,
682 drop_mbufs, idx, out);
685 nb_deq = rte_cryptodev_dequeue_burst(task->cdev_id,
686 task->qp_id, ops_burst, MAX_PKT_BURST);
687 for (j = 0; j < nb_deq; j++) {
688 mbufs[j] = ops_burst[j]->sym->m_src;
689 out[j] = task->handle_esp_finish(task, mbufs[j],
690 ops_burst[j]->status);
691 rte_crypto_op_free(ops_burst[j]);
693 nbr_tx_pkt += task->base.tx_pkt(&task->base, mbufs, nb_deq,
695 } while (nb_deq == MAX_PKT_BURST);
699 struct task_init task_init_esp_enc = {
701 .mode_str = "esp_enc",
702 .init = init_task_esp_enc,
703 .handle = handle_esp_bulk,
704 .size = sizeof(struct task_esp),
707 struct task_init task_init_esp_dec = {
709 .mode_str = "esp_dec",
710 .init = init_task_esp_dec,
711 .handle = handle_esp_bulk,
712 .size = sizeof(struct task_esp),
715 __attribute__((constructor)) static void reg_task_esp_enc(void)
717 reg_task(&task_init_esp_enc);
720 __attribute__((constructor)) static void reg_task_esp_dec(void)
722 reg_task(&task_init_esp_dec);