diff options
Diffstat (limited to 'common/dpdk/recipes-extended/dpdk/dpdk/dpdk-dev-v4-10-20-crypto-ccp-support-aes-cipher-algo.patch')
-rw-r--r-- | common/dpdk/recipes-extended/dpdk/dpdk/dpdk-dev-v4-10-20-crypto-ccp-support-aes-cipher-algo.patch | 449 |
1 files changed, 449 insertions, 0 deletions
diff --git a/common/dpdk/recipes-extended/dpdk/dpdk/dpdk-dev-v4-10-20-crypto-ccp-support-aes-cipher-algo.patch b/common/dpdk/recipes-extended/dpdk/dpdk/dpdk-dev-v4-10-20-crypto-ccp-support-aes-cipher-algo.patch new file mode 100644 index 00000000..b00686e6 --- /dev/null +++ b/common/dpdk/recipes-extended/dpdk/dpdk/dpdk-dev-v4-10-20-crypto-ccp-support-aes-cipher-algo.patch @@ -0,0 +1,449 @@ +From patchwork Fri Mar 9 08:35:10 2018 +Content-Type: text/plain; charset="utf-8" +MIME-Version: 1.0 +Content-Transfer-Encoding: 7bit +Subject: [dpdk-dev,v4,10/20] crypto/ccp: support aes cipher algo +From: Ravi Kumar <ravi1.kumar@amd.com> +X-Patchwork-Id: 35809 +X-Patchwork-Delegate: pablo.de.lara.guarch@intel.com +Message-Id: <1520584520-130522-10-git-send-email-Ravi1.kumar@amd.com> +List-Id: dev.dpdk.org +To: dev@dpdk.org +Cc: pablo.de.lara.guarch@intel.com +Date: Fri, 9 Mar 2018 03:35:10 -0500 + +Signed-off-by: Ravi Kumar <Ravi1.kumar@amd.com> +--- + drivers/crypto/ccp/ccp_crypto.c | 197 ++++++++++++++++++++++++++++++++++++++- + drivers/crypto/ccp/ccp_crypto.h | 13 +++ + drivers/crypto/ccp/ccp_dev.h | 53 +++++++++++ + drivers/crypto/ccp/ccp_pmd_ops.c | 60 ++++++++++++ + 4 files changed, 321 insertions(+), 2 deletions(-) + +diff --git a/drivers/crypto/ccp/ccp_crypto.c b/drivers/crypto/ccp/ccp_crypto.c +index c17e84f..b097355 100644 +--- a/drivers/crypto/ccp/ccp_crypto.c ++++ b/drivers/crypto/ccp/ccp_crypto.c +@@ -80,6 +80,7 @@ ccp_configure_session_cipher(struct ccp_session *sess, + const struct rte_crypto_sym_xform *xform) + { + const struct rte_crypto_cipher_xform *cipher_xform = NULL; ++ size_t i; + + cipher_xform = &xform->cipher; + +@@ -99,6 +100,21 @@ ccp_configure_session_cipher(struct ccp_session *sess, + sess->iv.length = cipher_xform->iv.length; + + switch (cipher_xform->algo) { ++ case RTE_CRYPTO_CIPHER_AES_CTR: ++ sess->cipher.algo = CCP_CIPHER_ALGO_AES_CTR; ++ sess->cipher.um.aes_mode = CCP_AES_MODE_CTR; ++ sess->cipher.engine = CCP_ENGINE_AES; ++ break; ++ case RTE_CRYPTO_CIPHER_AES_ECB: ++ sess->cipher.algo = CCP_CIPHER_ALGO_AES_CBC; ++ sess->cipher.um.aes_mode = CCP_AES_MODE_ECB; ++ sess->cipher.engine = CCP_ENGINE_AES; ++ break; ++ case RTE_CRYPTO_CIPHER_AES_CBC: ++ sess->cipher.algo = CCP_CIPHER_ALGO_AES_CBC; ++ sess->cipher.um.aes_mode = CCP_AES_MODE_CBC; ++ sess->cipher.engine = CCP_ENGINE_AES; ++ break; + default: + CCP_LOG_ERR("Unsupported cipher algo"); + return -1; +@@ -106,10 +122,27 @@ ccp_configure_session_cipher(struct ccp_session *sess, + + + switch (sess->cipher.engine) { ++ case CCP_ENGINE_AES: ++ if (sess->cipher.key_length == 16) ++ sess->cipher.ut.aes_type = CCP_AES_TYPE_128; ++ else if (sess->cipher.key_length == 24) ++ sess->cipher.ut.aes_type = CCP_AES_TYPE_192; ++ else if (sess->cipher.key_length == 32) ++ sess->cipher.ut.aes_type = CCP_AES_TYPE_256; ++ else { ++ CCP_LOG_ERR("Invalid cipher key length"); ++ return -1; ++ } ++ for (i = 0; i < sess->cipher.key_length ; i++) ++ sess->cipher.key_ccp[sess->cipher.key_length - i - 1] = ++ sess->cipher.key[i]; ++ break; + default: + CCP_LOG_ERR("Invalid CCP Engine"); + return -ENOTSUP; + } ++ sess->cipher.nonce_phys = rte_mem_virt2phy(sess->cipher.nonce); ++ sess->cipher.key_phys = rte_mem_virt2phy(sess->cipher.key_ccp); + return 0; + } + +@@ -235,6 +268,18 @@ ccp_cipher_slot(struct ccp_session *session) + int count = 0; + + switch (session->cipher.algo) { ++ case CCP_CIPHER_ALGO_AES_CBC: ++ count = 2; ++ /**< op + passthrough for iv */ ++ break; ++ case CCP_CIPHER_ALGO_AES_ECB: ++ count = 1; ++ /**<only op*/ ++ break; ++ case CCP_CIPHER_ALGO_AES_CTR: ++ count = 2; ++ /**< op + passthrough for iv */ ++ break; + default: + CCP_LOG_ERR("Unsupported cipher algo %d", + session->cipher.algo); +@@ -297,10 +342,146 @@ ccp_compute_slot_count(struct ccp_session *session) + return count; + } + ++static void ++ccp_perform_passthru(struct ccp_passthru *pst, ++ struct ccp_queue *cmd_q) ++{ ++ struct ccp_desc *desc; ++ union ccp_function function; ++ ++ desc = &cmd_q->qbase_desc[cmd_q->qidx]; ++ ++ CCP_CMD_ENGINE(desc) = CCP_ENGINE_PASSTHRU; ++ ++ CCP_CMD_SOC(desc) = 0; ++ CCP_CMD_IOC(desc) = 0; ++ CCP_CMD_INIT(desc) = 0; ++ CCP_CMD_EOM(desc) = 0; ++ CCP_CMD_PROT(desc) = 0; ++ ++ function.raw = 0; ++ CCP_PT_BYTESWAP(&function) = pst->byte_swap; ++ CCP_PT_BITWISE(&function) = pst->bit_mod; ++ CCP_CMD_FUNCTION(desc) = function.raw; ++ ++ CCP_CMD_LEN(desc) = pst->len; ++ ++ if (pst->dir) { ++ CCP_CMD_SRC_LO(desc) = (uint32_t)(pst->src_addr); ++ CCP_CMD_SRC_HI(desc) = high32_value(pst->src_addr); ++ CCP_CMD_SRC_MEM(desc) = CCP_MEMTYPE_SYSTEM; ++ ++ CCP_CMD_DST_LO(desc) = (uint32_t)(pst->dest_addr); ++ CCP_CMD_DST_HI(desc) = 0; ++ CCP_CMD_DST_MEM(desc) = CCP_MEMTYPE_SB; ++ ++ if (pst->bit_mod != CCP_PASSTHRU_BITWISE_NOOP) ++ CCP_CMD_LSB_ID(desc) = cmd_q->sb_key; ++ } else { ++ ++ CCP_CMD_SRC_LO(desc) = (uint32_t)(pst->src_addr); ++ CCP_CMD_SRC_HI(desc) = 0; ++ CCP_CMD_SRC_MEM(desc) = CCP_MEMTYPE_SB; ++ ++ CCP_CMD_DST_LO(desc) = (uint32_t)(pst->dest_addr); ++ CCP_CMD_DST_HI(desc) = high32_value(pst->dest_addr); ++ CCP_CMD_DST_MEM(desc) = CCP_MEMTYPE_SYSTEM; ++ } ++ ++ cmd_q->qidx = (cmd_q->qidx + 1) % COMMANDS_PER_QUEUE; ++} ++ ++static int ++ccp_perform_aes(struct rte_crypto_op *op, ++ struct ccp_queue *cmd_q, ++ struct ccp_batch_info *b_info) ++{ ++ struct ccp_session *session; ++ union ccp_function function; ++ uint8_t *lsb_buf; ++ struct ccp_passthru pst = {0}; ++ struct ccp_desc *desc; ++ phys_addr_t src_addr, dest_addr, key_addr; ++ uint8_t *iv; ++ ++ session = (struct ccp_session *)get_session_private_data( ++ op->sym->session, ++ ccp_cryptodev_driver_id); ++ function.raw = 0; ++ ++ iv = rte_crypto_op_ctod_offset(op, uint8_t *, session->iv.offset); ++ if (session->cipher.um.aes_mode != CCP_AES_MODE_ECB) { ++ if (session->cipher.um.aes_mode == CCP_AES_MODE_CTR) { ++ rte_memcpy(session->cipher.nonce + AES_BLOCK_SIZE, ++ iv, session->iv.length); ++ pst.src_addr = (phys_addr_t)session->cipher.nonce_phys; ++ CCP_AES_SIZE(&function) = 0x1F; ++ } else { ++ lsb_buf = ++ &(b_info->lsb_buf[b_info->lsb_buf_idx*CCP_SB_BYTES]); ++ rte_memcpy(lsb_buf + ++ (CCP_SB_BYTES - session->iv.length), ++ iv, session->iv.length); ++ pst.src_addr = b_info->lsb_buf_phys + ++ (b_info->lsb_buf_idx * CCP_SB_BYTES); ++ b_info->lsb_buf_idx++; ++ } ++ ++ pst.dest_addr = (phys_addr_t)(cmd_q->sb_iv * CCP_SB_BYTES); ++ pst.len = CCP_SB_BYTES; ++ pst.dir = 1; ++ pst.bit_mod = CCP_PASSTHRU_BITWISE_NOOP; ++ pst.byte_swap = CCP_PASSTHRU_BYTESWAP_256BIT; ++ ccp_perform_passthru(&pst, cmd_q); ++ } ++ ++ desc = &cmd_q->qbase_desc[cmd_q->qidx]; ++ ++ src_addr = rte_pktmbuf_mtophys_offset(op->sym->m_src, ++ op->sym->cipher.data.offset); ++ if (likely(op->sym->m_dst != NULL)) ++ dest_addr = rte_pktmbuf_mtophys_offset(op->sym->m_dst, ++ op->sym->cipher.data.offset); ++ else ++ dest_addr = src_addr; ++ key_addr = session->cipher.key_phys; ++ ++ /* prepare desc for aes command */ ++ CCP_CMD_ENGINE(desc) = CCP_ENGINE_AES; ++ CCP_CMD_INIT(desc) = 1; ++ CCP_CMD_EOM(desc) = 1; ++ ++ CCP_AES_ENCRYPT(&function) = session->cipher.dir; ++ CCP_AES_MODE(&function) = session->cipher.um.aes_mode; ++ CCP_AES_TYPE(&function) = session->cipher.ut.aes_type; ++ CCP_CMD_FUNCTION(desc) = function.raw; ++ ++ CCP_CMD_LEN(desc) = op->sym->cipher.data.length; ++ ++ CCP_CMD_SRC_LO(desc) = ((uint32_t)src_addr); ++ CCP_CMD_SRC_HI(desc) = high32_value(src_addr); ++ CCP_CMD_SRC_MEM(desc) = CCP_MEMTYPE_SYSTEM; ++ ++ CCP_CMD_DST_LO(desc) = ((uint32_t)dest_addr); ++ CCP_CMD_DST_HI(desc) = high32_value(dest_addr); ++ CCP_CMD_DST_MEM(desc) = CCP_MEMTYPE_SYSTEM; ++ ++ CCP_CMD_KEY_LO(desc) = ((uint32_t)key_addr); ++ CCP_CMD_KEY_HI(desc) = high32_value(key_addr); ++ CCP_CMD_KEY_MEM(desc) = CCP_MEMTYPE_SYSTEM; ++ ++ if (session->cipher.um.aes_mode != CCP_AES_MODE_ECB) ++ CCP_CMD_LSB_ID(desc) = cmd_q->sb_iv; ++ ++ cmd_q->qidx = (cmd_q->qidx + 1) % COMMANDS_PER_QUEUE; ++ op->status = RTE_CRYPTO_OP_STATUS_NOT_PROCESSED; ++ return 0; ++} ++ + static inline int + ccp_crypto_cipher(struct rte_crypto_op *op, +- struct ccp_queue *cmd_q __rte_unused, +- struct ccp_batch_info *b_info __rte_unused) ++ struct ccp_queue *cmd_q, ++ struct ccp_batch_info *b_info) + { + int result = 0; + struct ccp_session *session; +@@ -310,6 +491,18 @@ ccp_crypto_cipher(struct rte_crypto_op *op, + ccp_cryptodev_driver_id); + + switch (session->cipher.algo) { ++ case CCP_CIPHER_ALGO_AES_CBC: ++ result = ccp_perform_aes(op, cmd_q, b_info); ++ b_info->desccnt += 2; ++ break; ++ case CCP_CIPHER_ALGO_AES_CTR: ++ result = ccp_perform_aes(op, cmd_q, b_info); ++ b_info->desccnt += 2; ++ break; ++ case CCP_CIPHER_ALGO_AES_ECB: ++ result = ccp_perform_aes(op, cmd_q, b_info); ++ b_info->desccnt += 1; ++ break; + default: + CCP_LOG_ERR("Unsupported cipher algo %d", + session->cipher.algo); +diff --git a/drivers/crypto/ccp/ccp_crypto.h b/drivers/crypto/ccp/ccp_crypto.h +index 4455497..614cd47 100644 +--- a/drivers/crypto/ccp/ccp_crypto.h ++++ b/drivers/crypto/ccp/ccp_crypto.h +@@ -46,7 +46,20 @@ + + #include "ccp_dev.h" + ++#define AES_BLOCK_SIZE 16 ++#define CMAC_PAD_VALUE 0x80 ++#define CTR_NONCE_SIZE 4 ++#define CTR_IV_SIZE 8 + #define CCP_SHA3_CTX_SIZE 200 ++ ++/**Macro helpers for CCP command creation*/ ++#define CCP_AES_SIZE(p) ((p)->aes.size) ++#define CCP_AES_ENCRYPT(p) ((p)->aes.encrypt) ++#define CCP_AES_MODE(p) ((p)->aes.mode) ++#define CCP_AES_TYPE(p) ((p)->aes.type) ++#define CCP_PT_BYTESWAP(p) ((p)->pt.byteswap) ++#define CCP_PT_BITWISE(p) ((p)->pt.bitwise) ++ + /** + * CCP supported AES modes + */ +diff --git a/drivers/crypto/ccp/ccp_dev.h b/drivers/crypto/ccp/ccp_dev.h +index a5c9ef3..759afc1 100644 +--- a/drivers/crypto/ccp/ccp_dev.h ++++ b/drivers/crypto/ccp/ccp_dev.h +@@ -48,6 +48,7 @@ + /**< CCP sspecific */ + #define MAX_HW_QUEUES 5 + #define CCP_MAX_TRNG_RETRIES 10 ++#define CCP_ALIGN(x, y) ((((x) + (y - 1)) / y) * y) + + /**< CCP Register Mappings */ + #define Q_MASK_REG 0x000 +@@ -104,10 +105,52 @@ + #define LSB_SIZE 16 + #define LSB_ITEM_SIZE 32 + #define SLSB_MAP_SIZE (MAX_LSB_CNT * LSB_SIZE) ++#define LSB_ENTRY_NUMBER(LSB_ADDR) (LSB_ADDR / LSB_ITEM_SIZE) + + /* General CCP Defines */ + + #define CCP_SB_BYTES 32 ++/* Word 0 */ ++#define CCP_CMD_DW0(p) ((p)->dw0) ++#define CCP_CMD_SOC(p) (CCP_CMD_DW0(p).soc) ++#define CCP_CMD_IOC(p) (CCP_CMD_DW0(p).ioc) ++#define CCP_CMD_INIT(p) (CCP_CMD_DW0(p).init) ++#define CCP_CMD_EOM(p) (CCP_CMD_DW0(p).eom) ++#define CCP_CMD_FUNCTION(p) (CCP_CMD_DW0(p).function) ++#define CCP_CMD_ENGINE(p) (CCP_CMD_DW0(p).engine) ++#define CCP_CMD_PROT(p) (CCP_CMD_DW0(p).prot) ++ ++/* Word 1 */ ++#define CCP_CMD_DW1(p) ((p)->length) ++#define CCP_CMD_LEN(p) (CCP_CMD_DW1(p)) ++ ++/* Word 2 */ ++#define CCP_CMD_DW2(p) ((p)->src_lo) ++#define CCP_CMD_SRC_LO(p) (CCP_CMD_DW2(p)) ++ ++/* Word 3 */ ++#define CCP_CMD_DW3(p) ((p)->dw3) ++#define CCP_CMD_SRC_MEM(p) ((p)->dw3.src_mem) ++#define CCP_CMD_SRC_HI(p) ((p)->dw3.src_hi) ++#define CCP_CMD_LSB_ID(p) ((p)->dw3.lsb_cxt_id) ++#define CCP_CMD_FIX_SRC(p) ((p)->dw3.fixed) ++ ++/* Words 4/5 */ ++#define CCP_CMD_DW4(p) ((p)->dw4) ++#define CCP_CMD_DST_LO(p) (CCP_CMD_DW4(p).dst_lo) ++#define CCP_CMD_DW5(p) ((p)->dw5.fields.dst_hi) ++#define CCP_CMD_DST_HI(p) (CCP_CMD_DW5(p)) ++#define CCP_CMD_DST_MEM(p) ((p)->dw5.fields.dst_mem) ++#define CCP_CMD_FIX_DST(p) ((p)->dw5.fields.fixed) ++#define CCP_CMD_SHA_LO(p) ((p)->dw4.sha_len_lo) ++#define CCP_CMD_SHA_HI(p) ((p)->dw5.sha_len_hi) ++ ++/* Word 6/7 */ ++#define CCP_CMD_DW6(p) ((p)->key_lo) ++#define CCP_CMD_KEY_LO(p) (CCP_CMD_DW6(p)) ++#define CCP_CMD_DW7(p) ((p)->dw7) ++#define CCP_CMD_KEY_HI(p) ((p)->dw7.key_hi) ++#define CCP_CMD_KEY_MEM(p) ((p)->dw7.key_mem) + + /* bitmap */ + enum { +@@ -412,6 +455,16 @@ struct ccp_desc { + }; + + /** ++ * ccp memory type ++ */ ++enum ccp_memtype { ++ CCP_MEMTYPE_SYSTEM = 0, ++ CCP_MEMTYPE_SB, ++ CCP_MEMTYPE_LOCAL, ++ CCP_MEMTYPE_LAST, ++}; ++ ++/** + * cmd id to follow order + */ + enum ccp_cmd_order { +diff --git a/drivers/crypto/ccp/ccp_pmd_ops.c b/drivers/crypto/ccp/ccp_pmd_ops.c +index d483a74..5f56242 100644 +--- a/drivers/crypto/ccp/ccp_pmd_ops.c ++++ b/drivers/crypto/ccp/ccp_pmd_ops.c +@@ -39,6 +39,66 @@ + #include "ccp_crypto.h" + + static const struct rte_cryptodev_capabilities ccp_pmd_capabilities[] = { ++ { /* AES ECB */ ++ .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, ++ {.sym = { ++ .xform_type = RTE_CRYPTO_SYM_XFORM_CIPHER, ++ {.cipher = { ++ .algo = RTE_CRYPTO_CIPHER_AES_ECB, ++ .block_size = 16, ++ .key_size = { ++ .min = 16, ++ .max = 32, ++ .increment = 8 ++ }, ++ .iv_size = { ++ .min = 0, ++ .max = 0, ++ .increment = 0 ++ } ++ }, } ++ }, } ++ }, ++ { /* AES CBC */ ++ .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, ++ {.sym = { ++ .xform_type = RTE_CRYPTO_SYM_XFORM_CIPHER, ++ {.cipher = { ++ .algo = RTE_CRYPTO_CIPHER_AES_CBC, ++ .block_size = 16, ++ .key_size = { ++ .min = 16, ++ .max = 32, ++ .increment = 8 ++ }, ++ .iv_size = { ++ .min = 16, ++ .max = 16, ++ .increment = 0 ++ } ++ }, } ++ }, } ++ }, ++ { /* AES CTR */ ++ .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC, ++ {.sym = { ++ .xform_type = RTE_CRYPTO_SYM_XFORM_CIPHER, ++ {.cipher = { ++ .algo = RTE_CRYPTO_CIPHER_AES_CTR, ++ .block_size = 16, ++ .key_size = { ++ .min = 16, ++ .max = 32, ++ .increment = 8 ++ }, ++ .iv_size = { ++ .min = 16, ++ .max = 16, ++ .increment = 0 ++ } ++ }, } ++ }, } ++ }, + RTE_CRYPTODEV_END_OF_CAPABILITIES_LIST() + }; + |