/* Flag marking a final request. */ #define REQ_FINAL BIT(0) /* Flag marking a HMAC request. */ #define REQ_FLAGS_HMAC BIT(1) /* Flag set when HW HMAC is being used. */ #define REQ_FLAGS_HMAC_HW BIT(2) /* Flag set when SW HMAC is being used. */ #define REQ_FLAGS_HMAC_SW BIT(3)
/** * struct ocs_hcu_ctx: OCS HCU Transform context. * @hcu_dev: The OCS HCU device used by the transformation. * @key: The key (used only for HMAC transformations). * @key_len: The length of the key. * @is_sm3_tfm: Whether or not this is an SM3 transformation. * @is_hmac_tfm: Whether or not this is a HMAC transformation.
*/ struct ocs_hcu_ctx { struct ocs_hcu_dev *hcu_dev;
u8 key[SHA512_BLOCK_SIZE];
size_t key_len; bool is_sm3_tfm; bool is_hmac_tfm;
};
/** * struct ocs_hcu_rctx - Context for the request. * @hcu_dev: OCS HCU device to be used to service the request. * @flags: Flags tracking request status. * @algo: Algorithm to use for the request. * @blk_sz: Block size of the transformation / request. * @dig_sz: Digest size of the transformation / request. * @dma_list: OCS DMA linked list. * @hash_ctx: OCS HCU hashing context. * @buffer: Buffer to store: partial block of data and SW HMAC * artifacts (ipad, opad, etc.). * @buf_cnt: Number of bytes currently stored in the buffer. * @buf_dma_addr: The DMA address of @buffer (when mapped). * @buf_dma_count: The number of bytes in @buffer currently DMA-mapped. * @sg: Head of the scatterlist entries containing data. * @sg_data_total: Total data in the SG list at any time. * @sg_data_offset: Offset into the data of the current individual SG node. * @sg_dma_nents: Number of sg entries mapped in dma_list. * @nents: Number of entries in the scatterlist.
*/ struct ocs_hcu_rctx { struct ocs_hcu_dev *hcu_dev;
u32 flags; enum ocs_hcu_algo algo;
size_t blk_sz;
size_t dig_sz; struct ocs_hcu_dma_list *dma_list; struct ocs_hcu_hash_ctx hash_ctx; /* * Buffer is double the block size because we need space for SW HMAC * artifacts, i.e: * - ipad (1 block) + a possible partial block of data. * - opad (1 block) + digest of H(k ^ ipad || m)
*/
u8 buffer[2 * SHA512_BLOCK_SIZE];
size_t buf_cnt;
dma_addr_t buf_dma_addr;
size_t buf_dma_count; struct scatterlist *sg; unsignedint sg_data_total; unsignedint sg_data_offset; unsignedint sg_dma_nents; unsignedint nents;
};
/** * struct ocs_hcu_drv - Driver data * @dev_list: The list of HCU devices. * @lock: The lock protecting dev_list.
*/ struct ocs_hcu_drv { struct list_head dev_list;
spinlock_t lock; /* Protects dev_list. */
};
/* * Return the total amount of data in the request; that is: the data in the * request buffer + the data in the sg list.
*/ staticinlineunsignedint kmb_get_total_data(struct ocs_hcu_rctx *rctx)
{ return rctx->sg_data_total + rctx->buf_cnt;
}
/* Move remaining content of scatter-gather list to context buffer. */ staticint flush_sg_to_ocs_buffer(struct ocs_hcu_rctx *rctx)
{
size_t count;
if (rctx->sg_data_total > (sizeof(rctx->buffer) - rctx->buf_cnt)) {
WARN(1, "%s: sg data does not fit in buffer\n", __func__); return -EINVAL;
}
while (rctx->sg_data_total) { if (!rctx->sg) {
WARN(1, "%s: unexpected NULL sg\n", __func__); return -EINVAL;
} /* * If current sg has been fully processed, skip to the next * one.
*/ if (rctx->sg_data_offset == rctx->sg->length) {
rctx->sg = sg_next(rctx->sg);
rctx->sg_data_offset = 0; continue;
} /* * Determine the maximum data available to copy from the node. * Minimum of the length left in the sg node, or the total data * in the request.
*/
count = min(rctx->sg->length - rctx->sg_data_offset,
rctx->sg_data_total); /* Copy from scatter-list entry to context buffer. */
scatterwalk_map_and_copy(&rctx->buffer[rctx->buf_cnt],
rctx->sg, rctx->sg_data_offset,
count, 0);
/* If the HCU device for the request was previously set, return it. */ if (tctx->hcu_dev) return tctx->hcu_dev;
/* * Otherwise, get the first HCU device available (there should be one * and only one device).
*/
spin_lock_bh(&ocs_hcu.lock);
tctx->hcu_dev = list_first_entry_or_null(&ocs_hcu.dev_list, struct ocs_hcu_dev,
list);
spin_unlock_bh(&ocs_hcu.lock);
/* * Prepare for DMA operation: * - DMA-map request context buffer (if needed) * - DMA-map SG list (only the entries to be processed, see note below) * - Allocate OCS HCU DMA linked list (number of elements = SG entries to * process + context buffer (if not empty)). * - Add DMA-mapped request context buffer to OCS HCU DMA list. * - Add SG entries to DMA list. * * Note: if this is a final request, we process all the data in the SG list, * otherwise we can only process up to the maximum amount of block-aligned data * (the remainder will be put into the context buffer and processed in the next * request).
*/ staticint kmb_ocs_dma_prepare(struct ahash_request *req)
{ struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); struct device *dev = rctx->hcu_dev->dev; unsignedint remainder = 0; unsignedint total; int nents;
size_t count; int rc; int i;
/* This function should be called only when there is data to process. */
total = kmb_get_total_data(rctx); if (!total) return -EINVAL;
/* * If this is not a final DMA (terminated DMA), the data passed to the * HCU must be aligned to the block size; compute the remainder data to * be processed in the next request.
*/ if (!(rctx->flags & REQ_FINAL))
remainder = total % rctx->blk_sz;
/* Determine the number of scatter gather list entries to process. */
nents = sg_nents_for_len(req->src, rctx->sg_data_total - remainder);
if (nents < 0) return nents;
/* If there are entries to process, map them. */ if (nents) {
rctx->sg_dma_nents = dma_map_sg(dev, req->src, nents,
DMA_TO_DEVICE); if (!rctx->sg_dma_nents) {
dev_err(dev, "Failed to MAP SG\n");
rc = -ENOMEM; goto cleanup;
}
/* Save the value of nents to pass to dma_unmap_sg. */
rctx->nents = nents;
/* * The value returned by dma_map_sg() can be < nents; so update * nents accordingly.
*/
nents = rctx->sg_dma_nents;
}
/* * If context buffer is not empty, map it and add extra DMA entry for * it.
*/ if (rctx->buf_cnt) {
rctx->buf_dma_addr = dma_map_single(dev, rctx->buffer,
rctx->buf_cnt,
DMA_TO_DEVICE); if (dma_mapping_error(dev, rctx->buf_dma_addr)) {
dev_err(dev, "Failed to map request context buffer\n");
rc = -ENOMEM; goto cleanup;
}
rctx->buf_dma_count = rctx->buf_cnt; /* Increase number of dma entries. */
nents++;
}
/* Add the SG nodes to be processed to the DMA linked list. */
for_each_sg(req->src, rctx->sg, rctx->sg_dma_nents, i) { /* * The number of bytes to add to the list entry is the minimum * between: * - The DMA length of the SG entry. * - The data left to be processed.
*/
count = min(rctx->sg_data_total - remainder,
sg_dma_len(rctx->sg) - rctx->sg_data_offset); /* * Do not create a zero length DMA descriptor. Check in case of * zero length SG node.
*/ if (count == 0) continue; /* Add sg to HCU DMA list. */
rc = ocs_hcu_dma_list_add_tail(rctx->hcu_dev,
rctx->dma_list,
rctx->sg->dma_address,
count); if (rc) goto cleanup;
/* Update amount of data remaining in SG list. */
rctx->sg_data_total -= count;
/* * If remaining data is equal to remainder (note: 'less than' * case should never happen in practice), we are done: update * offset and exit the loop.
*/ if (rctx->sg_data_total <= remainder) {
WARN_ON(rctx->sg_data_total < remainder);
rctx->sg_data_offset += count; break;
}
/* * If we get here is because we need to process the next sg in * the list; set offset within the sg to 0.
*/
rctx->sg_data_offset = 0;
}
return 0;
cleanup:
dev_err(dev, "Failed to prepare DMA.\n");
kmb_ocs_hcu_dma_cleanup(req, rctx);
WARN(rctx->buf_cnt, "%s: Context buffer is not empty\n", __func__);
WARN(!(rctx->flags & REQ_FLAGS_HMAC_SW), "%s: HMAC_SW flag is not set\n", __func__); /* * Key length must be equal to block size. If key is shorter, * we pad it with zero (note: key cannot be longer, since * longer keys are hashed by kmb_ocs_hcu_setkey()).
*/ if (ctx->key_len > rctx->blk_sz) {
WARN(1, "%s: Invalid key length in tfm context\n", __func__); return -EINVAL;
}
memzero_explicit(&ctx->key[ctx->key_len],
rctx->blk_sz - ctx->key_len);
ctx->key_len = rctx->blk_sz; /* * Prepare IPAD for HMAC. Only done for first block. * HMAC(k,m) = H(k ^ opad || H(k ^ ipad || m)) * k ^ ipad will be first hashed block. * k ^ opad will be calculated in the final request. * Only needed if not using HW HMAC.
*/ for (i = 0; i < rctx->blk_sz; i++)
rctx->buffer[i] = ctx->key[i] ^ HMAC_IPAD_VALUE;
rctx->buf_cnt = rctx->blk_sz;
/* * If hardware HMAC flag is set, perform HMAC in hardware. * * NOTE: this flag implies REQ_FINAL && kmb_get_total_data(rctx)
*/ if (rctx->flags & REQ_FLAGS_HMAC_HW) { /* Map input data into the HCU DMA linked list. */
rc = kmb_ocs_dma_prepare(req); if (rc) goto error;
/* Unmap data and free DMA list regardless of return code. */
kmb_ocs_hcu_dma_cleanup(req, rctx);
/* Process previous return code. */ if (rc) goto error;
goto done;
}
/* Handle update request case. */ if (!(rctx->flags & REQ_FINAL)) { /* Update should always have input data. */ if (!kmb_get_total_data(rctx)) return -EINVAL;
/* Map input data into the HCU DMA linked list. */
rc = kmb_ocs_dma_prepare(req); if (rc) goto error;
/* Do hashing step. */
rc = ocs_hcu_hash_update(hcu_dev, &rctx->hash_ctx,
rctx->dma_list);
/* Unmap data and free DMA list regardless of return code. */
kmb_ocs_hcu_dma_cleanup(req, rctx);
/* Process previous return code. */ if (rc) goto error;
/* * Reset request buffer count (data in the buffer was just * processed).
*/
rctx->buf_cnt = 0; /* * Move remaining sg data into the request buffer, so that it * will be processed during the next request. * * NOTE: we have remaining data if kmb_get_total_data() was not * a multiple of block size.
*/
rc = flush_sg_to_ocs_buffer(rctx); if (rc) goto error;
goto done;
}
/* If we get here, this is a final request. */
/* If there is data to process, use finup. */ if (kmb_get_total_data(rctx)) { /* Map input data into the HCU DMA linked list. */
rc = kmb_ocs_dma_prepare(req); if (rc) goto error;
/* Do hashing step. */
rc = ocs_hcu_hash_finup(hcu_dev, &rctx->hash_ctx,
rctx->dma_list,
req->result, rctx->dig_sz); /* Free DMA list regardless of return code. */
kmb_ocs_hcu_dma_cleanup(req, rctx);
/* Process previous return code. */ if (rc) goto error;
} else { /* Otherwise (if we have no data), use final. */
rc = ocs_hcu_hash_final(hcu_dev, &rctx->hash_ctx, req->result,
rctx->dig_sz); if (rc) goto error;
}
/* * If we are finalizing a SW HMAC request, we just computed the result * of: H(k ^ ipad || m). * * We now need to complete the HMAC calculation with the OPAD step, * that is, we need to compute H(k ^ opad || digest), where digest is * the digest we just obtained, i.e., H(k ^ ipad || m).
*/ if (rctx->flags & REQ_FLAGS_HMAC_SW) { /* * Compute k ^ opad and store it in the request buffer (which * is not used anymore at this point). * Note: key has been padded / hashed already (so keylen == * blksz) .
*/
WARN_ON(tctx->key_len != rctx->blk_sz); for (i = 0; i < rctx->blk_sz; i++)
rctx->buffer[i] = tctx->key[i] ^ HMAC_OPAD_VALUE; /* Now append the digest to the rest of the buffer. */ for (i = 0; (i < rctx->dig_sz); i++)
rctx->buffer[rctx->blk_sz + i] = req->result[i];
/* Now hash the buffer to obtain the final HMAC. */
rc = ocs_hcu_digest(hcu_dev, rctx->algo, rctx->buffer,
rctx->blk_sz + rctx->dig_sz, req->result,
rctx->dig_sz); if (rc) goto error;
}
switch (rctx->dig_sz) { #ifdef CONFIG_CRYPTO_DEV_KEEMBAY_OCS_HCU_HMAC_SHA224 case SHA224_DIGEST_SIZE:
rctx->blk_sz = SHA224_BLOCK_SIZE;
rctx->algo = OCS_HCU_ALGO_SHA224; break; #endif/* CONFIG_CRYPTO_DEV_KEEMBAY_OCS_HCU_HMAC_SHA224 */ case SHA256_DIGEST_SIZE:
rctx->blk_sz = SHA256_BLOCK_SIZE; /* * SHA256 and SM3 have the same digest size: use info from tfm * context to find out which one we should use.
*/
rctx->algo = ctx->is_sm3_tfm ? OCS_HCU_ALGO_SM3 :
OCS_HCU_ALGO_SHA256; break; case SHA384_DIGEST_SIZE:
rctx->blk_sz = SHA384_BLOCK_SIZE;
rctx->algo = OCS_HCU_ALGO_SHA384; break; case SHA512_DIGEST_SIZE:
rctx->blk_sz = SHA512_BLOCK_SIZE;
rctx->algo = OCS_HCU_ALGO_SHA512; break; default: return -EINVAL;
}
/* * If we are doing HMAC, then we must use SW-assisted HMAC, since HW * HMAC does not support context switching (there it can only be used * with finup() or digest()).
*/ if (rctx->flags & REQ_FLAGS_HMAC &&
!(rctx->flags & REQ_FLAGS_HMAC_SW)) {
rctx->flags |= REQ_FLAGS_HMAC_SW;
rc = prepare_ipad(req); if (rc) return rc;
}
/* * If remaining sg_data fits into ctx buffer, just copy it there; we'll * process it at the next update() or final().
*/ if (rctx->sg_data_total <= (sizeof(rctx->buffer) - rctx->buf_cnt)) return flush_sg_to_ocs_buffer(rctx);
return kmb_ocs_hcu_handle_queue(req);
}
/* Common logic for kmb_ocs_hcu_final() and kmb_ocs_hcu_finup(). */ staticint kmb_ocs_hcu_fin_common(struct ahash_request *req)
{ struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); struct ocs_hcu_ctx *ctx = crypto_ahash_ctx(tfm); int rc;
rctx->flags |= REQ_FINAL;
/* * If this is a HMAC request and, so far, we didn't have to switch to * SW HMAC, check if we can use HW HMAC.
*/ if (rctx->flags & REQ_FLAGS_HMAC &&
!(rctx->flags & REQ_FLAGS_HMAC_SW)) { /* * If we are here, it means we never processed any data so far, * so we can use HW HMAC, but only if there is some data to * process (since OCS HW MAC does not support zero-length * messages) and the key length is supported by the hardware * (OCS HCU HW only supports length <= 64); if HW HMAC cannot * be used, fall back to SW-assisted HMAC.
*/ if (kmb_get_total_data(rctx) &&
ctx->key_len <= OCS_HCU_HW_KEY_LEN) {
rctx->flags |= REQ_FLAGS_HMAC_HW;
} else {
rctx->flags |= REQ_FLAGS_HMAC_SW;
rc = prepare_ipad(req); if (rc) return rc;
}
}
/* * Key length must be equal to block size: * - If key is shorter, we are done for now (the key will be padded * later on); this is to maximize the use of HW HMAC (which works * only for keys <= 64 bytes). * - If key is longer, we hash it.
*/ if (keylen <= blk_sz) {
memcpy(ctx->key, key, keylen);
ctx->key_len = keylen; return 0;
}
switch (digestsize) { #ifdef CONFIG_CRYPTO_DEV_KEEMBAY_OCS_HCU_HMAC_SHA224 case SHA224_DIGEST_SIZE:
alg_name = "sha224-keembay-ocs"; break; #endif/* CONFIG_CRYPTO_DEV_KEEMBAY_OCS_HCU_HMAC_SHA224 */ case SHA256_DIGEST_SIZE:
alg_name = ctx->is_sm3_tfm ? "sm3-keembay-ocs" : "sha256-keembay-ocs"; break; case SHA384_DIGEST_SIZE:
alg_name = "sha384-keembay-ocs"; break; case SHA512_DIGEST_SIZE:
alg_name = "sha512-keembay-ocs"; break; default: return -EINVAL;
}
ahash_tfm = crypto_alloc_ahash(alg_name, 0, 0); if (IS_ERR(ahash_tfm)) return PTR_ERR(ahash_tfm);
/* Function called when 'tfm' is de-initialized. */ staticvoid kmb_ocs_hcu_hmac_cra_exit(struct crypto_tfm *tfm)
{ struct ocs_hcu_ctx *ctx = crypto_tfm_ctx(tfm);
/* Clear the key. */
memzero_explicit(ctx->key, sizeof(ctx->key));
}
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.