if ((node == dev_to_node(&GET_DEV(tmp_dev)) ||
dev_to_node(&GET_DEV(tmp_dev)) < 0) &&
adf_dev_started(tmp_dev) &&
!list_empty(&tmp_dev->crypto_list)) {
ctr = atomic_read(&tmp_dev->ref_count); if (best > ctr) {
accel_dev = tmp_dev;
best = ctr;
}
}
}
if (!accel_dev) {
pr_debug_ratelimited("QAT: Could not find a device on node %d\n", node); /* Get any started device */
list_for_each_entry(tmp_dev, adf_devmgr_get_head(), list) { if (adf_dev_started(tmp_dev) &&
!list_empty(&tmp_dev->crypto_list)) {
accel_dev = tmp_dev; break;
}
}
}
if (!accel_dev) return NULL;
best = ~0;
list_for_each_entry(tmp_inst, &accel_dev->crypto_list, list) { unsignedlong ctr;
ctr = atomic_read(&tmp_inst->refctr); if (best > ctr) {
inst = tmp_inst;
best = ctr;
}
} if (inst) { if (adf_dev_get(accel_dev)) {
dev_err(&GET_DEV(accel_dev), "Could not increment dev refctr\n"); return NULL;
}
atomic_inc(&inst->refctr);
} return inst;
}
/** * qat_crypto_vf_dev_config() - create dev config required to create * crypto inst. * * @accel_dev: Pointer to acceleration device. * * Function creates device configuration required to create * asym, sym or, crypto instances * * Return: 0 on success, error code otherwise.
*/ int qat_crypto_vf_dev_config(struct adf_accel_dev *accel_dev)
{
u16 ring_to_svc_map = GET_HW_DATA(accel_dev)->ring_to_svc_map;
if (ring_to_svc_map != ADF_GEN2_DEFAULT_RING_TO_SRV_MAP) {
dev_err(&GET_DEV(accel_dev), "Unsupported ring/service mapping present on PF"); return -EFAULT;
}
staticint qat_crypto_create_instances(struct adf_accel_dev *accel_dev)
{ unsignedlong num_inst, num_msg_sym, num_msg_asym; char key[ADF_CFG_MAX_KEY_LEN_IN_BYTES]; char val[ADF_CFG_MAX_VAL_LEN_IN_BYTES]; unsignedlong sym_bank, asym_bank; struct qat_crypto_instance *inst; int msg_size; int ret; int i;
INIT_LIST_HEAD(&accel_dev->crypto_list);
ret = adf_cfg_get_param_value(accel_dev, SEC, ADF_NUM_CY, val); if (ret) return ret;
ret = kstrtoul(val, 0, &num_inst); if (ret) return ret;
for (i = 0; i < num_inst; i++) {
inst = kzalloc_node(sizeof(*inst), GFP_KERNEL,
dev_to_node(&GET_DEV(accel_dev))); if (!inst) {
ret = -ENOMEM; goto err;
}
staticint qat_crypto_event_handler(struct adf_accel_dev *accel_dev, enum adf_event event)
{ int ret;
switch (event) { case ADF_EVENT_INIT:
ret = qat_crypto_init(accel_dev); break; case ADF_EVENT_SHUTDOWN:
ret = qat_crypto_shutdown(accel_dev); break; case ADF_EVENT_RESTARTING: case ADF_EVENT_RESTARTED: case ADF_EVENT_START: case ADF_EVENT_STOP: default:
ret = 0;
} return ret;
}
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.