// SPDX-License-Identifier: (BSD-3-Clause OR GPL-2.0-only) /* Copyright(c) 2014 - 2020 Intel Corporation */ #include #include #include "adf_accel_devices.h" #include "adf_common_drv.h" #include "adf_transport.h" #include "adf_transport_access_macros.h" #include "adf_cfg.h" #include "adf_cfg_strings.h" #include "adf_gen2_hw_data.h" #include "qat_crypto.h" #include "icp_qat_fw.h" #define SEC ADF_KERNEL_SEC static struct service_hndl qat_crypto; void qat_crypto_put_instance(struct qat_crypto_instance *inst) { atomic_dec(&inst->refctr); adf_dev_put(inst->accel_dev); } static int qat_crypto_free_instances(struct adf_accel_dev *accel_dev) { struct qat_crypto_instance *inst, *tmp; int i; list_for_each_entry_safe(inst, tmp, &accel_dev->crypto_list, list) { for (i = 0; i < atomic_read(&inst->refctr); i++) qat_crypto_put_instance(inst); if (inst->sym_tx) adf_remove_ring(inst->sym_tx); if (inst->sym_rx) adf_remove_ring(inst->sym_rx); if (inst->pke_tx) adf_remove_ring(inst->pke_tx); if (inst->pke_rx) adf_remove_ring(inst->pke_rx); list_del(&inst->list); kfree(inst); } return 0; } struct qat_crypto_instance *qat_crypto_get_instance_node(int node) { struct adf_accel_dev *accel_dev = NULL, *tmp_dev; struct qat_crypto_instance *inst = NULL, *tmp_inst; unsigned long best = ~0; list_for_each_entry(tmp_dev, adf_devmgr_get_head(), list) { unsigned long ctr; if ((node == dev_to_node(&GET_DEV(tmp_dev)) || dev_to_node(&GET_DEV(tmp_dev)) < 0) && adf_dev_started(tmp_dev) && !list_empty(&tmp_dev->crypto_list)) { ctr = atomic_read(&tmp_dev->ref_count); if (best > ctr) { accel_dev = tmp_dev; best = ctr; } } } if (!accel_dev) { pr_info("QAT: Could not find a device on node %d\n", node); /* Get any started device */ list_for_each_entry(tmp_dev, adf_devmgr_get_head(), list) { if (adf_dev_started(tmp_dev) && !list_empty(&tmp_dev->crypto_list)) { accel_dev = tmp_dev; break; } } } if (!accel_dev) return NULL; best = ~0; list_for_each_entry(tmp_inst, &accel_dev->crypto_list, list) { unsigned long ctr; ctr = atomic_read(&tmp_inst->refctr); if (best > ctr) { inst = tmp_inst; best = ctr; } } if (inst) { if (adf_dev_get(accel_dev)) { dev_err(&GET_DEV(accel_dev), "Could not increment dev refctr\n"); return NULL; } atomic_inc(&inst->refctr); } return inst; } /** * qat_crypto_vf_dev_config() * create dev config required to create crypto inst. * * @accel_dev: Pointer to acceleration device. * * Function creates device configuration required to create * asym, sym or, crypto instances * * Return: 0 on success, error code otherwise. */ int qat_crypto_vf_dev_config(struct adf_accel_dev *accel_dev) { u16 ring_to_svc_map = GET_HW_DATA(accel_dev)->ring_to_svc_map; if (ring_to_svc_map != ADF_GEN2_DEFAULT_RING_TO_SRV_MAP) { dev_err(&GET_DEV(accel_dev), "Unsupported ring/service mapping present on PF"); return -EFAULT; } return qat_crypto_dev_config(accel_dev); } /** * qat_crypto_dev_config() - create dev config required to create crypto inst. * * @accel_dev: Pointer to acceleration device. * * Function creates device configuration required to create crypto instances * * Return: 0 on success, error code otherwise. */ int qat_crypto_dev_config(struct adf_accel_dev *accel_dev) { char key[ADF_CFG_MAX_KEY_LEN_IN_BYTES]; int banks = GET_MAX_BANKS(accel_dev); int cpus = num_online_cpus(); unsigned long val; int instances; int ret; int i; if (adf_hw_dev_has_crypto(accel_dev)) instances = min(cpus, banks); else instances = 0; ret = adf_cfg_section_add(accel_dev, ADF_KERNEL_SEC); if (ret) goto err; ret = adf_cfg_section_add(accel_dev, "Accelerator0"); if (ret) goto err; /* Temporarily set the number of crypto instances to zero to avoid * registering the crypto algorithms. * This will be removed when the algorithms will support the * CRYPTO_TFM_REQ_MAY_BACKLOG flag */ instances = 0; for (i = 0; i < instances; i++) { val = i; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_BANK_NUM, i); ret = adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, &val, ADF_DEC); if (ret) goto err; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_BANK_NUM, i); ret = adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, &val, ADF_DEC); if (ret) goto err; snprintf(key, sizeof(key), ADF_CY "%d" ADF_ETRMGR_CORE_AFFINITY, i); ret = adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, &val, ADF_DEC); if (ret) goto err; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_SIZE, i); val = 128; ret = adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, &val, ADF_DEC); if (ret) goto err; val = 512; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_SIZE, i); ret = adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, &val, ADF_DEC); if (ret) goto err; val = 0; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_TX, i); ret = adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, &val, ADF_DEC); if (ret) goto err; val = 2; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_TX, i); ret = adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, &val, ADF_DEC); if (ret) goto err; val = 8; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_RX, i); ret = adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, &val, ADF_DEC); if (ret) goto err; val = 10; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_RX, i); ret = adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, &val, ADF_DEC); if (ret) goto err; val = ADF_COALESCING_DEF_TIME; snprintf(key, sizeof(key), ADF_ETRMGR_COALESCE_TIMER_FORMAT, i); ret = adf_cfg_add_key_value_param(accel_dev, "Accelerator0", key, &val, ADF_DEC); if (ret) goto err; } val = i; ret = adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, ADF_NUM_CY, &val, ADF_DEC); if (ret) goto err; set_bit(ADF_STATUS_CONFIGURED, &accel_dev->status); return 0; err: dev_err(&GET_DEV(accel_dev), "Failed to start QAT accel dev\n"); return ret; } EXPORT_SYMBOL_GPL(qat_crypto_dev_config); static int qat_crypto_create_instances(struct adf_accel_dev *accel_dev) { unsigned long num_inst, num_msg_sym, num_msg_asym; char key[ADF_CFG_MAX_KEY_LEN_IN_BYTES]; char val[ADF_CFG_MAX_VAL_LEN_IN_BYTES]; unsigned long sym_bank, asym_bank; struct qat_crypto_instance *inst; int msg_size; int ret; int i; INIT_LIST_HEAD(&accel_dev->crypto_list); ret = adf_cfg_get_param_value(accel_dev, SEC, ADF_NUM_CY, val); if (ret) return ret; ret = kstrtoul(val, 0, &num_inst); if (ret) return ret; for (i = 0; i < num_inst; i++) { inst = kzalloc_node(sizeof(*inst), GFP_KERNEL, dev_to_node(&GET_DEV(accel_dev))); if (!inst) { ret = -ENOMEM; goto err; } list_add_tail(&inst->list, &accel_dev->crypto_list); inst->id = i; atomic_set(&inst->refctr, 0); inst->accel_dev = accel_dev; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_BANK_NUM, i); ret = adf_cfg_get_param_value(accel_dev, SEC, key, val); if (ret) goto err; ret = kstrtoul(val, 10, &sym_bank); if (ret) goto err; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_BANK_NUM, i); ret = adf_cfg_get_param_value(accel_dev, SEC, key, val); if (ret) goto err; ret = kstrtoul(val, 10, &asym_bank); if (ret) goto err; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_SIZE, i); ret = adf_cfg_get_param_value(accel_dev, SEC, key, val); if (ret) goto err; ret = kstrtoul(val, 10, &num_msg_sym); if (ret) goto err; num_msg_sym = num_msg_sym >> 1; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_SIZE, i); ret = adf_cfg_get_param_value(accel_dev, SEC, key, val); if (ret) goto err; ret = kstrtoul(val, 10, &num_msg_asym); if (ret) goto err; num_msg_asym = num_msg_asym >> 1; msg_size = ICP_QAT_FW_REQ_DEFAULT_SZ; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_TX, i); ret = adf_create_ring(accel_dev, SEC, sym_bank, num_msg_sym, msg_size, key, NULL, 0, &inst->sym_tx); if (ret) goto err; msg_size = msg_size >> 1; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_TX, i); ret = adf_create_ring(accel_dev, SEC, asym_bank, num_msg_asym, msg_size, key, NULL, 0, &inst->pke_tx); if (ret) goto err; msg_size = ICP_QAT_FW_RESP_DEFAULT_SZ; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_RX, i); ret = adf_create_ring(accel_dev, SEC, sym_bank, num_msg_sym, msg_size, key, qat_alg_callback, 0, &inst->sym_rx); if (ret) goto err; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_RX, i); ret = adf_create_ring(accel_dev, SEC, asym_bank, num_msg_asym, msg_size, key, qat_alg_asym_callback, 0, &inst->pke_rx); if (ret) goto err; } return 0; err: qat_crypto_free_instances(accel_dev); return ret; } static int qat_crypto_init(struct adf_accel_dev *accel_dev) { if (qat_crypto_create_instances(accel_dev)) return -EFAULT; return 0; } static int qat_crypto_shutdown(struct adf_accel_dev *accel_dev) { return qat_crypto_free_instances(accel_dev); } static int qat_crypto_event_handler(struct adf_accel_dev *accel_dev, enum adf_event event) { int ret; switch (event) { case ADF_EVENT_INIT: ret = qat_crypto_init(accel_dev); break; case ADF_EVENT_SHUTDOWN: ret = qat_crypto_shutdown(accel_dev); break; case ADF_EVENT_RESTARTING: case ADF_EVENT_RESTARTED: case ADF_EVENT_START: case ADF_EVENT_STOP: default: ret = 0; } return ret; } int qat_crypto_register(void) { memset(&qat_crypto, 0, sizeof(qat_crypto)); qat_crypto.event_hld = qat_crypto_event_handler; qat_crypto.name = "qat_crypto"; return adf_service_register(&qat_crypto); } int qat_crypto_unregister(void) { return adf_service_unregister(&qat_crypto); }