mirror of
https://github.com/torvalds/linux.git
synced 2024-12-04 18:13:04 +00:00
b42ab1c61a
When the Kunpeng accelerator executes tasks such as encryption and decryption have minimum requirements on the number of device queues. If the number of queues does not meet the requirement, the process initialization will fail. Therefore, the driver checks the number of queues on the device before registering the algorithm. If the number does not meet the requirements, the driver does not register the algorithm to crypto subsystem, the device is still added to the qm_list. Signed-off-by: Weili Qian <qianweili@huawei.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
650 lines
16 KiB
C
650 lines
16 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/* Copyright (c) 2019 HiSilicon Limited. */
|
|
#include <crypto/internal/acompress.h>
|
|
#include <linux/bitfield.h>
|
|
#include <linux/bitmap.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/scatterlist.h>
|
|
#include "zip.h"
|
|
|
|
/* hisi_zip_sqe dw3 */
|
|
#define HZIP_BD_STATUS_M GENMASK(7, 0)
|
|
/* hisi_zip_sqe dw7 */
|
|
#define HZIP_IN_SGE_DATA_OFFSET_M GENMASK(23, 0)
|
|
#define HZIP_SQE_TYPE_M GENMASK(31, 28)
|
|
/* hisi_zip_sqe dw8 */
|
|
#define HZIP_OUT_SGE_DATA_OFFSET_M GENMASK(23, 0)
|
|
/* hisi_zip_sqe dw9 */
|
|
#define HZIP_REQ_TYPE_M GENMASK(7, 0)
|
|
#define HZIP_ALG_TYPE_DEFLATE 0x01
|
|
#define HZIP_BUF_TYPE_M GENMASK(11, 8)
|
|
#define HZIP_SGL 0x1
|
|
|
|
#define HZIP_ALG_PRIORITY 300
|
|
#define HZIP_SGL_SGE_NR 10
|
|
|
|
#define HZIP_ALG_DEFLATE GENMASK(5, 4)
|
|
|
|
static DEFINE_MUTEX(zip_algs_lock);
|
|
static unsigned int zip_available_devs;
|
|
|
|
enum hisi_zip_alg_type {
|
|
HZIP_ALG_TYPE_COMP = 0,
|
|
HZIP_ALG_TYPE_DECOMP = 1,
|
|
};
|
|
|
|
enum {
|
|
HZIP_QPC_COMP,
|
|
HZIP_QPC_DECOMP,
|
|
HZIP_CTX_Q_NUM
|
|
};
|
|
|
|
#define COMP_NAME_TO_TYPE(alg_name) \
|
|
(!strcmp((alg_name), "deflate") ? HZIP_ALG_TYPE_DEFLATE : 0)
|
|
|
|
struct hisi_zip_req {
|
|
struct acomp_req *req;
|
|
struct hisi_acc_hw_sgl *hw_src;
|
|
struct hisi_acc_hw_sgl *hw_dst;
|
|
dma_addr_t dma_src;
|
|
dma_addr_t dma_dst;
|
|
u16 req_id;
|
|
};
|
|
|
|
struct hisi_zip_req_q {
|
|
struct hisi_zip_req *q;
|
|
unsigned long *req_bitmap;
|
|
rwlock_t req_lock;
|
|
u16 size;
|
|
};
|
|
|
|
struct hisi_zip_qp_ctx {
|
|
struct hisi_qp *qp;
|
|
struct hisi_zip_req_q req_q;
|
|
struct hisi_acc_sgl_pool *sgl_pool;
|
|
struct hisi_zip *zip_dev;
|
|
struct hisi_zip_ctx *ctx;
|
|
};
|
|
|
|
struct hisi_zip_sqe_ops {
|
|
u8 sqe_type;
|
|
void (*fill_addr)(struct hisi_zip_sqe *sqe, struct hisi_zip_req *req);
|
|
void (*fill_buf_size)(struct hisi_zip_sqe *sqe, struct hisi_zip_req *req);
|
|
void (*fill_buf_type)(struct hisi_zip_sqe *sqe, u8 buf_type);
|
|
void (*fill_req_type)(struct hisi_zip_sqe *sqe, u8 req_type);
|
|
void (*fill_tag)(struct hisi_zip_sqe *sqe, struct hisi_zip_req *req);
|
|
void (*fill_sqe_type)(struct hisi_zip_sqe *sqe, u8 sqe_type);
|
|
u32 (*get_tag)(struct hisi_zip_sqe *sqe);
|
|
u32 (*get_status)(struct hisi_zip_sqe *sqe);
|
|
u32 (*get_dstlen)(struct hisi_zip_sqe *sqe);
|
|
};
|
|
|
|
struct hisi_zip_ctx {
|
|
struct hisi_zip_qp_ctx qp_ctx[HZIP_CTX_Q_NUM];
|
|
const struct hisi_zip_sqe_ops *ops;
|
|
};
|
|
|
|
static int sgl_sge_nr_set(const char *val, const struct kernel_param *kp)
|
|
{
|
|
int ret;
|
|
u16 n;
|
|
|
|
if (!val)
|
|
return -EINVAL;
|
|
|
|
ret = kstrtou16(val, 10, &n);
|
|
if (ret || n == 0 || n > HISI_ACC_SGL_SGE_NR_MAX)
|
|
return -EINVAL;
|
|
|
|
return param_set_ushort(val, kp);
|
|
}
|
|
|
|
static const struct kernel_param_ops sgl_sge_nr_ops = {
|
|
.set = sgl_sge_nr_set,
|
|
.get = param_get_ushort,
|
|
};
|
|
|
|
static u16 sgl_sge_nr = HZIP_SGL_SGE_NR;
|
|
module_param_cb(sgl_sge_nr, &sgl_sge_nr_ops, &sgl_sge_nr, 0444);
|
|
MODULE_PARM_DESC(sgl_sge_nr, "Number of sge in sgl(1-255)");
|
|
|
|
static struct hisi_zip_req *hisi_zip_create_req(struct hisi_zip_qp_ctx *qp_ctx,
|
|
struct acomp_req *req)
|
|
{
|
|
struct hisi_zip_req_q *req_q = &qp_ctx->req_q;
|
|
struct hisi_zip_req *q = req_q->q;
|
|
struct hisi_zip_req *req_cache;
|
|
int req_id;
|
|
|
|
write_lock(&req_q->req_lock);
|
|
|
|
req_id = find_first_zero_bit(req_q->req_bitmap, req_q->size);
|
|
if (req_id >= req_q->size) {
|
|
write_unlock(&req_q->req_lock);
|
|
dev_dbg(&qp_ctx->qp->qm->pdev->dev, "req cache is full!\n");
|
|
return ERR_PTR(-EAGAIN);
|
|
}
|
|
set_bit(req_id, req_q->req_bitmap);
|
|
|
|
write_unlock(&req_q->req_lock);
|
|
|
|
req_cache = q + req_id;
|
|
req_cache->req_id = req_id;
|
|
req_cache->req = req;
|
|
|
|
return req_cache;
|
|
}
|
|
|
|
static void hisi_zip_remove_req(struct hisi_zip_qp_ctx *qp_ctx,
|
|
struct hisi_zip_req *req)
|
|
{
|
|
struct hisi_zip_req_q *req_q = &qp_ctx->req_q;
|
|
|
|
write_lock(&req_q->req_lock);
|
|
clear_bit(req->req_id, req_q->req_bitmap);
|
|
write_unlock(&req_q->req_lock);
|
|
}
|
|
|
|
static void hisi_zip_fill_addr(struct hisi_zip_sqe *sqe, struct hisi_zip_req *req)
|
|
{
|
|
sqe->source_addr_l = lower_32_bits(req->dma_src);
|
|
sqe->source_addr_h = upper_32_bits(req->dma_src);
|
|
sqe->dest_addr_l = lower_32_bits(req->dma_dst);
|
|
sqe->dest_addr_h = upper_32_bits(req->dma_dst);
|
|
}
|
|
|
|
static void hisi_zip_fill_buf_size(struct hisi_zip_sqe *sqe, struct hisi_zip_req *req)
|
|
{
|
|
struct acomp_req *a_req = req->req;
|
|
|
|
sqe->input_data_length = a_req->slen;
|
|
sqe->dest_avail_out = a_req->dlen;
|
|
}
|
|
|
|
static void hisi_zip_fill_buf_type(struct hisi_zip_sqe *sqe, u8 buf_type)
|
|
{
|
|
u32 val;
|
|
|
|
val = sqe->dw9 & ~HZIP_BUF_TYPE_M;
|
|
val |= FIELD_PREP(HZIP_BUF_TYPE_M, buf_type);
|
|
sqe->dw9 = val;
|
|
}
|
|
|
|
static void hisi_zip_fill_req_type(struct hisi_zip_sqe *sqe, u8 req_type)
|
|
{
|
|
u32 val;
|
|
|
|
val = sqe->dw9 & ~HZIP_REQ_TYPE_M;
|
|
val |= FIELD_PREP(HZIP_REQ_TYPE_M, req_type);
|
|
sqe->dw9 = val;
|
|
}
|
|
|
|
static void hisi_zip_fill_tag(struct hisi_zip_sqe *sqe, struct hisi_zip_req *req)
|
|
{
|
|
sqe->dw26 = req->req_id;
|
|
}
|
|
|
|
static void hisi_zip_fill_sqe_type(struct hisi_zip_sqe *sqe, u8 sqe_type)
|
|
{
|
|
u32 val;
|
|
|
|
val = sqe->dw7 & ~HZIP_SQE_TYPE_M;
|
|
val |= FIELD_PREP(HZIP_SQE_TYPE_M, sqe_type);
|
|
sqe->dw7 = val;
|
|
}
|
|
|
|
static void hisi_zip_fill_sqe(struct hisi_zip_ctx *ctx, struct hisi_zip_sqe *sqe,
|
|
u8 req_type, struct hisi_zip_req *req)
|
|
{
|
|
const struct hisi_zip_sqe_ops *ops = ctx->ops;
|
|
|
|
memset(sqe, 0, sizeof(struct hisi_zip_sqe));
|
|
|
|
ops->fill_addr(sqe, req);
|
|
ops->fill_buf_size(sqe, req);
|
|
ops->fill_buf_type(sqe, HZIP_SGL);
|
|
ops->fill_req_type(sqe, req_type);
|
|
ops->fill_tag(sqe, req);
|
|
ops->fill_sqe_type(sqe, ops->sqe_type);
|
|
}
|
|
|
|
static int hisi_zip_do_work(struct hisi_zip_qp_ctx *qp_ctx,
|
|
struct hisi_zip_req *req)
|
|
{
|
|
struct hisi_acc_sgl_pool *pool = qp_ctx->sgl_pool;
|
|
struct hisi_zip_dfx *dfx = &qp_ctx->zip_dev->dfx;
|
|
struct acomp_req *a_req = req->req;
|
|
struct hisi_qp *qp = qp_ctx->qp;
|
|
struct device *dev = &qp->qm->pdev->dev;
|
|
struct hisi_zip_sqe zip_sqe;
|
|
int ret;
|
|
|
|
if (unlikely(!a_req->src || !a_req->slen || !a_req->dst || !a_req->dlen))
|
|
return -EINVAL;
|
|
|
|
req->hw_src = hisi_acc_sg_buf_map_to_hw_sgl(dev, a_req->src, pool,
|
|
req->req_id << 1, &req->dma_src);
|
|
if (IS_ERR(req->hw_src)) {
|
|
dev_err(dev, "failed to map the src buffer to hw sgl (%ld)!\n",
|
|
PTR_ERR(req->hw_src));
|
|
return PTR_ERR(req->hw_src);
|
|
}
|
|
|
|
req->hw_dst = hisi_acc_sg_buf_map_to_hw_sgl(dev, a_req->dst, pool,
|
|
(req->req_id << 1) + 1,
|
|
&req->dma_dst);
|
|
if (IS_ERR(req->hw_dst)) {
|
|
ret = PTR_ERR(req->hw_dst);
|
|
dev_err(dev, "failed to map the dst buffer to hw slg (%d)!\n",
|
|
ret);
|
|
goto err_unmap_input;
|
|
}
|
|
|
|
hisi_zip_fill_sqe(qp_ctx->ctx, &zip_sqe, qp->req_type, req);
|
|
|
|
/* send command to start a task */
|
|
atomic64_inc(&dfx->send_cnt);
|
|
ret = hisi_qp_send(qp, &zip_sqe);
|
|
if (unlikely(ret < 0)) {
|
|
atomic64_inc(&dfx->send_busy_cnt);
|
|
ret = -EAGAIN;
|
|
dev_dbg_ratelimited(dev, "failed to send request!\n");
|
|
goto err_unmap_output;
|
|
}
|
|
|
|
return -EINPROGRESS;
|
|
|
|
err_unmap_output:
|
|
hisi_acc_sg_buf_unmap(dev, a_req->dst, req->hw_dst);
|
|
err_unmap_input:
|
|
hisi_acc_sg_buf_unmap(dev, a_req->src, req->hw_src);
|
|
return ret;
|
|
}
|
|
|
|
static u32 hisi_zip_get_tag(struct hisi_zip_sqe *sqe)
|
|
{
|
|
return sqe->dw26;
|
|
}
|
|
|
|
static u32 hisi_zip_get_status(struct hisi_zip_sqe *sqe)
|
|
{
|
|
return sqe->dw3 & HZIP_BD_STATUS_M;
|
|
}
|
|
|
|
static u32 hisi_zip_get_dstlen(struct hisi_zip_sqe *sqe)
|
|
{
|
|
return sqe->produced;
|
|
}
|
|
|
|
static void hisi_zip_acomp_cb(struct hisi_qp *qp, void *data)
|
|
{
|
|
struct hisi_zip_qp_ctx *qp_ctx = qp->qp_ctx;
|
|
const struct hisi_zip_sqe_ops *ops = qp_ctx->ctx->ops;
|
|
struct hisi_zip_dfx *dfx = &qp_ctx->zip_dev->dfx;
|
|
struct hisi_zip_req_q *req_q = &qp_ctx->req_q;
|
|
struct device *dev = &qp->qm->pdev->dev;
|
|
struct hisi_zip_sqe *sqe = data;
|
|
u32 tag = ops->get_tag(sqe);
|
|
struct hisi_zip_req *req = req_q->q + tag;
|
|
struct acomp_req *acomp_req = req->req;
|
|
int err = 0;
|
|
u32 status;
|
|
|
|
atomic64_inc(&dfx->recv_cnt);
|
|
status = ops->get_status(sqe);
|
|
if (unlikely(status != 0 && status != HZIP_NC_ERR)) {
|
|
dev_err(dev, "%scompress fail in qp%u: %u, output: %u\n",
|
|
(qp->alg_type == 0) ? "" : "de", qp->qp_id, status,
|
|
sqe->produced);
|
|
atomic64_inc(&dfx->err_bd_cnt);
|
|
err = -EIO;
|
|
}
|
|
|
|
hisi_acc_sg_buf_unmap(dev, acomp_req->src, req->hw_src);
|
|
hisi_acc_sg_buf_unmap(dev, acomp_req->dst, req->hw_dst);
|
|
|
|
acomp_req->dlen = ops->get_dstlen(sqe);
|
|
|
|
if (acomp_req->base.complete)
|
|
acomp_request_complete(acomp_req, err);
|
|
|
|
hisi_zip_remove_req(qp_ctx, req);
|
|
}
|
|
|
|
static int hisi_zip_acompress(struct acomp_req *acomp_req)
|
|
{
|
|
struct hisi_zip_ctx *ctx = crypto_tfm_ctx(acomp_req->base.tfm);
|
|
struct hisi_zip_qp_ctx *qp_ctx = &ctx->qp_ctx[HZIP_QPC_COMP];
|
|
struct device *dev = &qp_ctx->qp->qm->pdev->dev;
|
|
struct hisi_zip_req *req;
|
|
int ret;
|
|
|
|
req = hisi_zip_create_req(qp_ctx, acomp_req);
|
|
if (IS_ERR(req))
|
|
return PTR_ERR(req);
|
|
|
|
ret = hisi_zip_do_work(qp_ctx, req);
|
|
if (unlikely(ret != -EINPROGRESS)) {
|
|
dev_info_ratelimited(dev, "failed to do compress (%d)!\n", ret);
|
|
hisi_zip_remove_req(qp_ctx, req);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int hisi_zip_adecompress(struct acomp_req *acomp_req)
|
|
{
|
|
struct hisi_zip_ctx *ctx = crypto_tfm_ctx(acomp_req->base.tfm);
|
|
struct hisi_zip_qp_ctx *qp_ctx = &ctx->qp_ctx[HZIP_QPC_DECOMP];
|
|
struct device *dev = &qp_ctx->qp->qm->pdev->dev;
|
|
struct hisi_zip_req *req;
|
|
int ret;
|
|
|
|
req = hisi_zip_create_req(qp_ctx, acomp_req);
|
|
if (IS_ERR(req))
|
|
return PTR_ERR(req);
|
|
|
|
ret = hisi_zip_do_work(qp_ctx, req);
|
|
if (unlikely(ret != -EINPROGRESS)) {
|
|
dev_info_ratelimited(dev, "failed to do decompress (%d)!\n",
|
|
ret);
|
|
hisi_zip_remove_req(qp_ctx, req);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int hisi_zip_start_qp(struct hisi_qp *qp, struct hisi_zip_qp_ctx *qp_ctx,
|
|
int alg_type, int req_type)
|
|
{
|
|
struct device *dev = &qp->qm->pdev->dev;
|
|
int ret;
|
|
|
|
qp->req_type = req_type;
|
|
qp->alg_type = alg_type;
|
|
qp->qp_ctx = qp_ctx;
|
|
|
|
ret = hisi_qm_start_qp(qp, 0);
|
|
if (ret < 0) {
|
|
dev_err(dev, "failed to start qp (%d)!\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
qp_ctx->qp = qp;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void hisi_zip_release_qp(struct hisi_zip_qp_ctx *qp_ctx)
|
|
{
|
|
hisi_qm_stop_qp(qp_ctx->qp);
|
|
hisi_qm_free_qps(&qp_ctx->qp, 1);
|
|
}
|
|
|
|
static const struct hisi_zip_sqe_ops hisi_zip_ops = {
|
|
.sqe_type = 0x3,
|
|
.fill_addr = hisi_zip_fill_addr,
|
|
.fill_buf_size = hisi_zip_fill_buf_size,
|
|
.fill_buf_type = hisi_zip_fill_buf_type,
|
|
.fill_req_type = hisi_zip_fill_req_type,
|
|
.fill_tag = hisi_zip_fill_tag,
|
|
.fill_sqe_type = hisi_zip_fill_sqe_type,
|
|
.get_tag = hisi_zip_get_tag,
|
|
.get_status = hisi_zip_get_status,
|
|
.get_dstlen = hisi_zip_get_dstlen,
|
|
};
|
|
|
|
static int hisi_zip_ctx_init(struct hisi_zip_ctx *hisi_zip_ctx, u8 req_type, int node)
|
|
{
|
|
struct hisi_qp *qps[HZIP_CTX_Q_NUM] = { NULL };
|
|
struct hisi_zip_qp_ctx *qp_ctx;
|
|
struct hisi_zip *hisi_zip;
|
|
int ret, i, j;
|
|
|
|
ret = zip_create_qps(qps, HZIP_CTX_Q_NUM, node);
|
|
if (ret) {
|
|
pr_err("failed to create zip qps (%d)!\n", ret);
|
|
return -ENODEV;
|
|
}
|
|
|
|
hisi_zip = container_of(qps[0]->qm, struct hisi_zip, qm);
|
|
|
|
for (i = 0; i < HZIP_CTX_Q_NUM; i++) {
|
|
/* alg_type = 0 for compress, 1 for decompress in hw sqe */
|
|
qp_ctx = &hisi_zip_ctx->qp_ctx[i];
|
|
qp_ctx->ctx = hisi_zip_ctx;
|
|
ret = hisi_zip_start_qp(qps[i], qp_ctx, i, req_type);
|
|
if (ret) {
|
|
for (j = i - 1; j >= 0; j--)
|
|
hisi_qm_stop_qp(hisi_zip_ctx->qp_ctx[j].qp);
|
|
|
|
hisi_qm_free_qps(qps, HZIP_CTX_Q_NUM);
|
|
return ret;
|
|
}
|
|
|
|
qp_ctx->zip_dev = hisi_zip;
|
|
}
|
|
|
|
hisi_zip_ctx->ops = &hisi_zip_ops;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void hisi_zip_ctx_exit(struct hisi_zip_ctx *hisi_zip_ctx)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < HZIP_CTX_Q_NUM; i++)
|
|
hisi_zip_release_qp(&hisi_zip_ctx->qp_ctx[i]);
|
|
}
|
|
|
|
static int hisi_zip_create_req_q(struct hisi_zip_ctx *ctx)
|
|
{
|
|
u16 q_depth = ctx->qp_ctx[0].qp->sq_depth;
|
|
struct hisi_zip_req_q *req_q;
|
|
int i, ret;
|
|
|
|
for (i = 0; i < HZIP_CTX_Q_NUM; i++) {
|
|
req_q = &ctx->qp_ctx[i].req_q;
|
|
req_q->size = q_depth;
|
|
|
|
req_q->req_bitmap = bitmap_zalloc(req_q->size, GFP_KERNEL);
|
|
if (!req_q->req_bitmap) {
|
|
ret = -ENOMEM;
|
|
if (i == 0)
|
|
return ret;
|
|
|
|
goto err_free_comp_q;
|
|
}
|
|
rwlock_init(&req_q->req_lock);
|
|
|
|
req_q->q = kcalloc(req_q->size, sizeof(struct hisi_zip_req),
|
|
GFP_KERNEL);
|
|
if (!req_q->q) {
|
|
ret = -ENOMEM;
|
|
if (i == 0)
|
|
goto err_free_comp_bitmap;
|
|
else
|
|
goto err_free_decomp_bitmap;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_free_decomp_bitmap:
|
|
bitmap_free(ctx->qp_ctx[HZIP_QPC_DECOMP].req_q.req_bitmap);
|
|
err_free_comp_q:
|
|
kfree(ctx->qp_ctx[HZIP_QPC_COMP].req_q.q);
|
|
err_free_comp_bitmap:
|
|
bitmap_free(ctx->qp_ctx[HZIP_QPC_COMP].req_q.req_bitmap);
|
|
return ret;
|
|
}
|
|
|
|
static void hisi_zip_release_req_q(struct hisi_zip_ctx *ctx)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < HZIP_CTX_Q_NUM; i++) {
|
|
kfree(ctx->qp_ctx[i].req_q.q);
|
|
bitmap_free(ctx->qp_ctx[i].req_q.req_bitmap);
|
|
}
|
|
}
|
|
|
|
static int hisi_zip_create_sgl_pool(struct hisi_zip_ctx *ctx)
|
|
{
|
|
u16 q_depth = ctx->qp_ctx[0].qp->sq_depth;
|
|
struct hisi_zip_qp_ctx *tmp;
|
|
struct device *dev;
|
|
int i;
|
|
|
|
for (i = 0; i < HZIP_CTX_Q_NUM; i++) {
|
|
tmp = &ctx->qp_ctx[i];
|
|
dev = &tmp->qp->qm->pdev->dev;
|
|
tmp->sgl_pool = hisi_acc_create_sgl_pool(dev, q_depth << 1,
|
|
sgl_sge_nr);
|
|
if (IS_ERR(tmp->sgl_pool)) {
|
|
if (i == 1)
|
|
goto err_free_sgl_pool0;
|
|
return -ENOMEM;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_free_sgl_pool0:
|
|
hisi_acc_free_sgl_pool(&ctx->qp_ctx[HZIP_QPC_COMP].qp->qm->pdev->dev,
|
|
ctx->qp_ctx[HZIP_QPC_COMP].sgl_pool);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
static void hisi_zip_release_sgl_pool(struct hisi_zip_ctx *ctx)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < HZIP_CTX_Q_NUM; i++)
|
|
hisi_acc_free_sgl_pool(&ctx->qp_ctx[i].qp->qm->pdev->dev,
|
|
ctx->qp_ctx[i].sgl_pool);
|
|
}
|
|
|
|
static void hisi_zip_set_acomp_cb(struct hisi_zip_ctx *ctx,
|
|
void (*fn)(struct hisi_qp *, void *))
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < HZIP_CTX_Q_NUM; i++)
|
|
ctx->qp_ctx[i].qp->req_cb = fn;
|
|
}
|
|
|
|
static int hisi_zip_acomp_init(struct crypto_acomp *tfm)
|
|
{
|
|
const char *alg_name = crypto_tfm_alg_name(&tfm->base);
|
|
struct hisi_zip_ctx *ctx = crypto_tfm_ctx(&tfm->base);
|
|
struct device *dev;
|
|
int ret;
|
|
|
|
ret = hisi_zip_ctx_init(ctx, COMP_NAME_TO_TYPE(alg_name), tfm->base.node);
|
|
if (ret) {
|
|
pr_err("failed to init ctx (%d)!\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
dev = &ctx->qp_ctx[0].qp->qm->pdev->dev;
|
|
|
|
ret = hisi_zip_create_req_q(ctx);
|
|
if (ret) {
|
|
dev_err(dev, "failed to create request queue (%d)!\n", ret);
|
|
goto err_ctx_exit;
|
|
}
|
|
|
|
ret = hisi_zip_create_sgl_pool(ctx);
|
|
if (ret) {
|
|
dev_err(dev, "failed to create sgl pool (%d)!\n", ret);
|
|
goto err_release_req_q;
|
|
}
|
|
|
|
hisi_zip_set_acomp_cb(ctx, hisi_zip_acomp_cb);
|
|
|
|
return 0;
|
|
|
|
err_release_req_q:
|
|
hisi_zip_release_req_q(ctx);
|
|
err_ctx_exit:
|
|
hisi_zip_ctx_exit(ctx);
|
|
return ret;
|
|
}
|
|
|
|
static void hisi_zip_acomp_exit(struct crypto_acomp *tfm)
|
|
{
|
|
struct hisi_zip_ctx *ctx = crypto_tfm_ctx(&tfm->base);
|
|
|
|
hisi_zip_set_acomp_cb(ctx, NULL);
|
|
hisi_zip_release_sgl_pool(ctx);
|
|
hisi_zip_release_req_q(ctx);
|
|
hisi_zip_ctx_exit(ctx);
|
|
}
|
|
|
|
static struct acomp_alg hisi_zip_acomp_deflate = {
|
|
.init = hisi_zip_acomp_init,
|
|
.exit = hisi_zip_acomp_exit,
|
|
.compress = hisi_zip_acompress,
|
|
.decompress = hisi_zip_adecompress,
|
|
.base = {
|
|
.cra_name = "deflate",
|
|
.cra_driver_name = "hisi-deflate-acomp",
|
|
.cra_module = THIS_MODULE,
|
|
.cra_priority = HZIP_ALG_PRIORITY,
|
|
.cra_ctxsize = sizeof(struct hisi_zip_ctx),
|
|
}
|
|
};
|
|
|
|
static int hisi_zip_register_deflate(struct hisi_qm *qm)
|
|
{
|
|
int ret;
|
|
|
|
if (!hisi_zip_alg_support(qm, HZIP_ALG_DEFLATE))
|
|
return 0;
|
|
|
|
ret = crypto_register_acomp(&hisi_zip_acomp_deflate);
|
|
if (ret)
|
|
dev_err(&qm->pdev->dev, "failed to register to deflate (%d)!\n", ret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void hisi_zip_unregister_deflate(struct hisi_qm *qm)
|
|
{
|
|
if (!hisi_zip_alg_support(qm, HZIP_ALG_DEFLATE))
|
|
return;
|
|
|
|
crypto_unregister_acomp(&hisi_zip_acomp_deflate);
|
|
}
|
|
|
|
int hisi_zip_register_to_crypto(struct hisi_qm *qm)
|
|
{
|
|
int ret = 0;
|
|
|
|
mutex_lock(&zip_algs_lock);
|
|
if (zip_available_devs++)
|
|
goto unlock;
|
|
|
|
ret = hisi_zip_register_deflate(qm);
|
|
if (ret)
|
|
zip_available_devs--;
|
|
|
|
unlock:
|
|
mutex_unlock(&zip_algs_lock);
|
|
return ret;
|
|
}
|
|
|
|
void hisi_zip_unregister_from_crypto(struct hisi_qm *qm)
|
|
{
|
|
mutex_lock(&zip_algs_lock);
|
|
if (--zip_available_devs)
|
|
goto unlock;
|
|
|
|
hisi_zip_unregister_deflate(qm);
|
|
|
|
unlock:
|
|
mutex_unlock(&zip_algs_lock);
|
|
}
|