Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- commit 433cd2c617bfbac27a02e40fbcce1713c84ce441
- Author: Zain Wang <zain.wang@rock-chips.com>
- Date: Wed Nov 25 13:43:32 2015 +0800
- crypto: rockchip - add crypto driver for rk3288
- Crypto driver support:
- ecb(aes) cbc(aes) ecb(des) cbc(des) ecb(des3_ede) cbc(des3_ede)
- You can alloc tags above in your case.
- And other algorithms and platforms will be added later on.
- Signed-off-by: Zain Wang <zain.wang@rock-chips.com>
- Tested-by: Heiko Stuebner <heiko@sntech.de>
- Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
- diff --git a/drivers/crypto/Kconfig b/drivers/crypto/Kconfig
- index 5357bc1..95dccde 100644
- --- a/drivers/crypto/Kconfig
- +++ b/drivers/crypto/Kconfig
- @@ -497,4 +497,15 @@ config CRYPTO_DEV_SUN4I_SS
- To compile this driver as a module, choose M here: the module
- will be called sun4i-ss.
- +config CRYPTO_DEV_ROCKCHIP
- + tristate "Rockchip's Cryptographic Engine driver"
- + depends on OF && ARCH_ROCKCHIP
- + select CRYPTO_AES
- + select CRYPTO_DES
- + select CRYPTO_BLKCIPHER
- +
- + help
- + This driver interfaces with the hardware crypto accelerator.
- + Supporting cbc/ecb chainmode, and aes/des/des3_ede cipher mode.
- +
- endif # CRYPTO_HW
- diff --git a/drivers/crypto/Makefile b/drivers/crypto/Makefile
- index c3ced6fb..713de9d 100644
- --- a/drivers/crypto/Makefile
- +++ b/drivers/crypto/Makefile
- @@ -29,3 +29,4 @@ obj-$(CONFIG_CRYPTO_DEV_QAT) += qat/
- obj-$(CONFIG_CRYPTO_DEV_QCE) += qce/
- obj-$(CONFIG_CRYPTO_DEV_VMX) += vmx/
- obj-$(CONFIG_CRYPTO_DEV_SUN4I_SS) += sunxi-ss/
- +obj-$(CONFIG_CRYPTO_DEV_ROCKCHIP) += rockchip/
- diff --git a/drivers/crypto/rockchip/Makefile b/drivers/crypto/rockchip/Makefile
- new file mode 100644
- index 0000000..7051c6c
- --- /dev/null
- +++ b/drivers/crypto/rockchip/Makefile
- @@ -0,0 +1,3 @@
- +obj-$(CONFIG_CRYPTO_DEV_ROCKCHIP) += rk_crypto.o
- +rk_crypto-objs := rk3288_crypto.o \
- + rk3288_crypto_ablkcipher.o \
- diff --git a/drivers/crypto/rockchip/rk3288_crypto.c b/drivers/crypto/rockchip/rk3288_crypto.c
- new file mode 100644
- index 0000000..6b72f8d
- --- /dev/null
- +++ b/drivers/crypto/rockchip/rk3288_crypto.c
- @@ -0,0 +1,393 @@
- +/*
- + * Crypto acceleration support for Rockchip RK3288
- + *
- + * Copyright (c) 2015, Fuzhou Rockchip Electronics Co., Ltd
- + *
- + * Author: Zain Wang <zain.wang@rock-chips.com>
- + *
- + * This program is free software; you can redistribute it and/or modify it
- + * under the terms and conditions of the GNU General Public License,
- + * version 2, as published by the Free Software Foundation.
- + *
- + * Some ideas are from marvell-cesa.c and s5p-sss.c driver.
- + */
- +
- +#include "rk3288_crypto.h"
- +#include <linux/module.h>
- +#include <linux/platform_device.h>
- +#include <linux/of.h>
- +#include <linux/clk.h>
- +#include <linux/crypto.h>
- +#include <linux/reset.h>
- +
- +static int rk_crypto_enable_clk(struct rk_crypto_info *dev)
- +{
- + int err;
- +
- + err = clk_prepare_enable(dev->sclk);
- + if (err) {
- + dev_err(dev->dev, "[%s:%d], Couldn't enable clock sclk\n",
- + __func__, __LINE__);
- + goto err_return;
- + }
- + err = clk_prepare_enable(dev->aclk);
- + if (err) {
- + dev_err(dev->dev, "[%s:%d], Couldn't enable clock aclk\n",
- + __func__, __LINE__);
- + goto err_aclk;
- + }
- + err = clk_prepare_enable(dev->hclk);
- + if (err) {
- + dev_err(dev->dev, "[%s:%d], Couldn't enable clock hclk\n",
- + __func__, __LINE__);
- + goto err_hclk;
- + }
- + err = clk_prepare_enable(dev->dmaclk);
- + if (err) {
- + dev_err(dev->dev, "[%s:%d], Couldn't enable clock dmaclk\n",
- + __func__, __LINE__);
- + goto err_dmaclk;
- + }
- + return err;
- +err_dmaclk:
- + clk_disable_unprepare(dev->hclk);
- +err_hclk:
- + clk_disable_unprepare(dev->aclk);
- +err_aclk:
- + clk_disable_unprepare(dev->sclk);
- +err_return:
- + return err;
- +}
- +
- +static void rk_crypto_disable_clk(struct rk_crypto_info *dev)
- +{
- + clk_disable_unprepare(dev->dmaclk);
- + clk_disable_unprepare(dev->hclk);
- + clk_disable_unprepare(dev->aclk);
- + clk_disable_unprepare(dev->sclk);
- +}
- +
- +static int check_alignment(struct scatterlist *sg_src,
- + struct scatterlist *sg_dst,
- + int align_mask)
- +{
- + int in, out, align;
- +
- + in = IS_ALIGNED((uint32_t)sg_src->offset, 4) &&
- + IS_ALIGNED((uint32_t)sg_src->length, align_mask);
- + if (!sg_dst)
- + return in;
- + out = IS_ALIGNED((uint32_t)sg_dst->offset, 4) &&
- + IS_ALIGNED((uint32_t)sg_dst->length, align_mask);
- + align = in && out;
- +
- + return (align && (sg_src->length == sg_dst->length));
- +}
- +
- +static int rk_load_data(struct rk_crypto_info *dev,
- + struct scatterlist *sg_src,
- + struct scatterlist *sg_dst)
- +{
- + unsigned int count;
- +
- + dev->aligned = dev->aligned ?
- + check_alignment(sg_src, sg_dst, dev->align_size) :
- + dev->aligned;
- + if (dev->aligned) {
- + count = min(dev->left_bytes, sg_src->length);
- + dev->left_bytes -= count;
- +
- + if (!dma_map_sg(dev->dev, sg_src, 1, DMA_TO_DEVICE)) {
- + dev_err(dev->dev, "[%s:%d] dma_map_sg(src) error\n",
- + __func__, __LINE__);
- + return -EINVAL;
- + }
- + dev->addr_in = sg_dma_address(sg_src);
- +
- + if (sg_dst) {
- + if (!dma_map_sg(dev->dev, sg_dst, 1, DMA_FROM_DEVICE)) {
- + dev_err(dev->dev,
- + "[%s:%d] dma_map_sg(dst) error\n",
- + __func__, __LINE__);
- + dma_unmap_sg(dev->dev, sg_src, 1,
- + DMA_TO_DEVICE);
- + return -EINVAL;
- + }
- + dev->addr_out = sg_dma_address(sg_dst);
- + }
- + } else {
- + count = (dev->left_bytes > PAGE_SIZE) ?
- + PAGE_SIZE : dev->left_bytes;
- +
- + if (!sg_pcopy_to_buffer(dev->first, dev->nents,
- + dev->addr_vir, count,
- + dev->total - dev->left_bytes)) {
- + dev_err(dev->dev, "[%s:%d] pcopy err\n",
- + __func__, __LINE__);
- + return -EINVAL;
- + }
- + dev->left_bytes -= count;
- + sg_init_one(&dev->sg_tmp, dev->addr_vir, count);
- + if (!dma_map_sg(dev->dev, &dev->sg_tmp, 1, DMA_TO_DEVICE)) {
- + dev_err(dev->dev, "[%s:%d] dma_map_sg(sg_tmp) error\n",
- + __func__, __LINE__);
- + return -ENOMEM;
- + }
- + dev->addr_in = sg_dma_address(&dev->sg_tmp);
- +
- + if (sg_dst) {
- + if (!dma_map_sg(dev->dev, &dev->sg_tmp, 1,
- + DMA_FROM_DEVICE)) {
- + dev_err(dev->dev,
- + "[%s:%d] dma_map_sg(sg_tmp) error\n",
- + __func__, __LINE__);
- + dma_unmap_sg(dev->dev, &dev->sg_tmp, 1,
- + DMA_TO_DEVICE);
- + return -ENOMEM;
- + }
- + dev->addr_out = sg_dma_address(&dev->sg_tmp);
- + }
- + }
- + dev->count = count;
- + return 0;
- +}
- +
- +static void rk_unload_data(struct rk_crypto_info *dev)
- +{
- + struct scatterlist *sg_in, *sg_out;
- +
- + sg_in = dev->aligned ? dev->sg_src : &dev->sg_tmp;
- + dma_unmap_sg(dev->dev, sg_in, 1, DMA_TO_DEVICE);
- +
- + if (dev->sg_dst) {
- + sg_out = dev->aligned ? dev->sg_dst : &dev->sg_tmp;
- + dma_unmap_sg(dev->dev, sg_out, 1, DMA_FROM_DEVICE);
- + }
- +}
- +
- +static irqreturn_t rk_crypto_irq_handle(int irq, void *dev_id)
- +{
- + struct rk_crypto_info *dev = platform_get_drvdata(dev_id);
- + u32 interrupt_status;
- + int err = 0;
- +
- + spin_lock(&dev->lock);
- + interrupt_status = CRYPTO_READ(dev, RK_CRYPTO_INTSTS);
- + CRYPTO_WRITE(dev, RK_CRYPTO_INTSTS, interrupt_status);
- + if (interrupt_status & 0x0a) {
- + dev_warn(dev->dev, "DMA Error\n");
- + err = -EFAULT;
- + } else if (interrupt_status & 0x05) {
- + err = dev->update(dev);
- + }
- + if (err)
- + dev->complete(dev, err);
- + spin_unlock(&dev->lock);
- + return IRQ_HANDLED;
- +}
- +
- +static void rk_crypto_tasklet_cb(unsigned long data)
- +{
- + struct rk_crypto_info *dev = (struct rk_crypto_info *)data;
- + struct crypto_async_request *async_req, *backlog;
- + int err = 0;
- +
- + spin_lock(&dev->lock);
- + backlog = crypto_get_backlog(&dev->queue);
- + async_req = crypto_dequeue_request(&dev->queue);
- + spin_unlock(&dev->lock);
- + if (!async_req) {
- + dev_err(dev->dev, "async_req is NULL !!\n");
- + return;
- + }
- + if (backlog) {
- + backlog->complete(backlog, -EINPROGRESS);
- + backlog = NULL;
- + }
- +
- + if (crypto_tfm_alg_type(async_req->tfm) == CRYPTO_ALG_TYPE_ABLKCIPHER)
- + dev->ablk_req = ablkcipher_request_cast(async_req);
- + err = dev->start(dev);
- + if (err)
- + dev->complete(dev, err);
- +}
- +
- +static struct rk_crypto_tmp *rk_cipher_algs[] = {
- + &rk_ecb_aes_alg,
- + &rk_cbc_aes_alg,
- + &rk_ecb_des_alg,
- + &rk_cbc_des_alg,
- + &rk_ecb_des3_ede_alg,
- + &rk_cbc_des3_ede_alg,
- +};
- +
- +static int rk_crypto_register(struct rk_crypto_info *crypto_info)
- +{
- + unsigned int i, k;
- + int err = 0;
- +
- + for (i = 0; i < ARRAY_SIZE(rk_cipher_algs); i++) {
- + rk_cipher_algs[i]->dev = crypto_info;
- + err = crypto_register_alg(&rk_cipher_algs[i]->alg);
- + if (err)
- + goto err_cipher_algs;
- + }
- + return 0;
- +
- +err_cipher_algs:
- + for (k = 0; k < i; k++)
- + crypto_unregister_alg(&rk_cipher_algs[k]->alg);
- + return err;
- +}
- +
- +static void rk_crypto_unregister(void)
- +{
- + unsigned int i;
- +
- + for (i = 0; i < ARRAY_SIZE(rk_cipher_algs); i++)
- + crypto_unregister_alg(&rk_cipher_algs[i]->alg);
- +}
- +
- +static void rk_crypto_action(void *data)
- +{
- + struct rk_crypto_info *crypto_info = data;
- +
- + reset_control_assert(crypto_info->rst);
- +}
- +
- +static const struct of_device_id crypto_of_id_table[] = {
- + { .compatible = "rockchip,rk3288-crypto" },
- + {}
- +};
- +MODULE_DEVICE_TABLE(of, crypto_of_id_table);
- +
- +static int rk_crypto_probe(struct platform_device *pdev)
- +{
- + struct resource *res;
- + struct device *dev = &pdev->dev;
- + struct rk_crypto_info *crypto_info;
- + int err = 0;
- +
- + crypto_info = devm_kzalloc(&pdev->dev,
- + sizeof(*crypto_info), GFP_KERNEL);
- + if (!crypto_info) {
- + err = -ENOMEM;
- + goto err_crypto;
- + }
- +
- + crypto_info->rst = devm_reset_control_get(dev, "crypto-rst");
- + if (IS_ERR(crypto_info->rst)) {
- + err = PTR_ERR(crypto_info->rst);
- + goto err_crypto;
- + }
- +
- + reset_control_assert(crypto_info->rst);
- + usleep_range(10, 20);
- + reset_control_deassert(crypto_info->rst);
- +
- + err = devm_add_action(dev, rk_crypto_action, crypto_info);
- + if (err) {
- + reset_control_assert(crypto_info->rst);
- + goto err_crypto;
- + }
- +
- + spin_lock_init(&crypto_info->lock);
- +
- + res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
- + crypto_info->reg = devm_ioremap_resource(&pdev->dev, res);
- + if (IS_ERR(crypto_info->reg)) {
- + err = PTR_ERR(crypto_info->reg);
- + goto err_crypto;
- + }
- +
- + crypto_info->aclk = devm_clk_get(&pdev->dev, "aclk");
- + if (IS_ERR(crypto_info->aclk)) {
- + err = PTR_ERR(crypto_info->aclk);
- + goto err_crypto;
- + }
- +
- + crypto_info->hclk = devm_clk_get(&pdev->dev, "hclk");
- + if (IS_ERR(crypto_info->hclk)) {
- + err = PTR_ERR(crypto_info->hclk);
- + goto err_crypto;
- + }
- +
- + crypto_info->sclk = devm_clk_get(&pdev->dev, "sclk");
- + if (IS_ERR(crypto_info->sclk)) {
- + err = PTR_ERR(crypto_info->sclk);
- + goto err_crypto;
- + }
- +
- + crypto_info->dmaclk = devm_clk_get(&pdev->dev, "apb_pclk");
- + if (IS_ERR(crypto_info->dmaclk)) {
- + err = PTR_ERR(crypto_info->dmaclk);
- + goto err_crypto;
- + }
- +
- + crypto_info->irq = platform_get_irq(pdev, 0);
- + if (crypto_info->irq < 0) {
- + dev_warn(crypto_info->dev,
- + "control Interrupt is not available.\n");
- + err = crypto_info->irq;
- + goto err_crypto;
- + }
- +
- + err = devm_request_irq(&pdev->dev, crypto_info->irq,
- + rk_crypto_irq_handle, IRQF_SHARED,
- + "rk-crypto", pdev);
- +
- + if (err) {
- + dev_err(crypto_info->dev, "irq request failed.\n");
- + goto err_crypto;
- + }
- +
- + crypto_info->dev = &pdev->dev;
- + platform_set_drvdata(pdev, crypto_info);
- +
- + tasklet_init(&crypto_info->crypto_tasklet,
- + rk_crypto_tasklet_cb, (unsigned long)crypto_info);
- + crypto_init_queue(&crypto_info->queue, 50);
- +
- + crypto_info->enable_clk = rk_crypto_enable_clk;
- + crypto_info->disable_clk = rk_crypto_disable_clk;
- + crypto_info->load_data = rk_load_data;
- + crypto_info->unload_data = rk_unload_data;
- +
- + err = rk_crypto_register(crypto_info);
- + if (err) {
- + dev_err(dev, "err in register alg");
- + goto err_register_alg;
- + }
- +
- + dev_info(dev, "Crypto Accelerator successfully registered\n");
- + return 0;
- +
- +err_register_alg:
- + tasklet_kill(&crypto_info->crypto_tasklet);
- +err_crypto:
- + return err;
- +}
- +
- +static int rk_crypto_remove(struct platform_device *pdev)
- +{
- + struct rk_crypto_info *crypto_tmp = platform_get_drvdata(pdev);
- +
- + rk_crypto_unregister();
- + tasklet_kill(&crypto_tmp->crypto_tasklet);
- + return 0;
- +}
- +
- +static struct platform_driver crypto_driver = {
- + .probe = rk_crypto_probe,
- + .remove = rk_crypto_remove,
- + .driver = {
- + .name = "rk3288-crypto",
- + .of_match_table = crypto_of_id_table,
- + },
- +};
- +
- +module_platform_driver(crypto_driver);
- +
- +MODULE_AUTHOR("Zain Wang <zain.wang@rock-chips.com>");
- +MODULE_DESCRIPTION("Support for Rockchip's cryptographic engine");
- +MODULE_LICENSE("GPL");
- diff --git a/drivers/crypto/rockchip/rk3288_crypto.h b/drivers/crypto/rockchip/rk3288_crypto.h
- new file mode 100644
- index 0000000..e499c2c
- --- /dev/null
- +++ b/drivers/crypto/rockchip/rk3288_crypto.h
- @@ -0,0 +1,216 @@
- +#ifndef __RK3288_CRYPTO_H__
- +#define __RK3288_CRYPTO_H__
- +
- +#include <crypto/aes.h>
- +#include <crypto/des.h>
- +#include <crypto/algapi.h>
- +#include <linux/interrupt.h>
- +#include <linux/delay.h>
- +
- +#define _SBF(v, f) ((v) << (f))
- +
- +/* Crypto control registers*/
- +#define RK_CRYPTO_INTSTS 0x0000
- +#define RK_CRYPTO_PKA_DONE_INT BIT(5)
- +#define RK_CRYPTO_HASH_DONE_INT BIT(4)
- +#define RK_CRYPTO_HRDMA_ERR_INT BIT(3)
- +#define RK_CRYPTO_HRDMA_DONE_INT BIT(2)
- +#define RK_CRYPTO_BCDMA_ERR_INT BIT(1)
- +#define RK_CRYPTO_BCDMA_DONE_INT BIT(0)
- +
- +#define RK_CRYPTO_INTENA 0x0004
- +#define RK_CRYPTO_PKA_DONE_ENA BIT(5)
- +#define RK_CRYPTO_HASH_DONE_ENA BIT(4)
- +#define RK_CRYPTO_HRDMA_ERR_ENA BIT(3)
- +#define RK_CRYPTO_HRDMA_DONE_ENA BIT(2)
- +#define RK_CRYPTO_BCDMA_ERR_ENA BIT(1)
- +#define RK_CRYPTO_BCDMA_DONE_ENA BIT(0)
- +
- +#define RK_CRYPTO_CTRL 0x0008
- +#define RK_CRYPTO_WRITE_MASK _SBF(0xFFFF, 16)
- +#define RK_CRYPTO_TRNG_FLUSH BIT(9)
- +#define RK_CRYPTO_TRNG_START BIT(8)
- +#define RK_CRYPTO_PKA_FLUSH BIT(7)
- +#define RK_CRYPTO_HASH_FLUSH BIT(6)
- +#define RK_CRYPTO_BLOCK_FLUSH BIT(5)
- +#define RK_CRYPTO_PKA_START BIT(4)
- +#define RK_CRYPTO_HASH_START BIT(3)
- +#define RK_CRYPTO_BLOCK_START BIT(2)
- +#define RK_CRYPTO_TDES_START BIT(1)
- +#define RK_CRYPTO_AES_START BIT(0)
- +
- +#define RK_CRYPTO_CONF 0x000c
- +/* HASH Receive DMA Address Mode: fix | increment */
- +#define RK_CRYPTO_HR_ADDR_MODE BIT(8)
- +/* Block Transmit DMA Address Mode: fix | increment */
- +#define RK_CRYPTO_BT_ADDR_MODE BIT(7)
- +/* Block Receive DMA Address Mode: fix | increment */
- +#define RK_CRYPTO_BR_ADDR_MODE BIT(6)
- +#define RK_CRYPTO_BYTESWAP_HRFIFO BIT(5)
- +#define RK_CRYPTO_BYTESWAP_BTFIFO BIT(4)
- +#define RK_CRYPTO_BYTESWAP_BRFIFO BIT(3)
- +/* AES = 0 OR DES = 1 */
- +#define RK_CRYPTO_DESSEL BIT(2)
- +#define RK_CYYPTO_HASHINSEL_INDEPENDENT_SOURCE _SBF(0x00, 0)
- +#define RK_CYYPTO_HASHINSEL_BLOCK_CIPHER_INPUT _SBF(0x01, 0)
- +#define RK_CYYPTO_HASHINSEL_BLOCK_CIPHER_OUTPUT _SBF(0x02, 0)
- +
- +/* Block Receiving DMA Start Address Register */
- +#define RK_CRYPTO_BRDMAS 0x0010
- +/* Block Transmitting DMA Start Address Register */
- +#define RK_CRYPTO_BTDMAS 0x0014
- +/* Block Receiving DMA Length Register */
- +#define RK_CRYPTO_BRDMAL 0x0018
- +/* Hash Receiving DMA Start Address Register */
- +#define RK_CRYPTO_HRDMAS 0x001c
- +/* Hash Receiving DMA Length Register */
- +#define RK_CRYPTO_HRDMAL 0x0020
- +
- +/* AES registers */
- +#define RK_CRYPTO_AES_CTRL 0x0080
- +#define RK_CRYPTO_AES_BYTESWAP_CNT BIT(11)
- +#define RK_CRYPTO_AES_BYTESWAP_KEY BIT(10)
- +#define RK_CRYPTO_AES_BYTESWAP_IV BIT(9)
- +#define RK_CRYPTO_AES_BYTESWAP_DO BIT(8)
- +#define RK_CRYPTO_AES_BYTESWAP_DI BIT(7)
- +#define RK_CRYPTO_AES_KEY_CHANGE BIT(6)
- +#define RK_CRYPTO_AES_ECB_MODE _SBF(0x00, 4)
- +#define RK_CRYPTO_AES_CBC_MODE _SBF(0x01, 4)
- +#define RK_CRYPTO_AES_CTR_MODE _SBF(0x02, 4)
- +#define RK_CRYPTO_AES_128BIT_key _SBF(0x00, 2)
- +#define RK_CRYPTO_AES_192BIT_key _SBF(0x01, 2)
- +#define RK_CRYPTO_AES_256BIT_key _SBF(0x02, 2)
- +/* Slave = 0 / fifo = 1 */
- +#define RK_CRYPTO_AES_FIFO_MODE BIT(1)
- +/* Encryption = 0 , Decryption = 1 */
- +#define RK_CRYPTO_AES_DEC BIT(0)
- +
- +#define RK_CRYPTO_AES_STS 0x0084
- +#define RK_CRYPTO_AES_DONE BIT(0)
- +
- +/* AES Input Data 0-3 Register */
- +#define RK_CRYPTO_AES_DIN_0 0x0088
- +#define RK_CRYPTO_AES_DIN_1 0x008c
- +#define RK_CRYPTO_AES_DIN_2 0x0090
- +#define RK_CRYPTO_AES_DIN_3 0x0094
- +
- +/* AES output Data 0-3 Register */
- +#define RK_CRYPTO_AES_DOUT_0 0x0098
- +#define RK_CRYPTO_AES_DOUT_1 0x009c
- +#define RK_CRYPTO_AES_DOUT_2 0x00a0
- +#define RK_CRYPTO_AES_DOUT_3 0x00a4
- +
- +/* AES IV Data 0-3 Register */
- +#define RK_CRYPTO_AES_IV_0 0x00a8
- +#define RK_CRYPTO_AES_IV_1 0x00ac
- +#define RK_CRYPTO_AES_IV_2 0x00b0
- +#define RK_CRYPTO_AES_IV_3 0x00b4
- +
- +/* AES Key Data 0-3 Register */
- +#define RK_CRYPTO_AES_KEY_0 0x00b8
- +#define RK_CRYPTO_AES_KEY_1 0x00bc
- +#define RK_CRYPTO_AES_KEY_2 0x00c0
- +#define RK_CRYPTO_AES_KEY_3 0x00c4
- +#define RK_CRYPTO_AES_KEY_4 0x00c8
- +#define RK_CRYPTO_AES_KEY_5 0x00cc
- +#define RK_CRYPTO_AES_KEY_6 0x00d0
- +#define RK_CRYPTO_AES_KEY_7 0x00d4
- +
- +/* des/tdes */
- +#define RK_CRYPTO_TDES_CTRL 0x0100
- +#define RK_CRYPTO_TDES_BYTESWAP_KEY BIT(8)
- +#define RK_CRYPTO_TDES_BYTESWAP_IV BIT(7)
- +#define RK_CRYPTO_TDES_BYTESWAP_DO BIT(6)
- +#define RK_CRYPTO_TDES_BYTESWAP_DI BIT(5)
- +/* 0: ECB, 1: CBC */
- +#define RK_CRYPTO_TDES_CHAINMODE_CBC BIT(4)
- +/* TDES Key Mode, 0 : EDE, 1 : EEE */
- +#define RK_CRYPTO_TDES_EEE BIT(3)
- +/* 0: DES, 1:TDES */
- +#define RK_CRYPTO_TDES_SELECT BIT(2)
- +/* 0: Slave, 1:Fifo */
- +#define RK_CRYPTO_TDES_FIFO_MODE BIT(1)
- +/* Encryption = 0 , Decryption = 1 */
- +#define RK_CRYPTO_TDES_DEC BIT(0)
- +
- +#define RK_CRYPTO_TDES_STS 0x0104
- +#define RK_CRYPTO_TDES_DONE BIT(0)
- +
- +#define RK_CRYPTO_TDES_DIN_0 0x0108
- +#define RK_CRYPTO_TDES_DIN_1 0x010c
- +#define RK_CRYPTO_TDES_DOUT_0 0x0110
- +#define RK_CRYPTO_TDES_DOUT_1 0x0114
- +#define RK_CRYPTO_TDES_IV_0 0x0118
- +#define RK_CRYPTO_TDES_IV_1 0x011c
- +#define RK_CRYPTO_TDES_KEY1_0 0x0120
- +#define RK_CRYPTO_TDES_KEY1_1 0x0124
- +#define RK_CRYPTO_TDES_KEY2_0 0x0128
- +#define RK_CRYPTO_TDES_KEY2_1 0x012c
- +#define RK_CRYPTO_TDES_KEY3_0 0x0130
- +#define RK_CRYPTO_TDES_KEY3_1 0x0134
- +
- +#define CRYPTO_READ(dev, offset) \
- + readl_relaxed(((dev)->reg + (offset)))
- +#define CRYPTO_WRITE(dev, offset, val) \
- + writel_relaxed((val), ((dev)->reg + (offset)))
- +
- +struct rk_crypto_info {
- + struct device *dev;
- + struct clk *aclk;
- + struct clk *hclk;
- + struct clk *sclk;
- + struct clk *dmaclk;
- + struct reset_control *rst;
- + void __iomem *reg;
- + int irq;
- + struct crypto_queue queue;
- + struct tasklet_struct crypto_tasklet;
- + struct ablkcipher_request *ablk_req;
- + /* device lock */
- + spinlock_t lock;
- +
- + /* the public variable */
- + struct scatterlist *sg_src;
- + struct scatterlist *sg_dst;
- + struct scatterlist sg_tmp;
- + struct scatterlist *first;
- + unsigned int left_bytes;
- + void *addr_vir;
- + int aligned;
- + int align_size;
- + size_t nents;
- + unsigned int total;
- + unsigned int count;
- + u32 mode;
- + dma_addr_t addr_in;
- + dma_addr_t addr_out;
- + int (*start)(struct rk_crypto_info *dev);
- + int (*update)(struct rk_crypto_info *dev);
- + void (*complete)(struct rk_crypto_info *dev, int err);
- + int (*enable_clk)(struct rk_crypto_info *dev);
- + void (*disable_clk)(struct rk_crypto_info *dev);
- + int (*load_data)(struct rk_crypto_info *dev,
- + struct scatterlist *sg_src,
- + struct scatterlist *sg_dst);
- + void (*unload_data)(struct rk_crypto_info *dev);
- +};
- +
- +/* the private variable of cipher */
- +struct rk_cipher_ctx {
- + struct rk_crypto_info *dev;
- + unsigned int keylen;
- +};
- +
- +struct rk_crypto_tmp {
- + struct rk_crypto_info *dev;
- + struct crypto_alg alg;
- +};
- +
- +extern struct rk_crypto_tmp rk_ecb_aes_alg;
- +extern struct rk_crypto_tmp rk_cbc_aes_alg;
- +extern struct rk_crypto_tmp rk_ecb_des_alg;
- +extern struct rk_crypto_tmp rk_cbc_des_alg;
- +extern struct rk_crypto_tmp rk_ecb_des3_ede_alg;
- +extern struct rk_crypto_tmp rk_cbc_des3_ede_alg;
- +
- +#endif
- diff --git a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- new file mode 100644
- index 0000000..4a8f9de
- --- /dev/null
- +++ b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- @@ -0,0 +1,503 @@
- +/*
- + * Crypto acceleration support for Rockchip RK3288
- + *
- + * Copyright (c) 2015, Fuzhou Rockchip Electronics Co., Ltd
- + *
- + * Author: Zain Wang <zain.wang@rock-chips.com>
- + *
- + * This program is free software; you can redistribute it and/or modify it
- + * under the terms and conditions of the GNU General Public License,
- + * version 2, as published by the Free Software Foundation.
- + *
- + * Some ideas are from marvell-cesa.c and s5p-sss.c driver.
- + */
- +#include "rk3288_crypto.h"
- +
- +#define RK_CRYPTO_DEC BIT(0)
- +
- +static void rk_crypto_complete(struct rk_crypto_info *dev, int err)
- +{
- + if (dev->ablk_req->base.complete)
- + dev->ablk_req->base.complete(&dev->ablk_req->base, err);
- +}
- +
- +static int rk_handle_req(struct rk_crypto_info *dev,
- + struct ablkcipher_request *req)
- +{
- + int err;
- +
- + if (!IS_ALIGNED(req->nbytes, dev->align_size))
- + return -EINVAL;
- +
- + dev->left_bytes = req->nbytes;
- + dev->total = req->nbytes;
- + dev->sg_src = req->src;
- + dev->first = req->src;
- + dev->nents = sg_nents(req->src);
- + dev->sg_dst = req->dst;
- + dev->aligned = 1;
- + dev->ablk_req = req;
- +
- + spin_lock(&dev->lock);
- + err = ablkcipher_enqueue_request(&dev->queue, req);
- + spin_unlock(&dev->lock);
- + tasklet_schedule(&dev->crypto_tasklet);
- + return err;
- +}
- +
- +static int rk_aes_setkey(struct crypto_ablkcipher *cipher,
- + const u8 *key, unsigned int keylen)
- +{
- + struct crypto_tfm *tfm = crypto_ablkcipher_tfm(cipher);
- + struct rk_cipher_ctx *ctx = crypto_tfm_ctx(tfm);
- +
- + if (keylen != AES_KEYSIZE_128 && keylen != AES_KEYSIZE_192 &&
- + keylen != AES_KEYSIZE_256) {
- + crypto_ablkcipher_set_flags(cipher, CRYPTO_TFM_RES_BAD_KEY_LEN);
- + return -EINVAL;
- + }
- + ctx->keylen = keylen;
- + memcpy_toio(ctx->dev->reg + RK_CRYPTO_AES_KEY_0, key, keylen);
- + return 0;
- +}
- +
- +static int rk_tdes_setkey(struct crypto_ablkcipher *cipher,
- + const u8 *key, unsigned int keylen)
- +{
- + struct crypto_tfm *tfm = crypto_ablkcipher_tfm(cipher);
- + struct rk_cipher_ctx *ctx = crypto_tfm_ctx(tfm);
- + u32 tmp[DES_EXPKEY_WORDS];
- +
- + if (keylen != DES_KEY_SIZE && keylen != DES3_EDE_KEY_SIZE) {
- + crypto_ablkcipher_set_flags(cipher, CRYPTO_TFM_RES_BAD_KEY_LEN);
- + return -EINVAL;
- + }
- +
- + if (keylen == DES_KEY_SIZE) {
- + if (!des_ekey(tmp, key) &&
- + (tfm->crt_flags & CRYPTO_TFM_REQ_WEAK_KEY)) {
- + tfm->crt_flags |= CRYPTO_TFM_RES_WEAK_KEY;
- + return -EINVAL;
- + }
- + }
- +
- + ctx->keylen = keylen;
- + memcpy_toio(ctx->dev->reg + RK_CRYPTO_TDES_KEY1_0, key, keylen);
- + return 0;
- +}
- +
- +static int rk_aes_ecb_encrypt(struct ablkcipher_request *req)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- + struct rk_crypto_info *dev = ctx->dev;
- +
- + dev->mode = RK_CRYPTO_AES_ECB_MODE;
- + return rk_handle_req(dev, req);
- +}
- +
- +static int rk_aes_ecb_decrypt(struct ablkcipher_request *req)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- + struct rk_crypto_info *dev = ctx->dev;
- +
- + dev->mode = RK_CRYPTO_AES_ECB_MODE | RK_CRYPTO_DEC;
- + return rk_handle_req(dev, req);
- +}
- +
- +static int rk_aes_cbc_encrypt(struct ablkcipher_request *req)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- + struct rk_crypto_info *dev = ctx->dev;
- +
- + dev->mode = RK_CRYPTO_AES_CBC_MODE;
- + return rk_handle_req(dev, req);
- +}
- +
- +static int rk_aes_cbc_decrypt(struct ablkcipher_request *req)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- + struct rk_crypto_info *dev = ctx->dev;
- +
- + dev->mode = RK_CRYPTO_AES_CBC_MODE | RK_CRYPTO_DEC;
- + return rk_handle_req(dev, req);
- +}
- +
- +static int rk_des_ecb_encrypt(struct ablkcipher_request *req)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- + struct rk_crypto_info *dev = ctx->dev;
- +
- + dev->mode = 0;
- + return rk_handle_req(dev, req);
- +}
- +
- +static int rk_des_ecb_decrypt(struct ablkcipher_request *req)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- + struct rk_crypto_info *dev = ctx->dev;
- +
- + dev->mode = RK_CRYPTO_DEC;
- + return rk_handle_req(dev, req);
- +}
- +
- +static int rk_des_cbc_encrypt(struct ablkcipher_request *req)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- + struct rk_crypto_info *dev = ctx->dev;
- +
- + dev->mode = RK_CRYPTO_TDES_CHAINMODE_CBC;
- + return rk_handle_req(dev, req);
- +}
- +
- +static int rk_des_cbc_decrypt(struct ablkcipher_request *req)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- + struct rk_crypto_info *dev = ctx->dev;
- +
- + dev->mode = RK_CRYPTO_TDES_CHAINMODE_CBC | RK_CRYPTO_DEC;
- + return rk_handle_req(dev, req);
- +}
- +
- +static int rk_des3_ede_ecb_encrypt(struct ablkcipher_request *req)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- + struct rk_crypto_info *dev = ctx->dev;
- +
- + dev->mode = RK_CRYPTO_TDES_SELECT;
- + return rk_handle_req(dev, req);
- +}
- +
- +static int rk_des3_ede_ecb_decrypt(struct ablkcipher_request *req)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- + struct rk_crypto_info *dev = ctx->dev;
- +
- + dev->mode = RK_CRYPTO_TDES_SELECT | RK_CRYPTO_DEC;
- + return rk_handle_req(dev, req);
- +}
- +
- +static int rk_des3_ede_cbc_encrypt(struct ablkcipher_request *req)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- + struct rk_crypto_info *dev = ctx->dev;
- +
- + dev->mode = RK_CRYPTO_TDES_SELECT | RK_CRYPTO_TDES_CHAINMODE_CBC;
- + return rk_handle_req(dev, req);
- +}
- +
- +static int rk_des3_ede_cbc_decrypt(struct ablkcipher_request *req)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- + struct rk_crypto_info *dev = ctx->dev;
- +
- + dev->mode = RK_CRYPTO_TDES_SELECT | RK_CRYPTO_TDES_CHAINMODE_CBC |
- + RK_CRYPTO_DEC;
- + return rk_handle_req(dev, req);
- +}
- +
- +static void rk_ablk_hw_init(struct rk_crypto_info *dev)
- +{
- + struct crypto_ablkcipher *cipher =
- + crypto_ablkcipher_reqtfm(dev->ablk_req);
- + struct crypto_tfm *tfm = crypto_ablkcipher_tfm(cipher);
- + struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(cipher);
- + u32 ivsize, block, conf_reg = 0;
- +
- + block = crypto_tfm_alg_blocksize(tfm);
- + ivsize = crypto_ablkcipher_ivsize(cipher);
- +
- + if (block == DES_BLOCK_SIZE) {
- + dev->mode |= RK_CRYPTO_TDES_FIFO_MODE |
- + RK_CRYPTO_TDES_BYTESWAP_KEY |
- + RK_CRYPTO_TDES_BYTESWAP_IV;
- + CRYPTO_WRITE(dev, RK_CRYPTO_TDES_CTRL, dev->mode);
- + memcpy_toio(dev->reg + RK_CRYPTO_TDES_IV_0,
- + dev->ablk_req->info, ivsize);
- + conf_reg = RK_CRYPTO_DESSEL;
- + } else {
- + dev->mode |= RK_CRYPTO_AES_FIFO_MODE |
- + RK_CRYPTO_AES_KEY_CHANGE |
- + RK_CRYPTO_AES_BYTESWAP_KEY |
- + RK_CRYPTO_AES_BYTESWAP_IV;
- + if (ctx->keylen == AES_KEYSIZE_192)
- + dev->mode |= RK_CRYPTO_AES_192BIT_key;
- + else if (ctx->keylen == AES_KEYSIZE_256)
- + dev->mode |= RK_CRYPTO_AES_256BIT_key;
- + CRYPTO_WRITE(dev, RK_CRYPTO_AES_CTRL, dev->mode);
- + memcpy_toio(dev->reg + RK_CRYPTO_AES_IV_0,
- + dev->ablk_req->info, ivsize);
- + }
- + conf_reg |= RK_CRYPTO_BYTESWAP_BTFIFO |
- + RK_CRYPTO_BYTESWAP_BRFIFO;
- + CRYPTO_WRITE(dev, RK_CRYPTO_CONF, conf_reg);
- + CRYPTO_WRITE(dev, RK_CRYPTO_INTENA,
- + RK_CRYPTO_BCDMA_ERR_ENA | RK_CRYPTO_BCDMA_DONE_ENA);
- +}
- +
- +static void crypto_dma_start(struct rk_crypto_info *dev)
- +{
- + CRYPTO_WRITE(dev, RK_CRYPTO_BRDMAS, dev->addr_in);
- + CRYPTO_WRITE(dev, RK_CRYPTO_BRDMAL, dev->count / 4);
- + CRYPTO_WRITE(dev, RK_CRYPTO_BTDMAS, dev->addr_out);
- + CRYPTO_WRITE(dev, RK_CRYPTO_CTRL, RK_CRYPTO_BLOCK_START |
- + _SBF(RK_CRYPTO_BLOCK_START, 16));
- +}
- +
- +static int rk_set_data_start(struct rk_crypto_info *dev)
- +{
- + int err;
- +
- + err = dev->load_data(dev, dev->sg_src, dev->sg_dst);
- + if (!err)
- + crypto_dma_start(dev);
- + return err;
- +}
- +
- +static int rk_ablk_start(struct rk_crypto_info *dev)
- +{
- + int err;
- +
- + spin_lock(&dev->lock);
- + rk_ablk_hw_init(dev);
- + err = rk_set_data_start(dev);
- + spin_unlock(&dev->lock);
- + return err;
- +}
- +
- +static void rk_iv_copyback(struct rk_crypto_info *dev)
- +{
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(dev->ablk_req);
- + u32 ivsize = crypto_ablkcipher_ivsize(tfm);
- +
- + if (ivsize == DES_BLOCK_SIZE)
- + memcpy_fromio(dev->ablk_req->info,
- + dev->reg + RK_CRYPTO_TDES_IV_0, ivsize);
- + else if (ivsize == AES_BLOCK_SIZE)
- + memcpy_fromio(dev->ablk_req->info,
- + dev->reg + RK_CRYPTO_AES_IV_0, ivsize);
- +}
- +
- +/* return:
- + * true some err was occurred
- + * fault no err, continue
- + */
- +static int rk_ablk_rx(struct rk_crypto_info *dev)
- +{
- + int err = 0;
- +
- + dev->unload_data(dev);
- + if (!dev->aligned) {
- + if (!sg_pcopy_from_buffer(dev->ablk_req->dst, dev->nents,
- + dev->addr_vir, dev->count,
- + dev->total - dev->left_bytes -
- + dev->count)) {
- + err = -EINVAL;
- + goto out_rx;
- + }
- + }
- + if (dev->left_bytes) {
- + if (dev->aligned) {
- + if (sg_is_last(dev->sg_src)) {
- + dev_err(dev->dev, "[%s:%d] Lack of data\n",
- + __func__, __LINE__);
- + err = -ENOMEM;
- + goto out_rx;
- + }
- + dev->sg_src = sg_next(dev->sg_src);
- + dev->sg_dst = sg_next(dev->sg_dst);
- + }
- + err = rk_set_data_start(dev);
- + } else {
- + rk_iv_copyback(dev);
- + /* here show the calculation is over without any err */
- + dev->complete(dev, 0);
- + }
- +out_rx:
- + return err;
- +}
- +
- +static int rk_ablk_cra_init(struct crypto_tfm *tfm)
- +{
- + struct rk_cipher_ctx *ctx = crypto_tfm_ctx(tfm);
- + struct crypto_alg *alg = tfm->__crt_alg;
- + struct rk_crypto_tmp *algt;
- +
- + algt = container_of(alg, struct rk_crypto_tmp, alg);
- +
- + ctx->dev = algt->dev;
- + ctx->dev->align_size = crypto_tfm_alg_alignmask(tfm) + 1;
- + ctx->dev->start = rk_ablk_start;
- + ctx->dev->update = rk_ablk_rx;
- + ctx->dev->complete = rk_crypto_complete;
- + ctx->dev->addr_vir = (char *)__get_free_page(GFP_KERNEL);
- +
- + return ctx->dev->addr_vir ? ctx->dev->enable_clk(ctx->dev) : -ENOMEM;
- +}
- +
- +static void rk_ablk_cra_exit(struct crypto_tfm *tfm)
- +{
- + struct rk_cipher_ctx *ctx = crypto_tfm_ctx(tfm);
- +
- + free_page((unsigned long)ctx->dev->addr_vir);
- + ctx->dev->disable_clk(ctx->dev);
- +}
- +
- +struct rk_crypto_tmp rk_ecb_aes_alg = {
- + .alg = {
- + .cra_name = "ecb(aes)",
- + .cra_driver_name = "ecb-aes-rk",
- + .cra_priority = 300,
- + .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER |
- + CRYPTO_ALG_ASYNC,
- + .cra_blocksize = AES_BLOCK_SIZE,
- + .cra_ctxsize = sizeof(struct rk_cipher_ctx),
- + .cra_alignmask = 0x0f,
- + .cra_type = &crypto_ablkcipher_type,
- + .cra_module = THIS_MODULE,
- + .cra_init = rk_ablk_cra_init,
- + .cra_exit = rk_ablk_cra_exit,
- + .cra_u.ablkcipher = {
- + .min_keysize = AES_MIN_KEY_SIZE,
- + .max_keysize = AES_MAX_KEY_SIZE,
- + .setkey = rk_aes_setkey,
- + .encrypt = rk_aes_ecb_encrypt,
- + .decrypt = rk_aes_ecb_decrypt,
- + }
- + }
- +};
- +
- +struct rk_crypto_tmp rk_cbc_aes_alg = {
- + .alg = {
- + .cra_name = "cbc(aes)",
- + .cra_driver_name = "cbc-aes-rk",
- + .cra_priority = 300,
- + .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER |
- + CRYPTO_ALG_ASYNC,
- + .cra_blocksize = AES_BLOCK_SIZE,
- + .cra_ctxsize = sizeof(struct rk_cipher_ctx),
- + .cra_alignmask = 0x0f,
- + .cra_type = &crypto_ablkcipher_type,
- + .cra_module = THIS_MODULE,
- + .cra_init = rk_ablk_cra_init,
- + .cra_exit = rk_ablk_cra_exit,
- + .cra_u.ablkcipher = {
- + .min_keysize = AES_MIN_KEY_SIZE,
- + .max_keysize = AES_MAX_KEY_SIZE,
- + .ivsize = AES_BLOCK_SIZE,
- + .setkey = rk_aes_setkey,
- + .encrypt = rk_aes_cbc_encrypt,
- + .decrypt = rk_aes_cbc_decrypt,
- + }
- + }
- +};
- +
- +struct rk_crypto_tmp rk_ecb_des_alg = {
- + .alg = {
- + .cra_name = "ecb(des)",
- + .cra_driver_name = "ecb-des-rk",
- + .cra_priority = 300,
- + .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER |
- + CRYPTO_ALG_ASYNC,
- + .cra_blocksize = DES_BLOCK_SIZE,
- + .cra_ctxsize = sizeof(struct rk_cipher_ctx),
- + .cra_alignmask = 0x07,
- + .cra_type = &crypto_ablkcipher_type,
- + .cra_module = THIS_MODULE,
- + .cra_init = rk_ablk_cra_init,
- + .cra_exit = rk_ablk_cra_exit,
- + .cra_u.ablkcipher = {
- + .min_keysize = DES_KEY_SIZE,
- + .max_keysize = DES_KEY_SIZE,
- + .setkey = rk_tdes_setkey,
- + .encrypt = rk_des_ecb_encrypt,
- + .decrypt = rk_des_ecb_decrypt,
- + }
- + }
- +};
- +
- +struct rk_crypto_tmp rk_cbc_des_alg = {
- + .alg = {
- + .cra_name = "cbc(des)",
- + .cra_driver_name = "cbc-des-rk",
- + .cra_priority = 300,
- + .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER |
- + CRYPTO_ALG_ASYNC,
- + .cra_blocksize = DES_BLOCK_SIZE,
- + .cra_ctxsize = sizeof(struct rk_cipher_ctx),
- + .cra_alignmask = 0x07,
- + .cra_type = &crypto_ablkcipher_type,
- + .cra_module = THIS_MODULE,
- + .cra_init = rk_ablk_cra_init,
- + .cra_exit = rk_ablk_cra_exit,
- + .cra_u.ablkcipher = {
- + .min_keysize = DES_KEY_SIZE,
- + .max_keysize = DES_KEY_SIZE,
- + .ivsize = DES_BLOCK_SIZE,
- + .setkey = rk_tdes_setkey,
- + .encrypt = rk_des_cbc_encrypt,
- + .decrypt = rk_des_cbc_decrypt,
- + }
- + }
- +};
- +
- +struct rk_crypto_tmp rk_ecb_des3_ede_alg = {
- + .alg = {
- + .cra_name = "ecb(des3_ede)",
- + .cra_driver_name = "ecb-des3-ede-rk",
- + .cra_priority = 300,
- + .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER |
- + CRYPTO_ALG_ASYNC,
- + .cra_blocksize = DES_BLOCK_SIZE,
- + .cra_ctxsize = sizeof(struct rk_cipher_ctx),
- + .cra_alignmask = 0x07,
- + .cra_type = &crypto_ablkcipher_type,
- + .cra_module = THIS_MODULE,
- + .cra_init = rk_ablk_cra_init,
- + .cra_exit = rk_ablk_cra_exit,
- + .cra_u.ablkcipher = {
- + .min_keysize = DES3_EDE_KEY_SIZE,
- + .max_keysize = DES3_EDE_KEY_SIZE,
- + .ivsize = DES_BLOCK_SIZE,
- + .setkey = rk_tdes_setkey,
- + .encrypt = rk_des3_ede_ecb_encrypt,
- + .decrypt = rk_des3_ede_ecb_decrypt,
- + }
- + }
- +};
- +
- +struct rk_crypto_tmp rk_cbc_des3_ede_alg = {
- + .alg = {
- + .cra_name = "cbc(des3_ede)",
- + .cra_driver_name = "cbc-des3-ede-rk",
- + .cra_priority = 300,
- + .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER |
- + CRYPTO_ALG_ASYNC,
- + .cra_blocksize = DES_BLOCK_SIZE,
- + .cra_ctxsize = sizeof(struct rk_cipher_ctx),
- + .cra_alignmask = 0x07,
- + .cra_type = &crypto_ablkcipher_type,
- + .cra_module = THIS_MODULE,
- + .cra_init = rk_ablk_cra_init,
- + .cra_exit = rk_ablk_cra_exit,
- + .cra_u.ablkcipher = {
- + .min_keysize = DES3_EDE_KEY_SIZE,
- + .max_keysize = DES3_EDE_KEY_SIZE,
- + .ivsize = DES_BLOCK_SIZE,
- + .setkey = rk_tdes_setkey,
- + .encrypt = rk_des3_ede_cbc_encrypt,
- + .decrypt = rk_des3_ede_cbc_decrypt,
- + }
- + }
- +};
- commit ac7c8e6b6dc959d285382c7e9cdfe608205f0c68
- Author: Heiko Stuebner <heiko@sntech.de>
- Date: Sat Nov 28 13:27:48 2015 +0100
- crypto: rockchip - fix possible deadlock
- Lockdep warns about a possible deadlock resulting from the use of regular
- spin_locks:
- =================================
- [ INFO: inconsistent lock state ]
- 4.4.0-rc2+ #2724 Not tainted
- ---------------------------------
- inconsistent {SOFTIRQ-ON-W} -> {IN-SOFTIRQ-W} usage.
- ksoftirqd/0/3 [HC0[0]:SC1[1]:HE1:SE0] takes:
- (&(&crypto_info->lock)->rlock){+.?...}, at: [<bf14a65c>] rk_crypto_tasklet_cb+0x24/0xb4 [rk_crypto]
- {SOFTIRQ-ON-W} state was registered at:
- [<c007f4ac>] lock_acquire+0x178/0x218
- [<c0759bac>] _raw_spin_lock+0x54/0x64
- [<bf14af88>] rk_handle_req+0x7c/0xbc [rk_crypto]
- [<bf14b040>] rk_des_ecb_encrypt+0x2c/0x30 [rk_crypto]
- [<bf14b05c>] rk_aes_ecb_encrypt+0x18/0x1c [rk_crypto]
- [<c028c820>] skcipher_encrypt_ablkcipher+0x64/0x68
- [<c0290770>] __test_skcipher+0x2a8/0x8dc
- [<c0292e94>] test_skcipher+0x38/0xc4
- [<c0292fb0>] alg_test_skcipher+0x90/0xb0
- [<c0292158>] alg_test+0x1e8/0x280
- [<c028f6f4>] cryptomgr_test+0x34/0x54
- [<c004bbe8>] kthread+0xf4/0x10c
- [<c0010010>] ret_from_fork+0x14/0x24
- irq event stamp: 10672
- hardirqs last enabled at (10672): [<c002fac8>] tasklet_action+0x48/0x104
- hardirqs last disabled at (10671): [<c002faa0>] tasklet_action+0x20/0x104
- softirqs last enabled at (10658): [<c002ef84>] __do_softirq+0x358/0x49c
- softirqs last disabled at (10669): [<c002f108>] run_ksoftirqd+0x40/0x80
- other info that might help us debug this:
- Possible unsafe locking scenario:
- CPU0
- ----
- lock(&(&crypto_info->lock)->rlock);
- <Interrupt>
- lock(&(&crypto_info->lock)->rlock);
- *** DEADLOCK ***
- Fix this by moving to irq-disabling spinlocks.
- Signed-off-by: Heiko Stuebner <heiko@sntech.de>
- Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
- diff --git a/drivers/crypto/rockchip/rk3288_crypto.c b/drivers/crypto/rockchip/rk3288_crypto.c
- index 6b72f8d..da9c73d 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto.c
- @@ -190,12 +190,13 @@ static void rk_crypto_tasklet_cb(unsigned long data)
- {
- struct rk_crypto_info *dev = (struct rk_crypto_info *)data;
- struct crypto_async_request *async_req, *backlog;
- + unsigned long flags;
- int err = 0;
- - spin_lock(&dev->lock);
- + spin_lock_irqsave(&dev->lock, flags);
- backlog = crypto_get_backlog(&dev->queue);
- async_req = crypto_dequeue_request(&dev->queue);
- - spin_unlock(&dev->lock);
- + spin_unlock_irqrestore(&dev->lock, flags);
- if (!async_req) {
- dev_err(dev->dev, "async_req is NULL !!\n");
- return;
- diff --git a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- index 4a8f9de..d98b681 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- @@ -24,6 +24,7 @@ static void rk_crypto_complete(struct rk_crypto_info *dev, int err)
- static int rk_handle_req(struct rk_crypto_info *dev,
- struct ablkcipher_request *req)
- {
- + unsigned long flags;
- int err;
- if (!IS_ALIGNED(req->nbytes, dev->align_size))
- @@ -38,9 +39,9 @@ static int rk_handle_req(struct rk_crypto_info *dev,
- dev->aligned = 1;
- dev->ablk_req = req;
- - spin_lock(&dev->lock);
- + spin_lock_irqsave(&dev->lock, flags);
- err = ablkcipher_enqueue_request(&dev->queue, req);
- - spin_unlock(&dev->lock);
- + spin_unlock_irqrestore(&dev->lock, flags);
- tasklet_schedule(&dev->crypto_tasklet);
- return err;
- }
- @@ -267,12 +268,13 @@ static int rk_set_data_start(struct rk_crypto_info *dev)
- static int rk_ablk_start(struct rk_crypto_info *dev)
- {
- + unsigned long flags;
- int err;
- - spin_lock(&dev->lock);
- + spin_lock_irqsave(&dev->lock, flags);
- rk_ablk_hw_init(dev);
- err = rk_set_data_start(dev);
- - spin_unlock(&dev->lock);
- + spin_unlock_irqrestore(&dev->lock, flags);
- return err;
- }
- commit bfd927ffa219ac81082b2dcc61a1c4037869befc
- Author: Zain Wang <zain.wang@rock-chips.com>
- Date: Tue Feb 16 10:15:01 2016 +0800
- crypto: rockchip - add hash support for crypto engine in rk3288
- Add md5 sha1 sha256 support for crypto engine in rk3288.
- Signed-off-by: Zain Wang <zain.wang@rock-chips.com>
- Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
- diff --git a/drivers/crypto/Kconfig b/drivers/crypto/Kconfig
- index fed3ffb..477fffd 100644
- --- a/drivers/crypto/Kconfig
- +++ b/drivers/crypto/Kconfig
- @@ -508,6 +508,10 @@ config CRYPTO_DEV_ROCKCHIP
- depends on OF && ARCH_ROCKCHIP
- select CRYPTO_AES
- select CRYPTO_DES
- + select CRYPTO_MD5
- + select CRYPTO_SHA1
- + select CRYPTO_SHA256
- + select CRYPTO_HASH
- select CRYPTO_BLKCIPHER
- help
- diff --git a/drivers/crypto/rockchip/Makefile b/drivers/crypto/rockchip/Makefile
- index 7051c6c..30f9129 100644
- --- a/drivers/crypto/rockchip/Makefile
- +++ b/drivers/crypto/rockchip/Makefile
- @@ -1,3 +1,4 @@
- obj-$(CONFIG_CRYPTO_DEV_ROCKCHIP) += rk_crypto.o
- rk_crypto-objs := rk3288_crypto.o \
- rk3288_crypto_ablkcipher.o \
- + rk3288_crypto_ahash.o
- diff --git a/drivers/crypto/rockchip/rk3288_crypto.c b/drivers/crypto/rockchip/rk3288_crypto.c
- index da9c73d..af50825 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto.c
- @@ -208,6 +208,8 @@ static void rk_crypto_tasklet_cb(unsigned long data)
- if (crypto_tfm_alg_type(async_req->tfm) == CRYPTO_ALG_TYPE_ABLKCIPHER)
- dev->ablk_req = ablkcipher_request_cast(async_req);
- + else
- + dev->ahash_req = ahash_request_cast(async_req);
- err = dev->start(dev);
- if (err)
- dev->complete(dev, err);
- @@ -220,6 +222,9 @@ static void rk_crypto_tasklet_cb(unsigned long data)
- &rk_cbc_des_alg,
- &rk_ecb_des3_ede_alg,
- &rk_cbc_des3_ede_alg,
- + &rk_ahash_sha1,
- + &rk_ahash_sha256,
- + &rk_ahash_md5,
- };
- static int rk_crypto_register(struct rk_crypto_info *crypto_info)
- @@ -229,15 +234,24 @@ static int rk_crypto_register(struct rk_crypto_info *crypto_info)
- for (i = 0; i < ARRAY_SIZE(rk_cipher_algs); i++) {
- rk_cipher_algs[i]->dev = crypto_info;
- - err = crypto_register_alg(&rk_cipher_algs[i]->alg);
- + if (rk_cipher_algs[i]->type == ALG_TYPE_CIPHER)
- + err = crypto_register_alg(
- + &rk_cipher_algs[i]->alg.crypto);
- + else
- + err = crypto_register_ahash(
- + &rk_cipher_algs[i]->alg.hash);
- if (err)
- goto err_cipher_algs;
- }
- return 0;
- err_cipher_algs:
- - for (k = 0; k < i; k++)
- - crypto_unregister_alg(&rk_cipher_algs[k]->alg);
- + for (k = 0; k < i; k++) {
- + if (rk_cipher_algs[i]->type == ALG_TYPE_CIPHER)
- + crypto_unregister_alg(&rk_cipher_algs[k]->alg.crypto);
- + else
- + crypto_unregister_ahash(&rk_cipher_algs[i]->alg.hash);
- + }
- return err;
- }
- @@ -245,8 +259,12 @@ static void rk_crypto_unregister(void)
- {
- unsigned int i;
- - for (i = 0; i < ARRAY_SIZE(rk_cipher_algs); i++)
- - crypto_unregister_alg(&rk_cipher_algs[i]->alg);
- + for (i = 0; i < ARRAY_SIZE(rk_cipher_algs); i++) {
- + if (rk_cipher_algs[i]->type == ALG_TYPE_CIPHER)
- + crypto_unregister_alg(&rk_cipher_algs[i]->alg.crypto);
- + else
- + crypto_unregister_ahash(&rk_cipher_algs[i]->alg.hash);
- + }
- }
- static void rk_crypto_action(void *data)
- diff --git a/drivers/crypto/rockchip/rk3288_crypto.h b/drivers/crypto/rockchip/rk3288_crypto.h
- index e499c2c..d7b71fe 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto.h
- +++ b/drivers/crypto/rockchip/rk3288_crypto.h
- @@ -6,6 +6,10 @@
- #include <crypto/algapi.h>
- #include <linux/interrupt.h>
- #include <linux/delay.h>
- +#include <crypto/internal/hash.h>
- +
- +#include <crypto/md5.h>
- +#include <crypto/sha.h>
- #define _SBF(v, f) ((v) << (f))
- @@ -149,6 +153,28 @@
- #define RK_CRYPTO_TDES_KEY3_0 0x0130
- #define RK_CRYPTO_TDES_KEY3_1 0x0134
- +/* HASH */
- +#define RK_CRYPTO_HASH_CTRL 0x0180
- +#define RK_CRYPTO_HASH_SWAP_DO BIT(3)
- +#define RK_CRYPTO_HASH_SWAP_DI BIT(2)
- +#define RK_CRYPTO_HASH_SHA1 _SBF(0x00, 0)
- +#define RK_CRYPTO_HASH_MD5 _SBF(0x01, 0)
- +#define RK_CRYPTO_HASH_SHA256 _SBF(0x02, 0)
- +#define RK_CRYPTO_HASH_PRNG _SBF(0x03, 0)
- +
- +#define RK_CRYPTO_HASH_STS 0x0184
- +#define RK_CRYPTO_HASH_DONE BIT(0)
- +
- +#define RK_CRYPTO_HASH_MSG_LEN 0x0188
- +#define RK_CRYPTO_HASH_DOUT_0 0x018c
- +#define RK_CRYPTO_HASH_DOUT_1 0x0190
- +#define RK_CRYPTO_HASH_DOUT_2 0x0194
- +#define RK_CRYPTO_HASH_DOUT_3 0x0198
- +#define RK_CRYPTO_HASH_DOUT_4 0x019c
- +#define RK_CRYPTO_HASH_DOUT_5 0x01a0
- +#define RK_CRYPTO_HASH_DOUT_6 0x01a4
- +#define RK_CRYPTO_HASH_DOUT_7 0x01a8
- +
- #define CRYPTO_READ(dev, offset) \
- readl_relaxed(((dev)->reg + (offset)))
- #define CRYPTO_WRITE(dev, offset, val) \
- @@ -166,6 +192,7 @@ struct rk_crypto_info {
- struct crypto_queue queue;
- struct tasklet_struct crypto_tasklet;
- struct ablkcipher_request *ablk_req;
- + struct ahash_request *ahash_req;
- /* device lock */
- spinlock_t lock;
- @@ -195,15 +222,36 @@ struct rk_crypto_info {
- void (*unload_data)(struct rk_crypto_info *dev);
- };
- +/* the private variable of hash */
- +struct rk_ahash_ctx {
- + struct rk_crypto_info *dev;
- + /* for fallback */
- + struct crypto_ahash *fallback_tfm;
- +};
- +
- +/* the privete variable of hash for fallback */
- +struct rk_ahash_rctx {
- + struct ahash_request fallback_req;
- +};
- +
- /* the private variable of cipher */
- struct rk_cipher_ctx {
- struct rk_crypto_info *dev;
- unsigned int keylen;
- };
- +enum alg_type {
- + ALG_TYPE_HASH,
- + ALG_TYPE_CIPHER,
- +};
- +
- struct rk_crypto_tmp {
- - struct rk_crypto_info *dev;
- - struct crypto_alg alg;
- + struct rk_crypto_info *dev;
- + union {
- + struct crypto_alg crypto;
- + struct ahash_alg hash;
- + } alg;
- + enum alg_type type;
- };
- extern struct rk_crypto_tmp rk_ecb_aes_alg;
- @@ -213,4 +261,8 @@ struct rk_crypto_tmp {
- extern struct rk_crypto_tmp rk_ecb_des3_ede_alg;
- extern struct rk_crypto_tmp rk_cbc_des3_ede_alg;
- +extern struct rk_crypto_tmp rk_ahash_sha1;
- +extern struct rk_crypto_tmp rk_ahash_sha256;
- +extern struct rk_crypto_tmp rk_ahash_md5;
- +
- #endif
- diff --git a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- index d98b681..b5a3afe 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- @@ -336,7 +336,7 @@ static int rk_ablk_cra_init(struct crypto_tfm *tfm)
- struct crypto_alg *alg = tfm->__crt_alg;
- struct rk_crypto_tmp *algt;
- - algt = container_of(alg, struct rk_crypto_tmp, alg);
- + algt = container_of(alg, struct rk_crypto_tmp, alg.crypto);
- ctx->dev = algt->dev;
- ctx->dev->align_size = crypto_tfm_alg_alignmask(tfm) + 1;
- @@ -357,7 +357,8 @@ static void rk_ablk_cra_exit(struct crypto_tfm *tfm)
- }
- struct rk_crypto_tmp rk_ecb_aes_alg = {
- - .alg = {
- + .type = ALG_TYPE_CIPHER,
- + .alg.crypto = {
- .cra_name = "ecb(aes)",
- .cra_driver_name = "ecb-aes-rk",
- .cra_priority = 300,
- @@ -381,7 +382,8 @@ struct rk_crypto_tmp rk_ecb_aes_alg = {
- };
- struct rk_crypto_tmp rk_cbc_aes_alg = {
- - .alg = {
- + .type = ALG_TYPE_CIPHER,
- + .alg.crypto = {
- .cra_name = "cbc(aes)",
- .cra_driver_name = "cbc-aes-rk",
- .cra_priority = 300,
- @@ -406,7 +408,8 @@ struct rk_crypto_tmp rk_cbc_aes_alg = {
- };
- struct rk_crypto_tmp rk_ecb_des_alg = {
- - .alg = {
- + .type = ALG_TYPE_CIPHER,
- + .alg.crypto = {
- .cra_name = "ecb(des)",
- .cra_driver_name = "ecb-des-rk",
- .cra_priority = 300,
- @@ -430,7 +433,8 @@ struct rk_crypto_tmp rk_ecb_des_alg = {
- };
- struct rk_crypto_tmp rk_cbc_des_alg = {
- - .alg = {
- + .type = ALG_TYPE_CIPHER,
- + .alg.crypto = {
- .cra_name = "cbc(des)",
- .cra_driver_name = "cbc-des-rk",
- .cra_priority = 300,
- @@ -455,7 +459,8 @@ struct rk_crypto_tmp rk_cbc_des_alg = {
- };
- struct rk_crypto_tmp rk_ecb_des3_ede_alg = {
- - .alg = {
- + .type = ALG_TYPE_CIPHER,
- + .alg.crypto = {
- .cra_name = "ecb(des3_ede)",
- .cra_driver_name = "ecb-des3-ede-rk",
- .cra_priority = 300,
- @@ -480,7 +485,8 @@ struct rk_crypto_tmp rk_ecb_des3_ede_alg = {
- };
- struct rk_crypto_tmp rk_cbc_des3_ede_alg = {
- - .alg = {
- + .type = ALG_TYPE_CIPHER,
- + .alg.crypto = {
- .cra_name = "cbc(des3_ede)",
- .cra_driver_name = "cbc-des3-ede-rk",
- .cra_priority = 300,
- diff --git a/drivers/crypto/rockchip/rk3288_crypto_ahash.c b/drivers/crypto/rockchip/rk3288_crypto_ahash.c
- new file mode 100644
- index 0000000..7185882
- --- /dev/null
- +++ b/drivers/crypto/rockchip/rk3288_crypto_ahash.c
- @@ -0,0 +1,404 @@
- +/*
- + * Crypto acceleration support for Rockchip RK3288
- + *
- + * Copyright (c) 2015, Fuzhou Rockchip Electronics Co., Ltd
- + *
- + * Author: Zain Wang <zain.wang@rock-chips.com>
- + *
- + * This program is free software; you can redistribute it and/or modify it
- + * under the terms and conditions of the GNU General Public License,
- + * version 2, as published by the Free Software Foundation.
- + *
- + * Some ideas are from marvell/cesa.c and s5p-sss.c driver.
- + */
- +#include "rk3288_crypto.h"
- +
- +/*
- + * IC can not process zero message hash,
- + * so we put the fixed hash out when met zero message.
- + */
- +
- +static int zero_message_process(struct ahash_request *req)
- +{
- + struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
- + int rk_digest_size = crypto_ahash_digestsize(tfm);
- +
- + switch (rk_digest_size) {
- + case SHA1_DIGEST_SIZE:
- + memcpy(req->result, sha1_zero_message_hash, rk_digest_size);
- + break;
- + case SHA256_DIGEST_SIZE:
- + memcpy(req->result, sha256_zero_message_hash, rk_digest_size);
- + break;
- + case MD5_DIGEST_SIZE:
- + memcpy(req->result, md5_zero_message_hash, rk_digest_size);
- + break;
- + default:
- + return -EINVAL;
- + }
- +
- + return 0;
- +}
- +
- +static void rk_ahash_crypto_complete(struct rk_crypto_info *dev, int err)
- +{
- + if (dev->ahash_req->base.complete)
- + dev->ahash_req->base.complete(&dev->ahash_req->base, err);
- +}
- +
- +static void rk_ahash_reg_init(struct rk_crypto_info *dev)
- +{
- + int reg_status = 0;
- +
- + reg_status = CRYPTO_READ(dev, RK_CRYPTO_CTRL) |
- + RK_CRYPTO_HASH_FLUSH | _SBF(0xffff, 16);
- + CRYPTO_WRITE(dev, RK_CRYPTO_CTRL, reg_status);
- +
- + reg_status = CRYPTO_READ(dev, RK_CRYPTO_CTRL);
- + reg_status &= (~RK_CRYPTO_HASH_FLUSH);
- + reg_status |= _SBF(0xffff, 16);
- + CRYPTO_WRITE(dev, RK_CRYPTO_CTRL, reg_status);
- +
- + memset_io(dev->reg + RK_CRYPTO_HASH_DOUT_0, 0, 32);
- +
- + CRYPTO_WRITE(dev, RK_CRYPTO_INTENA, RK_CRYPTO_HRDMA_ERR_ENA |
- + RK_CRYPTO_HRDMA_DONE_ENA);
- +
- + CRYPTO_WRITE(dev, RK_CRYPTO_INTSTS, RK_CRYPTO_HRDMA_ERR_INT |
- + RK_CRYPTO_HRDMA_DONE_INT);
- +
- + CRYPTO_WRITE(dev, RK_CRYPTO_HASH_CTRL, dev->mode |
- + RK_CRYPTO_HASH_SWAP_DO);
- +
- + CRYPTO_WRITE(dev, RK_CRYPTO_CONF, RK_CRYPTO_BYTESWAP_HRFIFO |
- + RK_CRYPTO_BYTESWAP_BRFIFO |
- + RK_CRYPTO_BYTESWAP_BTFIFO);
- +
- + CRYPTO_WRITE(dev, RK_CRYPTO_HASH_MSG_LEN, dev->total);
- +}
- +
- +static int rk_ahash_init(struct ahash_request *req)
- +{
- + struct rk_ahash_rctx *rctx = ahash_request_ctx(req);
- + struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
- + struct rk_ahash_ctx *ctx = crypto_ahash_ctx(tfm);
- +
- + ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback_tfm);
- + rctx->fallback_req.base.flags = req->base.flags &
- + CRYPTO_TFM_REQ_MAY_SLEEP;
- +
- + return crypto_ahash_init(&rctx->fallback_req);
- +}
- +
- +static int rk_ahash_update(struct ahash_request *req)
- +{
- + struct rk_ahash_rctx *rctx = ahash_request_ctx(req);
- + struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
- + struct rk_ahash_ctx *ctx = crypto_ahash_ctx(tfm);
- +
- + ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback_tfm);
- + rctx->fallback_req.base.flags = req->base.flags &
- + CRYPTO_TFM_REQ_MAY_SLEEP;
- + rctx->fallback_req.nbytes = req->nbytes;
- + rctx->fallback_req.src = req->src;
- +
- + return crypto_ahash_update(&rctx->fallback_req);
- +}
- +
- +static int rk_ahash_final(struct ahash_request *req)
- +{
- + struct rk_ahash_rctx *rctx = ahash_request_ctx(req);
- + struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
- + struct rk_ahash_ctx *ctx = crypto_ahash_ctx(tfm);
- +
- + ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback_tfm);
- + rctx->fallback_req.base.flags = req->base.flags &
- + CRYPTO_TFM_REQ_MAY_SLEEP;
- + rctx->fallback_req.result = req->result;
- +
- + return crypto_ahash_final(&rctx->fallback_req);
- +}
- +
- +static int rk_ahash_finup(struct ahash_request *req)
- +{
- + struct rk_ahash_rctx *rctx = ahash_request_ctx(req);
- + struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
- + struct rk_ahash_ctx *ctx = crypto_ahash_ctx(tfm);
- +
- + ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback_tfm);
- + rctx->fallback_req.base.flags = req->base.flags &
- + CRYPTO_TFM_REQ_MAY_SLEEP;
- +
- + rctx->fallback_req.nbytes = req->nbytes;
- + rctx->fallback_req.src = req->src;
- + rctx->fallback_req.result = req->result;
- +
- + return crypto_ahash_finup(&rctx->fallback_req);
- +}
- +
- +static int rk_ahash_import(struct ahash_request *req, const void *in)
- +{
- + struct rk_ahash_rctx *rctx = ahash_request_ctx(req);
- + struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
- + struct rk_ahash_ctx *ctx = crypto_ahash_ctx(tfm);
- +
- + ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback_tfm);
- + rctx->fallback_req.base.flags = req->base.flags &
- + CRYPTO_TFM_REQ_MAY_SLEEP;
- +
- + return crypto_ahash_import(&rctx->fallback_req, in);
- +}
- +
- +static int rk_ahash_export(struct ahash_request *req, void *out)
- +{
- + struct rk_ahash_rctx *rctx = ahash_request_ctx(req);
- + struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
- + struct rk_ahash_ctx *ctx = crypto_ahash_ctx(tfm);
- +
- + ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback_tfm);
- + rctx->fallback_req.base.flags = req->base.flags &
- + CRYPTO_TFM_REQ_MAY_SLEEP;
- +
- + return crypto_ahash_export(&rctx->fallback_req, out);
- +}
- +
- +static int rk_ahash_digest(struct ahash_request *req)
- +{
- + struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
- + struct rk_ahash_ctx *tctx = crypto_tfm_ctx(req->base.tfm);
- + struct rk_crypto_info *dev = NULL;
- + unsigned long flags;
- + int ret;
- +
- + if (!req->nbytes)
- + return zero_message_process(req);
- +
- + dev = tctx->dev;
- + dev->total = req->nbytes;
- + dev->left_bytes = req->nbytes;
- + dev->aligned = 0;
- + dev->mode = 0;
- + dev->align_size = 4;
- + dev->sg_dst = NULL;
- + dev->sg_src = req->src;
- + dev->first = req->src;
- + dev->nents = sg_nents(req->src);
- +
- + switch (crypto_ahash_digestsize(tfm)) {
- + case SHA1_DIGEST_SIZE:
- + dev->mode = RK_CRYPTO_HASH_SHA1;
- + break;
- + case SHA256_DIGEST_SIZE:
- + dev->mode = RK_CRYPTO_HASH_SHA256;
- + break;
- + case MD5_DIGEST_SIZE:
- + dev->mode = RK_CRYPTO_HASH_MD5;
- + break;
- + default:
- + return -EINVAL;
- + }
- +
- + rk_ahash_reg_init(dev);
- +
- + spin_lock_irqsave(&dev->lock, flags);
- + ret = crypto_enqueue_request(&dev->queue, &req->base);
- + spin_unlock_irqrestore(&dev->lock, flags);
- +
- + tasklet_schedule(&dev->crypto_tasklet);
- +
- + /*
- + * it will take some time to process date after last dma transmission.
- + *
- + * waiting time is relative with the last date len,
- + * so cannot set a fixed time here.
- + * 10-50 makes system not call here frequently wasting
- + * efficiency, and make it response quickly when dma
- + * complete.
- + */
- + while (!CRYPTO_READ(dev, RK_CRYPTO_HASH_STS))
- + usleep_range(10, 50);
- +
- + memcpy_fromio(req->result, dev->reg + RK_CRYPTO_HASH_DOUT_0,
- + crypto_ahash_digestsize(tfm));
- +
- + return 0;
- +}
- +
- +static void crypto_ahash_dma_start(struct rk_crypto_info *dev)
- +{
- + CRYPTO_WRITE(dev, RK_CRYPTO_HRDMAS, dev->addr_in);
- + CRYPTO_WRITE(dev, RK_CRYPTO_HRDMAL, (dev->count + 3) / 4);
- + CRYPTO_WRITE(dev, RK_CRYPTO_CTRL, RK_CRYPTO_HASH_START |
- + (RK_CRYPTO_HASH_START << 16));
- +}
- +
- +static int rk_ahash_set_data_start(struct rk_crypto_info *dev)
- +{
- + int err;
- +
- + err = dev->load_data(dev, dev->sg_src, NULL);
- + if (!err)
- + crypto_ahash_dma_start(dev);
- + return err;
- +}
- +
- +static int rk_ahash_start(struct rk_crypto_info *dev)
- +{
- + return rk_ahash_set_data_start(dev);
- +}
- +
- +static int rk_ahash_crypto_rx(struct rk_crypto_info *dev)
- +{
- + int err = 0;
- +
- + dev->unload_data(dev);
- + if (dev->left_bytes) {
- + if (dev->aligned) {
- + if (sg_is_last(dev->sg_src)) {
- + dev_warn(dev->dev, "[%s:%d], Lack of data\n",
- + __func__, __LINE__);
- + err = -ENOMEM;
- + goto out_rx;
- + }
- + dev->sg_src = sg_next(dev->sg_src);
- + }
- + err = rk_ahash_set_data_start(dev);
- + } else {
- + dev->complete(dev, 0);
- + }
- +
- +out_rx:
- + return err;
- +}
- +
- +static int rk_cra_hash_init(struct crypto_tfm *tfm)
- +{
- + struct rk_ahash_ctx *tctx = crypto_tfm_ctx(tfm);
- + struct rk_crypto_tmp *algt;
- + struct ahash_alg *alg = __crypto_ahash_alg(tfm->__crt_alg);
- +
- + const char *alg_name = crypto_tfm_alg_name(tfm);
- +
- + algt = container_of(alg, struct rk_crypto_tmp, alg.hash);
- +
- + tctx->dev = algt->dev;
- + tctx->dev->addr_vir = (void *)__get_free_page(GFP_KERNEL);
- + if (!tctx->dev->addr_vir) {
- + dev_err(tctx->dev->dev, "failed to kmalloc for addr_vir\n");
- + return -ENOMEM;
- + }
- + tctx->dev->start = rk_ahash_start;
- + tctx->dev->update = rk_ahash_crypto_rx;
- + tctx->dev->complete = rk_ahash_crypto_complete;
- +
- + /* for fallback */
- + tctx->fallback_tfm = crypto_alloc_ahash(alg_name, 0,
- + CRYPTO_ALG_NEED_FALLBACK);
- + if (IS_ERR(tctx->fallback_tfm)) {
- + dev_err(tctx->dev->dev, "Could not load fallback driver.\n");
- + return PTR_ERR(tctx->fallback_tfm);
- + }
- + crypto_ahash_set_reqsize(__crypto_ahash_cast(tfm),
- + sizeof(struct rk_ahash_rctx) +
- + crypto_ahash_reqsize(tctx->fallback_tfm));
- +
- + return tctx->dev->enable_clk(tctx->dev);
- +}
- +
- +static void rk_cra_hash_exit(struct crypto_tfm *tfm)
- +{
- + struct rk_ahash_ctx *tctx = crypto_tfm_ctx(tfm);
- +
- + free_page((unsigned long)tctx->dev->addr_vir);
- + return tctx->dev->disable_clk(tctx->dev);
- +}
- +
- +struct rk_crypto_tmp rk_ahash_sha1 = {
- + .type = ALG_TYPE_HASH,
- + .alg.hash = {
- + .init = rk_ahash_init,
- + .update = rk_ahash_update,
- + .final = rk_ahash_final,
- + .finup = rk_ahash_finup,
- + .export = rk_ahash_export,
- + .import = rk_ahash_import,
- + .digest = rk_ahash_digest,
- + .halg = {
- + .digestsize = SHA1_DIGEST_SIZE,
- + .statesize = sizeof(struct sha1_state),
- + .base = {
- + .cra_name = "sha1",
- + .cra_driver_name = "rk-sha1",
- + .cra_priority = 300,
- + .cra_flags = CRYPTO_ALG_ASYNC |
- + CRYPTO_ALG_NEED_FALLBACK,
- + .cra_blocksize = SHA1_BLOCK_SIZE,
- + .cra_ctxsize = sizeof(struct rk_ahash_ctx),
- + .cra_alignmask = 3,
- + .cra_init = rk_cra_hash_init,
- + .cra_exit = rk_cra_hash_exit,
- + .cra_module = THIS_MODULE,
- + }
- + }
- + }
- +};
- +
- +struct rk_crypto_tmp rk_ahash_sha256 = {
- + .type = ALG_TYPE_HASH,
- + .alg.hash = {
- + .init = rk_ahash_init,
- + .update = rk_ahash_update,
- + .final = rk_ahash_final,
- + .finup = rk_ahash_finup,
- + .export = rk_ahash_export,
- + .import = rk_ahash_import,
- + .digest = rk_ahash_digest,
- + .halg = {
- + .digestsize = SHA256_DIGEST_SIZE,
- + .statesize = sizeof(struct sha256_state),
- + .base = {
- + .cra_name = "sha256",
- + .cra_driver_name = "rk-sha256",
- + .cra_priority = 300,
- + .cra_flags = CRYPTO_ALG_ASYNC |
- + CRYPTO_ALG_NEED_FALLBACK,
- + .cra_blocksize = SHA256_BLOCK_SIZE,
- + .cra_ctxsize = sizeof(struct rk_ahash_ctx),
- + .cra_alignmask = 3,
- + .cra_init = rk_cra_hash_init,
- + .cra_exit = rk_cra_hash_exit,
- + .cra_module = THIS_MODULE,
- + }
- + }
- + }
- +};
- +
- +struct rk_crypto_tmp rk_ahash_md5 = {
- + .type = ALG_TYPE_HASH,
- + .alg.hash = {
- + .init = rk_ahash_init,
- + .update = rk_ahash_update,
- + .final = rk_ahash_final,
- + .finup = rk_ahash_finup,
- + .export = rk_ahash_export,
- + .import = rk_ahash_import,
- + .digest = rk_ahash_digest,
- + .halg = {
- + .digestsize = MD5_DIGEST_SIZE,
- + .statesize = sizeof(struct md5_state),
- + .base = {
- + .cra_name = "md5",
- + .cra_driver_name = "rk-md5",
- + .cra_priority = 300,
- + .cra_flags = CRYPTO_ALG_ASYNC |
- + CRYPTO_ALG_NEED_FALLBACK,
- + .cra_blocksize = SHA1_BLOCK_SIZE,
- + .cra_ctxsize = sizeof(struct rk_ahash_ctx),
- + .cra_alignmask = 3,
- + .cra_init = rk_cra_hash_init,
- + .cra_exit = rk_cra_hash_exit,
- + .cra_module = THIS_MODULE,
- + }
- + }
- + }
- +};
- commit 641eacd15696a65b08880985701c8082872da136
- Author: Zain Wang <wzz@rock-chips.com>
- Date: Mon Jul 24 09:23:13 2017 +0800
- crypto: rockchip - move the crypto completion from interrupt context
- It's illegal to call the completion function from hardirq context,
- it will cause runtime tests to fail. Let's build a new task (done_task)
- for moving update operation from hardirq context.
- Signed-off-by: zain wang <wzz@rock-chips.com>
- Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
- diff --git a/drivers/crypto/rockchip/rk3288_crypto.c b/drivers/crypto/rockchip/rk3288_crypto.c
- index d0f80c6..c2b1dd7 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto.c
- @@ -169,24 +169,22 @@ static irqreturn_t rk_crypto_irq_handle(int irq, void *dev_id)
- {
- struct rk_crypto_info *dev = platform_get_drvdata(dev_id);
- u32 interrupt_status;
- - int err = 0;
- spin_lock(&dev->lock);
- interrupt_status = CRYPTO_READ(dev, RK_CRYPTO_INTSTS);
- CRYPTO_WRITE(dev, RK_CRYPTO_INTSTS, interrupt_status);
- +
- if (interrupt_status & 0x0a) {
- dev_warn(dev->dev, "DMA Error\n");
- - err = -EFAULT;
- - } else if (interrupt_status & 0x05) {
- - err = dev->update(dev);
- + dev->err = -EFAULT;
- }
- - if (err)
- - dev->complete(dev, err);
- + tasklet_schedule(&dev->done_task);
- +
- spin_unlock(&dev->lock);
- return IRQ_HANDLED;
- }
- -static void rk_crypto_tasklet_cb(unsigned long data)
- +static void rk_crypto_queue_task_cb(unsigned long data)
- {
- struct rk_crypto_info *dev = (struct rk_crypto_info *)data;
- struct crypto_async_request *async_req, *backlog;
- @@ -210,11 +208,26 @@ static void rk_crypto_tasklet_cb(unsigned long data)
- dev->ablk_req = ablkcipher_request_cast(async_req);
- else
- dev->ahash_req = ahash_request_cast(async_req);
- + dev->err = 0;
- err = dev->start(dev);
- if (err)
- dev->complete(dev, err);
- }
- +static void rk_crypto_done_task_cb(unsigned long data)
- +{
- + struct rk_crypto_info *dev = (struct rk_crypto_info *)data;
- +
- + if (dev->err) {
- + dev->complete(dev, dev->err);
- + return;
- + }
- +
- + dev->err = dev->update(dev);
- + if (dev->err)
- + dev->complete(dev, dev->err);
- +}
- +
- static struct rk_crypto_tmp *rk_cipher_algs[] = {
- &rk_ecb_aes_alg,
- &rk_cbc_aes_alg,
- @@ -361,8 +374,10 @@ static int rk_crypto_probe(struct platform_device *pdev)
- crypto_info->dev = &pdev->dev;
- platform_set_drvdata(pdev, crypto_info);
- - tasklet_init(&crypto_info->crypto_tasklet,
- - rk_crypto_tasklet_cb, (unsigned long)crypto_info);
- + tasklet_init(&crypto_info->queue_task,
- + rk_crypto_queue_task_cb, (unsigned long)crypto_info);
- + tasklet_init(&crypto_info->done_task,
- + rk_crypto_done_task_cb, (unsigned long)crypto_info);
- crypto_init_queue(&crypto_info->queue, 50);
- crypto_info->enable_clk = rk_crypto_enable_clk;
- @@ -380,7 +395,8 @@ static int rk_crypto_probe(struct platform_device *pdev)
- return 0;
- err_register_alg:
- - tasklet_kill(&crypto_info->crypto_tasklet);
- + tasklet_kill(&crypto_info->queue_task);
- + tasklet_kill(&crypto_info->done_task);
- err_crypto:
- return err;
- }
- @@ -390,7 +406,8 @@ static int rk_crypto_remove(struct platform_device *pdev)
- struct rk_crypto_info *crypto_tmp = platform_get_drvdata(pdev);
- rk_crypto_unregister();
- - tasklet_kill(&crypto_tmp->crypto_tasklet);
- + tasklet_kill(&crypto_tmp->done_task);
- + tasklet_kill(&crypto_tmp->queue_task);
- return 0;
- }
- diff --git a/drivers/crypto/rockchip/rk3288_crypto.h b/drivers/crypto/rockchip/rk3288_crypto.h
- index d7b71fe..65ad1c2 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto.h
- +++ b/drivers/crypto/rockchip/rk3288_crypto.h
- @@ -190,9 +190,11 @@ struct rk_crypto_info {
- void __iomem *reg;
- int irq;
- struct crypto_queue queue;
- - struct tasklet_struct crypto_tasklet;
- + struct tasklet_struct queue_task;
- + struct tasklet_struct done_task;
- struct ablkcipher_request *ablk_req;
- struct ahash_request *ahash_req;
- + int err;
- /* device lock */
- spinlock_t lock;
- diff --git a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- index b5a3afe..8787e44 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- @@ -42,7 +42,7 @@ static int rk_handle_req(struct rk_crypto_info *dev,
- spin_lock_irqsave(&dev->lock, flags);
- err = ablkcipher_enqueue_request(&dev->queue, req);
- spin_unlock_irqrestore(&dev->lock, flags);
- - tasklet_schedule(&dev->crypto_tasklet);
- + tasklet_schedule(&dev->queue_task);
- return err;
- }
- diff --git a/drivers/crypto/rockchip/rk3288_crypto_ahash.c b/drivers/crypto/rockchip/rk3288_crypto_ahash.c
- index 7185882..9b55585 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto_ahash.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto_ahash.c
- @@ -204,7 +204,7 @@ static int rk_ahash_digest(struct ahash_request *req)
- ret = crypto_enqueue_request(&dev->queue, &req->base);
- spin_unlock_irqrestore(&dev->lock, flags);
- - tasklet_schedule(&dev->crypto_tasklet);
- + tasklet_schedule(&dev->queue_task);
- /*
- * it will take some time to process date after last dma transmission.
- commit 9a42e4eed3fcd7ba8dff6622384cd08bfe5ef707
- Author: Zain Wang <wzz@rock-chips.com>
- Date: Mon Jul 24 09:23:14 2017 +0800
- crypto: rockchip - return the err code when unable dequeue the crypto request
- Sometime we would unable to dequeue the crypto request, in this case,
- we should finish crypto and return the err code.
- Signed-off-by: zain wang <wzz@rock-chips.com>
- Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
- diff --git a/drivers/crypto/rockchip/rk3288_crypto.c b/drivers/crypto/rockchip/rk3288_crypto.c
- index c2b1dd7..57c3783 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto.c
- @@ -187,27 +187,8 @@ static irqreturn_t rk_crypto_irq_handle(int irq, void *dev_id)
- static void rk_crypto_queue_task_cb(unsigned long data)
- {
- struct rk_crypto_info *dev = (struct rk_crypto_info *)data;
- - struct crypto_async_request *async_req, *backlog;
- - unsigned long flags;
- int err = 0;
- - spin_lock_irqsave(&dev->lock, flags);
- - backlog = crypto_get_backlog(&dev->queue);
- - async_req = crypto_dequeue_request(&dev->queue);
- - spin_unlock_irqrestore(&dev->lock, flags);
- - if (!async_req) {
- - dev_err(dev->dev, "async_req is NULL !!\n");
- - return;
- - }
- - if (backlog) {
- - backlog->complete(backlog, -EINPROGRESS);
- - backlog = NULL;
- - }
- -
- - if (crypto_tfm_alg_type(async_req->tfm) == CRYPTO_ALG_TYPE_ABLKCIPHER)
- - dev->ablk_req = ablkcipher_request_cast(async_req);
- - else
- - dev->ahash_req = ahash_request_cast(async_req);
- dev->err = 0;
- err = dev->start(dev);
- if (err)
- diff --git a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- index 8787e44..dbe78de 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- @@ -25,6 +25,7 @@ static int rk_handle_req(struct rk_crypto_info *dev,
- struct ablkcipher_request *req)
- {
- unsigned long flags;
- + struct crypto_async_request *async_req, *backlog;
- int err;
- if (!IS_ALIGNED(req->nbytes, dev->align_size))
- @@ -41,7 +42,21 @@ static int rk_handle_req(struct rk_crypto_info *dev,
- spin_lock_irqsave(&dev->lock, flags);
- err = ablkcipher_enqueue_request(&dev->queue, req);
- + backlog = crypto_get_backlog(&dev->queue);
- + async_req = crypto_dequeue_request(&dev->queue);
- spin_unlock_irqrestore(&dev->lock, flags);
- +
- + if (!async_req) {
- + dev_err(dev->dev, "async_req is NULL !!\n");
- + return err;
- + }
- + if (backlog) {
- + backlog->complete(backlog, -EINPROGRESS);
- + backlog = NULL;
- + }
- +
- + dev->ablk_req = ablkcipher_request_cast(async_req);
- +
- tasklet_schedule(&dev->queue_task);
- return err;
- }
- diff --git a/drivers/crypto/rockchip/rk3288_crypto_ahash.c b/drivers/crypto/rockchip/rk3288_crypto_ahash.c
- index 9b55585..ebc46e0 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto_ahash.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto_ahash.c
- @@ -166,6 +166,7 @@ static int rk_ahash_digest(struct ahash_request *req)
- {
- struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
- struct rk_ahash_ctx *tctx = crypto_tfm_ctx(req->base.tfm);
- + struct crypto_async_request *async_req, *backlog;
- struct rk_crypto_info *dev = NULL;
- unsigned long flags;
- int ret;
- @@ -202,8 +203,21 @@ static int rk_ahash_digest(struct ahash_request *req)
- spin_lock_irqsave(&dev->lock, flags);
- ret = crypto_enqueue_request(&dev->queue, &req->base);
- + backlog = crypto_get_backlog(&dev->queue);
- + async_req = crypto_dequeue_request(&dev->queue);
- spin_unlock_irqrestore(&dev->lock, flags);
- + if (!async_req) {
- + dev_err(dev->dev, "async_req is NULL !!\n");
- + return ret;
- + }
- + if (backlog) {
- + backlog->complete(backlog, -EINPROGRESS);
- + backlog = NULL;
- + }
- +
- + dev->ahash_req = ahash_request_cast(async_req);
- +
- tasklet_schedule(&dev->queue_task);
- /*
- commit 5a7801f6634b1e2888bcb1a85bedc50e46dcd757
- Author: Zain Wang <wzz@rock-chips.com>
- Date: Tue Aug 15 15:48:15 2017 +0800
- crypto: rockchip - Don't dequeue the request when device is busy
- The device can only process one request at a time. So if multiple
- requests came at the same time, we can enqueue them first, and
- dequeue them one by one when the device is idle.
- Signed-off-by: zain wang <wzz@rock-chips.com>
- Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
- diff --git a/drivers/crypto/rockchip/rk3288_crypto.c b/drivers/crypto/rockchip/rk3288_crypto.c
- index 57c3783..c9d622a 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto.c
- @@ -184,15 +184,53 @@ static irqreturn_t rk_crypto_irq_handle(int irq, void *dev_id)
- return IRQ_HANDLED;
- }
- +static int rk_crypto_enqueue(struct rk_crypto_info *dev,
- + struct crypto_async_request *async_req)
- +{
- + unsigned long flags;
- + int ret;
- +
- + spin_lock_irqsave(&dev->lock, flags);
- + ret = crypto_enqueue_request(&dev->queue, async_req);
- + if (dev->busy) {
- + spin_unlock_irqrestore(&dev->lock, flags);
- + return ret;
- + }
- + dev->busy = true;
- + spin_unlock_irqrestore(&dev->lock, flags);
- + tasklet_schedule(&dev->queue_task);
- +
- + return ret;
- +}
- +
- static void rk_crypto_queue_task_cb(unsigned long data)
- {
- struct rk_crypto_info *dev = (struct rk_crypto_info *)data;
- + struct crypto_async_request *async_req, *backlog;
- + unsigned long flags;
- int err = 0;
- dev->err = 0;
- + spin_lock_irqsave(&dev->lock, flags);
- + backlog = crypto_get_backlog(&dev->queue);
- + async_req = crypto_dequeue_request(&dev->queue);
- +
- + if (!async_req) {
- + dev->busy = false;
- + spin_unlock_irqrestore(&dev->lock, flags);
- + return;
- + }
- + spin_unlock_irqrestore(&dev->lock, flags);
- +
- + if (backlog) {
- + backlog->complete(backlog, -EINPROGRESS);
- + backlog = NULL;
- + }
- +
- + dev->async_req = async_req;
- err = dev->start(dev);
- if (err)
- - dev->complete(dev, err);
- + dev->complete(dev->async_req, err);
- }
- static void rk_crypto_done_task_cb(unsigned long data)
- @@ -200,13 +238,13 @@ static void rk_crypto_done_task_cb(unsigned long data)
- struct rk_crypto_info *dev = (struct rk_crypto_info *)data;
- if (dev->err) {
- - dev->complete(dev, dev->err);
- + dev->complete(dev->async_req, dev->err);
- return;
- }
- dev->err = dev->update(dev);
- if (dev->err)
- - dev->complete(dev, dev->err);
- + dev->complete(dev->async_req, dev->err);
- }
- static struct rk_crypto_tmp *rk_cipher_algs[] = {
- @@ -365,6 +403,8 @@ static int rk_crypto_probe(struct platform_device *pdev)
- crypto_info->disable_clk = rk_crypto_disable_clk;
- crypto_info->load_data = rk_load_data;
- crypto_info->unload_data = rk_unload_data;
- + crypto_info->enqueue = rk_crypto_enqueue;
- + crypto_info->busy = false;
- err = rk_crypto_register(crypto_info);
- if (err) {
- diff --git a/drivers/crypto/rockchip/rk3288_crypto.h b/drivers/crypto/rockchip/rk3288_crypto.h
- index 65ad1c2..ab6a1b4 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto.h
- +++ b/drivers/crypto/rockchip/rk3288_crypto.h
- @@ -192,8 +192,7 @@ struct rk_crypto_info {
- struct crypto_queue queue;
- struct tasklet_struct queue_task;
- struct tasklet_struct done_task;
- - struct ablkcipher_request *ablk_req;
- - struct ahash_request *ahash_req;
- + struct crypto_async_request *async_req;
- int err;
- /* device lock */
- spinlock_t lock;
- @@ -210,18 +209,20 @@ struct rk_crypto_info {
- size_t nents;
- unsigned int total;
- unsigned int count;
- - u32 mode;
- dma_addr_t addr_in;
- dma_addr_t addr_out;
- + bool busy;
- int (*start)(struct rk_crypto_info *dev);
- int (*update)(struct rk_crypto_info *dev);
- - void (*complete)(struct rk_crypto_info *dev, int err);
- + void (*complete)(struct crypto_async_request *base, int err);
- int (*enable_clk)(struct rk_crypto_info *dev);
- void (*disable_clk)(struct rk_crypto_info *dev);
- int (*load_data)(struct rk_crypto_info *dev,
- struct scatterlist *sg_src,
- struct scatterlist *sg_dst);
- void (*unload_data)(struct rk_crypto_info *dev);
- + int (*enqueue)(struct rk_crypto_info *dev,
- + struct crypto_async_request *async_req);
- };
- /* the private variable of hash */
- @@ -234,12 +235,14 @@ struct rk_ahash_ctx {
- /* the privete variable of hash for fallback */
- struct rk_ahash_rctx {
- struct ahash_request fallback_req;
- + u32 mode;
- };
- /* the private variable of cipher */
- struct rk_cipher_ctx {
- struct rk_crypto_info *dev;
- unsigned int keylen;
- + u32 mode;
- };
- enum alg_type {
- diff --git a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- index dbe78de..639c15c 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c
- @@ -15,50 +15,19 @@
- #define RK_CRYPTO_DEC BIT(0)
- -static void rk_crypto_complete(struct rk_crypto_info *dev, int err)
- +static void rk_crypto_complete(struct crypto_async_request *base, int err)
- {
- - if (dev->ablk_req->base.complete)
- - dev->ablk_req->base.complete(&dev->ablk_req->base, err);
- + if (base->complete)
- + base->complete(base, err);
- }
- static int rk_handle_req(struct rk_crypto_info *dev,
- struct ablkcipher_request *req)
- {
- - unsigned long flags;
- - struct crypto_async_request *async_req, *backlog;
- - int err;
- -
- if (!IS_ALIGNED(req->nbytes, dev->align_size))
- return -EINVAL;
- -
- - dev->left_bytes = req->nbytes;
- - dev->total = req->nbytes;
- - dev->sg_src = req->src;
- - dev->first = req->src;
- - dev->nents = sg_nents(req->src);
- - dev->sg_dst = req->dst;
- - dev->aligned = 1;
- - dev->ablk_req = req;
- -
- - spin_lock_irqsave(&dev->lock, flags);
- - err = ablkcipher_enqueue_request(&dev->queue, req);
- - backlog = crypto_get_backlog(&dev->queue);
- - async_req = crypto_dequeue_request(&dev->queue);
- - spin_unlock_irqrestore(&dev->lock, flags);
- -
- - if (!async_req) {
- - dev_err(dev->dev, "async_req is NULL !!\n");
- - return err;
- - }
- - if (backlog) {
- - backlog->complete(backlog, -EINPROGRESS);
- - backlog = NULL;
- - }
- -
- - dev->ablk_req = ablkcipher_request_cast(async_req);
- -
- - tasklet_schedule(&dev->queue_task);
- - return err;
- + else
- + return dev->enqueue(dev, &req->base);
- }
- static int rk_aes_setkey(struct crypto_ablkcipher *cipher,
- @@ -108,7 +77,7 @@ static int rk_aes_ecb_encrypt(struct ablkcipher_request *req)
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- struct rk_crypto_info *dev = ctx->dev;
- - dev->mode = RK_CRYPTO_AES_ECB_MODE;
- + ctx->mode = RK_CRYPTO_AES_ECB_MODE;
- return rk_handle_req(dev, req);
- }
- @@ -118,7 +87,7 @@ static int rk_aes_ecb_decrypt(struct ablkcipher_request *req)
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- struct rk_crypto_info *dev = ctx->dev;
- - dev->mode = RK_CRYPTO_AES_ECB_MODE | RK_CRYPTO_DEC;
- + ctx->mode = RK_CRYPTO_AES_ECB_MODE | RK_CRYPTO_DEC;
- return rk_handle_req(dev, req);
- }
- @@ -128,7 +97,7 @@ static int rk_aes_cbc_encrypt(struct ablkcipher_request *req)
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- struct rk_crypto_info *dev = ctx->dev;
- - dev->mode = RK_CRYPTO_AES_CBC_MODE;
- + ctx->mode = RK_CRYPTO_AES_CBC_MODE;
- return rk_handle_req(dev, req);
- }
- @@ -138,7 +107,7 @@ static int rk_aes_cbc_decrypt(struct ablkcipher_request *req)
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- struct rk_crypto_info *dev = ctx->dev;
- - dev->mode = RK_CRYPTO_AES_CBC_MODE | RK_CRYPTO_DEC;
- + ctx->mode = RK_CRYPTO_AES_CBC_MODE | RK_CRYPTO_DEC;
- return rk_handle_req(dev, req);
- }
- @@ -148,7 +117,7 @@ static int rk_des_ecb_encrypt(struct ablkcipher_request *req)
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- struct rk_crypto_info *dev = ctx->dev;
- - dev->mode = 0;
- + ctx->mode = 0;
- return rk_handle_req(dev, req);
- }
- @@ -158,7 +127,7 @@ static int rk_des_ecb_decrypt(struct ablkcipher_request *req)
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- struct rk_crypto_info *dev = ctx->dev;
- - dev->mode = RK_CRYPTO_DEC;
- + ctx->mode = RK_CRYPTO_DEC;
- return rk_handle_req(dev, req);
- }
- @@ -168,7 +137,7 @@ static int rk_des_cbc_encrypt(struct ablkcipher_request *req)
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- struct rk_crypto_info *dev = ctx->dev;
- - dev->mode = RK_CRYPTO_TDES_CHAINMODE_CBC;
- + ctx->mode = RK_CRYPTO_TDES_CHAINMODE_CBC;
- return rk_handle_req(dev, req);
- }
- @@ -178,7 +147,7 @@ static int rk_des_cbc_decrypt(struct ablkcipher_request *req)
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- struct rk_crypto_info *dev = ctx->dev;
- - dev->mode = RK_CRYPTO_TDES_CHAINMODE_CBC | RK_CRYPTO_DEC;
- + ctx->mode = RK_CRYPTO_TDES_CHAINMODE_CBC | RK_CRYPTO_DEC;
- return rk_handle_req(dev, req);
- }
- @@ -188,7 +157,7 @@ static int rk_des3_ede_ecb_encrypt(struct ablkcipher_request *req)
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- struct rk_crypto_info *dev = ctx->dev;
- - dev->mode = RK_CRYPTO_TDES_SELECT;
- + ctx->mode = RK_CRYPTO_TDES_SELECT;
- return rk_handle_req(dev, req);
- }
- @@ -198,7 +167,7 @@ static int rk_des3_ede_ecb_decrypt(struct ablkcipher_request *req)
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- struct rk_crypto_info *dev = ctx->dev;
- - dev->mode = RK_CRYPTO_TDES_SELECT | RK_CRYPTO_DEC;
- + ctx->mode = RK_CRYPTO_TDES_SELECT | RK_CRYPTO_DEC;
- return rk_handle_req(dev, req);
- }
- @@ -208,7 +177,7 @@ static int rk_des3_ede_cbc_encrypt(struct ablkcipher_request *req)
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- struct rk_crypto_info *dev = ctx->dev;
- - dev->mode = RK_CRYPTO_TDES_SELECT | RK_CRYPTO_TDES_CHAINMODE_CBC;
- + ctx->mode = RK_CRYPTO_TDES_SELECT | RK_CRYPTO_TDES_CHAINMODE_CBC;
- return rk_handle_req(dev, req);
- }
- @@ -218,15 +187,16 @@ static int rk_des3_ede_cbc_decrypt(struct ablkcipher_request *req)
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
- struct rk_crypto_info *dev = ctx->dev;
- - dev->mode = RK_CRYPTO_TDES_SELECT | RK_CRYPTO_TDES_CHAINMODE_CBC |
- + ctx->mode = RK_CRYPTO_TDES_SELECT | RK_CRYPTO_TDES_CHAINMODE_CBC |
- RK_CRYPTO_DEC;
- return rk_handle_req(dev, req);
- }
- static void rk_ablk_hw_init(struct rk_crypto_info *dev)
- {
- - struct crypto_ablkcipher *cipher =
- - crypto_ablkcipher_reqtfm(dev->ablk_req);
- + struct ablkcipher_request *req =
- + ablkcipher_request_cast(dev->async_req);
- + struct crypto_ablkcipher *cipher = crypto_ablkcipher_reqtfm(req);
- struct crypto_tfm *tfm = crypto_ablkcipher_tfm(cipher);
- struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(cipher);
- u32 ivsize, block, conf_reg = 0;
- @@ -235,25 +205,23 @@ static void rk_ablk_hw_init(struct rk_crypto_info *dev)
- ivsize = crypto_ablkcipher_ivsize(cipher);
- if (block == DES_BLOCK_SIZE) {
- - dev->mode |= RK_CRYPTO_TDES_FIFO_MODE |
- + ctx->mode |= RK_CRYPTO_TDES_FIFO_MODE |
- RK_CRYPTO_TDES_BYTESWAP_KEY |
- RK_CRYPTO_TDES_BYTESWAP_IV;
- - CRYPTO_WRITE(dev, RK_CRYPTO_TDES_CTRL, dev->mode);
- - memcpy_toio(dev->reg + RK_CRYPTO_TDES_IV_0,
- - dev->ablk_req->info, ivsize);
- + CRYPTO_WRITE(dev, RK_CRYPTO_TDES_CTRL, ctx->mode);
- + memcpy_toio(dev->reg + RK_CRYPTO_TDES_IV_0, req->info, ivsize);
- conf_reg = RK_CRYPTO_DESSEL;
- } else {
- - dev->mode |= RK_CRYPTO_AES_FIFO_MODE |
- + ctx->mode |= RK_CRYPTO_AES_FIFO_MODE |
- RK_CRYPTO_AES_KEY_CHANGE |
- RK_CRYPTO_AES_BYTESWAP_KEY |
- RK_CRYPTO_AES_BYTESWAP_IV;
- if (ctx->keylen == AES_KEYSIZE_192)
- - dev->mode |= RK_CRYPTO_AES_192BIT_key;
- + ctx->mode |= RK_CRYPTO_AES_192BIT_key;
- else if (ctx->keylen == AES_KEYSIZE_256)
- - dev->mode |= RK_CRYPTO_AES_256BIT_key;
- - CRYPTO_WRITE(dev, RK_CRYPTO_AES_CTRL, dev->mode);
- - memcpy_toio(dev->reg + RK_CRYPTO_AES_IV_0,
- - dev->ablk_req->info, ivsize);
- + ctx->mode |= RK_CRYPTO_AES_256BIT_key;
- + CRYPTO_WRITE(dev, RK_CRYPTO_AES_CTRL, ctx->mode);
- + memcpy_toio(dev->reg + RK_CRYPTO_AES_IV_0, req->info, ivsize);
- }
- conf_reg |= RK_CRYPTO_BYTESWAP_BTFIFO |
- RK_CRYPTO_BYTESWAP_BRFIFO;
- @@ -283,8 +251,18 @@ static int rk_set_data_start(struct rk_crypto_info *dev)
- static int rk_ablk_start(struct rk_crypto_info *dev)
- {
- + struct ablkcipher_request *req =
- + ablkcipher_request_cast(dev->async_req);
- unsigned long flags;
- - int err;
- + int err = 0;
- +
- + dev->left_bytes = req->nbytes;
- + dev->total = req->nbytes;
- + dev->sg_src = req->src;
- + dev->first = req->src;
- + dev->nents = sg_nents(req->src);
- + dev->sg_dst = req->dst;
- + dev->aligned = 1;
- spin_lock_irqsave(&dev->lock, flags);
- rk_ablk_hw_init(dev);
- @@ -295,15 +273,16 @@ static int rk_ablk_start(struct rk_crypto_info *dev)
- static void rk_iv_copyback(struct rk_crypto_info *dev)
- {
- - struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(dev->ablk_req);
- + struct ablkcipher_request *req =
- + ablkcipher_request_cast(dev->async_req);
- + struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
- u32 ivsize = crypto_ablkcipher_ivsize(tfm);
- if (ivsize == DES_BLOCK_SIZE)
- - memcpy_fromio(dev->ablk_req->info,
- - dev->reg + RK_CRYPTO_TDES_IV_0, ivsize);
- + memcpy_fromio(req->info, dev->reg + RK_CRYPTO_TDES_IV_0,
- + ivsize);
- else if (ivsize == AES_BLOCK_SIZE)
- - memcpy_fromio(dev->ablk_req->info,
- - dev->reg + RK_CRYPTO_AES_IV_0, ivsize);
- + memcpy_fromio(req->info, dev->reg + RK_CRYPTO_AES_IV_0, ivsize);
- }
- /* return:
- @@ -313,10 +292,12 @@ static void rk_iv_copyback(struct rk_crypto_info *dev)
- static int rk_ablk_rx(struct rk_crypto_info *dev)
- {
- int err = 0;
- + struct ablkcipher_request *req =
- + ablkcipher_request_cast(dev->async_req);
- dev->unload_data(dev);
- if (!dev->aligned) {
- - if (!sg_pcopy_from_buffer(dev->ablk_req->dst, dev->nents,
- + if (!sg_pcopy_from_buffer(req->dst, dev->nents,
- dev->addr_vir, dev->count,
- dev->total - dev->left_bytes -
- dev->count)) {
- @@ -339,7 +320,8 @@ static int rk_ablk_rx(struct rk_crypto_info *dev)
- } else {
- rk_iv_copyback(dev);
- /* here show the calculation is over without any err */
- - dev->complete(dev, 0);
- + dev->complete(dev->async_req, 0);
- + tasklet_schedule(&dev->queue_task);
- }
- out_rx:
- return err;
- diff --git a/drivers/crypto/rockchip/rk3288_crypto_ahash.c b/drivers/crypto/rockchip/rk3288_crypto_ahash.c
- index ebc46e0..821a506 100644
- --- a/drivers/crypto/rockchip/rk3288_crypto_ahash.c
- +++ b/drivers/crypto/rockchip/rk3288_crypto_ahash.c
- @@ -40,14 +40,16 @@ static int zero_message_process(struct ahash_request *req)
- return 0;
- }
- -static void rk_ahash_crypto_complete(struct rk_crypto_info *dev, int err)
- +static void rk_ahash_crypto_complete(struct crypto_async_request *base, int err)
- {
- - if (dev->ahash_req->base.complete)
- - dev->ahash_req->base.complete(&dev->ahash_req->base, err);
- + if (base->complete)
- + base->complete(base, err);
- }
- static void rk_ahash_reg_init(struct rk_crypto_info *dev)
- {
- + struct ahash_request *req = ahash_request_cast(dev->async_req);
- + struct rk_ahash_rctx *rctx = ahash_request_ctx(req);
- int reg_status = 0;
- reg_status = CRYPTO_READ(dev, RK_CRYPTO_CTRL) |
- @@ -67,7 +69,7 @@ static void rk_ahash_reg_init(struct rk_crypto_info *dev)
- CRYPTO_WRITE(dev, RK_CRYPTO_INTSTS, RK_CRYPTO_HRDMA_ERR_INT |
- RK_CRYPTO_HRDMA_DONE_INT);
- - CRYPTO_WRITE(dev, RK_CRYPTO_HASH_CTRL, dev->mode |
- + CRYPTO_WRITE(dev, RK_CRYPTO_HASH_CTRL, rctx->mode |
- RK_CRYPTO_HASH_SWAP_DO);
- CRYPTO_WRITE(dev, RK_CRYPTO_CONF, RK_CRYPTO_BYTESWAP_HRFIFO |
- @@ -164,78 +166,13 @@ static int rk_ahash_export(struct ahash_request *req, void *out)
- static int rk_ahash_digest(struct ahash_request *req)
- {
- - struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
- struct rk_ahash_ctx *tctx = crypto_tfm_ctx(req->base.tfm);
- - struct crypto_async_request *async_req, *backlog;
- - struct rk_crypto_info *dev = NULL;
- - unsigned long flags;
- - int ret;
- + struct rk_crypto_info *dev = tctx->dev;
- if (!req->nbytes)
- return zero_message_process(req);
- -
- - dev = tctx->dev;
- - dev->total = req->nbytes;
- - dev->left_bytes = req->nbytes;
- - dev->aligned = 0;
- - dev->mode = 0;
- - dev->align_size = 4;
- - dev->sg_dst = NULL;
- - dev->sg_src = req->src;
- - dev->first = req->src;
- - dev->nents = sg_nents(req->src);
- -
- - switch (crypto_ahash_digestsize(tfm)) {
- - case SHA1_DIGEST_SIZE:
- - dev->mode = RK_CRYPTO_HASH_SHA1;
- - break;
- - case SHA256_DIGEST_SIZE:
- - dev->mode = RK_CRYPTO_HASH_SHA256;
- - break;
- - case MD5_DIGEST_SIZE:
- - dev->mode = RK_CRYPTO_HASH_MD5;
- - break;
- - default:
- - return -EINVAL;
- - }
- -
- - rk_ahash_reg_init(dev);
- -
- - spin_lock_irqsave(&dev->lock, flags);
- - ret = crypto_enqueue_request(&dev->queue, &req->base);
- - backlog = crypto_get_backlog(&dev->queue);
- - async_req = crypto_dequeue_request(&dev->queue);
- - spin_unlock_irqrestore(&dev->lock, flags);
- -
- - if (!async_req) {
- - dev_err(dev->dev, "async_req is NULL !!\n");
- - return ret;
- - }
- - if (backlog) {
- - backlog->complete(backlog, -EINPROGRESS);
- - backlog = NULL;
- - }
- -
- - dev->ahash_req = ahash_request_cast(async_req);
- -
- - tasklet_schedule(&dev->queue_task);
- -
- - /*
- - * it will take some time to process date after last dma transmission.
- - *
- - * waiting time is relative with the last date len,
- - * so cannot set a fixed time here.
- - * 10-50 makes system not call here frequently wasting
- - * efficiency, and make it response quickly when dma
- - * complete.
- - */
- - while (!CRYPTO_READ(dev, RK_CRYPTO_HASH_STS))
- - usleep_range(10, 50);
- -
- - memcpy_fromio(req->result, dev->reg + RK_CRYPTO_HASH_DOUT_0,
- - crypto_ahash_digestsize(tfm));
- -
- - return 0;
- + else
- + return dev->enqueue(dev, &req->base);
- }
- static void crypto_ahash_dma_start(struct rk_crypto_info *dev)
- @@ -258,12 +195,45 @@ static int rk_ahash_set_data_start(struct rk_crypto_info *dev)
- static int rk_ahash_start(struct rk_crypto_info *dev)
- {
- + struct ahash_request *req = ahash_request_cast(dev->async_req);
- + struct crypto_ahash *tfm;
- + struct rk_ahash_rctx *rctx;
- +
- + dev->total = req->nbytes;
- + dev->left_bytes = req->nbytes;
- + dev->aligned = 0;
- + dev->align_size = 4;
- + dev->sg_dst = NULL;
- + dev->sg_src = req->src;
- + dev->first = req->src;
- + dev->nents = sg_nents(req->src);
- + rctx = ahash_request_ctx(req);
- + rctx->mode = 0;
- +
- + tfm = crypto_ahash_reqtfm(req);
- + switch (crypto_ahash_digestsize(tfm)) {
- + case SHA1_DIGEST_SIZE:
- + rctx->mode = RK_CRYPTO_HASH_SHA1;
- + break;
- + case SHA256_DIGEST_SIZE:
- + rctx->mode = RK_CRYPTO_HASH_SHA256;
- + break;
- + case MD5_DIGEST_SIZE:
- + rctx->mode = RK_CRYPTO_HASH_MD5;
- + break;
- + default:
- + return -EINVAL;
- + }
- +
- + rk_ahash_reg_init(dev);
- return rk_ahash_set_data_start(dev);
- }
- static int rk_ahash_crypto_rx(struct rk_crypto_info *dev)
- {
- int err = 0;
- + struct ahash_request *req = ahash_request_cast(dev->async_req);
- + struct crypto_ahash *tfm;
- dev->unload_data(dev);
- if (dev->left_bytes) {
- @@ -278,7 +248,24 @@ static int rk_ahash_crypto_rx(struct rk_crypto_info *dev)
- }
- err = rk_ahash_set_data_start(dev);
- } else {
- - dev->complete(dev, 0);
- + /*
- + * it will take some time to process date after last dma
- + * transmission.
- + *
- + * waiting time is relative with the last date len,
- + * so cannot set a fixed time here.
- + * 10us makes system not call here frequently wasting
- + * efficiency, and make it response quickly when dma
- + * complete.
- + */
- + while (!CRYPTO_READ(dev, RK_CRYPTO_HASH_STS))
- + udelay(10);
- +
- + tfm = crypto_ahash_reqtfm(req);
- + memcpy_fromio(req->result, dev->reg + RK_CRYPTO_HASH_DOUT_0,
- + crypto_ahash_digestsize(tfm));
- + dev->complete(dev->async_req, 0);
- + tasklet_schedule(&dev->queue_task);
- }
- out_rx:
- commit 0c4c78de0417ced1da92351a3013e631860ea576
- Author: LABBE Corentin <clabbe.montjoie@gmail.com>
- Date: Thu Dec 17 13:45:39 2015 +0100
- crypto: hash - add zero length message hash for shax and md5
- Some crypto drivers cannot process empty data message and return a
- precalculated hash for md5/sha1/sha224/sha256.
- This patch add thoses precalculated hash in include/crypto.
- Signed-off-by: LABBE Corentin <clabbe.montjoie@gmail.com>
- Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
- diff --git a/crypto/md5.c b/crypto/md5.c
- index 33d17e9..2355a7c 100644
- --- a/crypto/md5.c
- +++ b/crypto/md5.c
- @@ -24,6 +24,12 @@
- #include <linux/cryptohash.h>
- #include <asm/byteorder.h>
- +const u8 md5_zero_message_hash[MD5_DIGEST_SIZE] = {
- + 0xd4, 0x1d, 0x8c, 0xd9, 0x8f, 0x00, 0xb2, 0x04,
- + 0xe9, 0x80, 0x09, 0x98, 0xec, 0xf8, 0x42, 0x7e,
- +};
- +EXPORT_SYMBOL_GPL(md5_zero_message_hash);
- +
- /* XXX: this stuff can be optimized */
- static inline void le32_to_cpu_array(u32 *buf, unsigned int words)
- {
- diff --git a/crypto/sha1_generic.c b/crypto/sha1_generic.c
- index 39e3acc..6877cbb 100644
- --- a/crypto/sha1_generic.c
- +++ b/crypto/sha1_generic.c
- @@ -26,6 +26,13 @@
- #include <crypto/sha1_base.h>
- #include <asm/byteorder.h>
- +const u8 sha1_zero_message_hash[SHA1_DIGEST_SIZE] = {
- + 0xda, 0x39, 0xa3, 0xee, 0x5e, 0x6b, 0x4b, 0x0d,
- + 0x32, 0x55, 0xbf, 0xef, 0x95, 0x60, 0x18, 0x90,
- + 0xaf, 0xd8, 0x07, 0x09
- +};
- +EXPORT_SYMBOL_GPL(sha1_zero_message_hash);
- +
- static void sha1_generic_block_fn(struct sha1_state *sst, u8 const *src,
- int blocks)
- {
- diff --git a/crypto/sha256_generic.c b/crypto/sha256_generic.c
- index 7843116..8f9c47e 100644
- --- a/crypto/sha256_generic.c
- +++ b/crypto/sha256_generic.c
- @@ -27,6 +27,22 @@
- #include <asm/byteorder.h>
- #include <asm/unaligned.h>
- +const u8 sha224_zero_message_hash[SHA224_DIGEST_SIZE] = {
- + 0xd1, 0x4a, 0x02, 0x8c, 0x2a, 0x3a, 0x2b, 0xc9, 0x47,
- + 0x61, 0x02, 0xbb, 0x28, 0x82, 0x34, 0xc4, 0x15, 0xa2,
- + 0xb0, 0x1f, 0x82, 0x8e, 0xa6, 0x2a, 0xc5, 0xb3, 0xe4,
- + 0x2f
- +};
- +EXPORT_SYMBOL_GPL(sha224_zero_message_hash);
- +
- +const u8 sha256_zero_message_hash[SHA256_DIGEST_SIZE] = {
- + 0xe3, 0xb0, 0xc4, 0x42, 0x98, 0xfc, 0x1c, 0x14,
- + 0x9a, 0xfb, 0xf4, 0xc8, 0x99, 0x6f, 0xb9, 0x24,
- + 0x27, 0xae, 0x41, 0xe4, 0x64, 0x9b, 0x93, 0x4c,
- + 0xa4, 0x95, 0x99, 0x1b, 0x78, 0x52, 0xb8, 0x55
- +};
- +EXPORT_SYMBOL_GPL(sha256_zero_message_hash);
- +
- static inline u32 Ch(u32 x, u32 y, u32 z)
- {
- return z ^ (x & (y ^ z));
- diff --git a/include/crypto/md5.h b/include/crypto/md5.h
- index 146af825..327deac 100644
- --- a/include/crypto/md5.h
- +++ b/include/crypto/md5.h
- @@ -13,6 +13,8 @@
- #define MD5_H2 0x98badcfeUL
- #define MD5_H3 0x10325476UL
- +extern const u8 md5_zero_message_hash[MD5_DIGEST_SIZE];
- +
- struct md5_state {
- u32 hash[MD5_HASH_WORDS];
- u32 block[MD5_BLOCK_WORDS];
- diff --git a/include/crypto/sha.h b/include/crypto/sha.h
- index dd7905a..c94d3eb 100644
- --- a/include/crypto/sha.h
- +++ b/include/crypto/sha.h
- @@ -64,6 +64,12 @@
- #define SHA512_H6 0x1f83d9abfb41bd6bULL
- #define SHA512_H7 0x5be0cd19137e2179ULL
- +extern const u8 sha1_zero_message_hash[SHA1_DIGEST_SIZE];
- +
- +extern const u8 sha224_zero_message_hash[SHA224_DIGEST_SIZE];
- +
- +extern const u8 sha256_zero_message_hash[SHA256_DIGEST_SIZE];
- +
- struct sha1_state {
- u32 state[SHA1_DIGEST_SIZE / 4];
- u64 count;
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement