On Fri, Oct 30, 2015 at 04:22:46PM +0800, Zain Wang wrote:
> Crypto driver support cbc/ecb two chainmode, and aes/des/des3 three cipher 
> mode.
> The names registered are:
>     ecb(aes) cbc(aes) ecb(des) cbc(des) ecb(des3_ede) cbc(des3_ede)
> You can alloc tags above in your case.
> 
> And other algorithms and platforms will be added later on.
> 
> Signed-off-by: Zain Wang <zain.w...@rock-chips.com>
> ---
>  drivers/crypto/Makefile                            |   1 +
>  drivers/crypto/rk_crypto/Makefile                  |   3 +
>  drivers/crypto/rk_crypto/rk3288_crypto.c           | 393 ++++++++++++++++
>  drivers/crypto/rk_crypto/rk3288_crypto.h           | 291 ++++++++++++
>  .../crypto/rk_crypto/rk3288_crypto_ablkcipher.c    | 502 
> +++++++++++++++++++++
>  5 files changed, 1190 insertions(+)
>  create mode 100644 drivers/crypto/rk_crypto/Makefile
>  create mode 100644 drivers/crypto/rk_crypto/rk3288_crypto.c
>  create mode 100644 drivers/crypto/rk_crypto/rk3288_crypto.h
>  create mode 100644 drivers/crypto/rk_crypto/rk3288_crypto_ablkcipher.c
> 
> diff --git a/drivers/crypto/Makefile b/drivers/crypto/Makefile
> index c3ced6f..00d103c 100644
> --- a/drivers/crypto/Makefile
> +++ b/drivers/crypto/Makefile
> @@ -29,3 +29,4 @@ obj-$(CONFIG_CRYPTO_DEV_QAT) += qat/
>  obj-$(CONFIG_CRYPTO_DEV_QCE) += qce/
>  obj-$(CONFIG_CRYPTO_DEV_VMX) += vmx/
>  obj-$(CONFIG_CRYPTO_DEV_SUN4I_SS) += sunxi-ss/
> +obj-$(CONFIG_CRYPTO_DEV_RK3288) += rk_crypto/
> diff --git a/drivers/crypto/rk_crypto/Makefile 
> b/drivers/crypto/rk_crypto/Makefile
> new file mode 100644
> index 0000000..0f62d87
> --- /dev/null
> +++ b/drivers/crypto/rk_crypto/Makefile
> @@ -0,0 +1,3 @@
> +obj-$(CONFIG_CRYPTO_DEV_RK3288) += rk_crypto_driver.o
> +rk_crypto_driver-objs := rk3288_crypto.o \
> +                      rk3288_crypto_ablkcipher.o \
> diff --git a/drivers/crypto/rk_crypto/rk3288_crypto.c 
> b/drivers/crypto/rk_crypto/rk3288_crypto.c
> new file mode 100644
> index 0000000..fe55d7e
> --- /dev/null
> +++ b/drivers/crypto/rk_crypto/rk3288_crypto.c
> @@ -0,0 +1,393 @@
> +/*
> + *Crypto acceleration support for Rockchip RK3288
> + *
> + * Copyright (c) 2015, Fuzhou Rockchip Electronics Co., Ltd
> + *
> + * Author: Zain Wang <zain.w...@rock-chips.com>
> + *
> + * This program is free software; you can redistribute it and/or modify it
> + * under the terms and conditions of the GNU General Public License,
> + * version 2, as published by the Free Software Foundation.
> + *
> + * Some ideas are from marvell/cesa.c and s5p-sss.c driver.
> + */
> +
> +#include "rk3288_crypto.h"
> +#include <linux/module.h>
> +#include <linux/platform_device.h>
> +#include <linux/of.h>
> +#include <linux/clk.h>
> +#include <linux/crypto.h>
> +
> +struct crypto_info_t *crypto_p;
> +
> +static int rk_crypto_enable_clk(struct crypto_info_t *dev)
> +{
> +     if (clk_prepare_enable(dev->clk)) {
> +             dev_err(dev->dev, "[%s:%d], Couldn't enable clock 'clk'\n",
> +                                             __func__, __LINE__);
> +             return -ENOENT;
> +     }
> +     if (clk_prepare_enable(dev->aclk)) {
> +             dev_err(dev->dev, "[%s:%d], Couldn't enable clock 'aclk'\n",
> +                                             __func__, __LINE__);
> +             goto err_aclk;
> +     }
> +     if (clk_prepare_enable(dev->hclk)) {
> +             dev_err(dev->dev, "[%s:%d], Couldn't enable clock 'hclk'\n",
> +                                             __func__, __LINE__);
> +             goto err_hclk;
> +     }
> +     if (clk_prepare_enable(dev->pclk)) {
> +             dev_err(dev->dev, "[%s:%d], Couldn't enable clock 'pclk'\n",
> +                                             __func__, __LINE__);
> +             goto err_pclk;
> +     }
> +     return 0;
> +
> +err_pclk:
> +     clk_disable_unprepare(dev->hclk);
> +err_hclk:
> +     clk_disable_unprepare(dev->aclk);
> +err_aclk:
> +     clk_disable_unprepare(dev->clk);
> +
> +     return -ENOENT;
> +}
Please return the error value given by clk_prepare_enable()

> +
> +static void rk_crypto_disable_clk(struct crypto_info_t *dev)
> +{
> +     clk_disable_unprepare(dev->hclk);
> +     clk_disable_unprepare(dev->aclk);
> +     clk_disable_unprepare(dev->pclk);
> +     clk_disable_unprepare(dev->clk);
> +}
> +
> +static int check_alignment(struct scatterlist *sg_src,
> +                        struct scatterlist *sg_dst,
> +                        int align_mask)
> +{
> +     int in, out, align;
> +
> +     in = IS_ALIGNED((u32)sg_src->offset, 4) &&
> +          IS_ALIGNED(sg_src->length, align_mask);
> +     if (sg_dst == NULL)
> +             return in;
> +     out = IS_ALIGNED((u32)sg_dst->offset, 4) &&
> +           IS_ALIGNED(sg_dst->length, align_mask);
> +     align = in && out;
> +
> +     return (align && (sg_src->length == sg_dst->length));
> +}
> +
> +static int rk_load_data(struct crypto_info_t *dev,
> +                           struct scatterlist *sg_src,
> +                           struct scatterlist *sg_dst)
> +{
> +     uint32_t count;
u32 is prefered over uint32_t

> +     int ret;
> +
> +     dev->aligned = dev->aligned ?
> +             check_alignment(sg_src, sg_dst, dev->align_size) :
> +             dev->aligned;
> +     if (dev->aligned) {
> +             count = min(dev->left_bytes, sg_src->length);
> +             dev->left_bytes -= count;
> +
> +             ret = dma_map_sg(dev->dev, sg_src, 1, DMA_TO_DEVICE);
> +             if (!ret) {
> +                     dev_err(dev->dev, "[%s:%d] dma_map_sg(src)  error\n",
> +                                                     __func__, __LINE__);
> +                     return -EINVAL;
> +             }
> +             dev->addr_in = sg_dma_address(sg_src);
> +
> +             if (sg_dst != NULL) {
> +                     ret = dma_map_sg(dev->dev, sg_dst, 1, DMA_FROM_DEVICE);
> +                     if (!ret) {
> +                             dev_err(dev->dev,
> +                                     "[%s:%d] dma_map_sg(dst)  error\n",
> +                                     __func__, __LINE__);
> +                             dma_unmap_sg(dev->dev, sg_src, 1,
> +                                             DMA_TO_DEVICE);
> +                             return -EINVAL;
> +                     }
> +                     dev->addr_out = sg_dma_address(sg_dst);
> +             }
> +     } else {
> +             count = (dev->left_bytes > PAGE_SIZE) ?
> +                     PAGE_SIZE : dev->left_bytes;
> +
> +             ret = sg_pcopy_to_buffer(dev->first, dev->nents,
> +                                      dev->addr_vir, count,
> +                                      dev->total - dev->left_bytes);
> +             if (!ret) {
> +                     dev_err(dev->dev, "[%s:%d] pcopy err\n",
> +                                             __func__, __LINE__);
> +                     return -EINVAL;
> +             }
> +             dev->left_bytes -= count;
> +             sg_init_one(&dev->sg_tmp, dev->addr_vir, count);
> +             ret = dma_map_sg(dev->dev, &dev->sg_tmp, 1, DMA_TO_DEVICE);
> +             if (!ret) {
> +                     dev_err(dev->dev, "[%s:%d] dma_map_sg(sg_tmp)  error\n",
> +                                                     __func__, __LINE__);
> +                     return -ENOMEM;
> +             }
> +             dev->addr_in = sg_dma_address(&dev->sg_tmp);
> +
> +             if (sg_dst != NULL) {
> +                     ret = dma_map_sg(dev->dev, &dev->sg_tmp, 1,
> +                                      DMA_FROM_DEVICE);
> +                     if (!ret) {
> +                             dev_err(dev->dev,
> +                                     "[%s:%d] dma_map_sg(sg_tmp)  error\n",
> +                                     __func__, __LINE__);
> +                             dma_unmap_sg(dev->dev, &dev->sg_tmp, 1,
> +                                          DMA_TO_DEVICE);
> +                             return -ENOMEM;
> +                     }
> +                     dev->addr_out = sg_dma_address(&dev->sg_tmp);
> +             }
> +     }
> +     dev->count = count;
> +     return 0;
> +}
> +
> +static void rk_unload_data(struct crypto_info_t *dev)
> +{
> +     struct scatterlist *sg_in, *sg_out;
> +
> +     sg_in = dev->aligned ? dev->sg_src : &dev->sg_tmp;
> +     dma_unmap_sg(dev->dev, sg_in, 1, DMA_TO_DEVICE);
> +
> +     if (dev->sg_dst != NULL) {
> +             sg_out = dev->aligned ? dev->sg_dst : &dev->sg_tmp;
> +             dma_unmap_sg(dev->dev, sg_out, 1, DMA_FROM_DEVICE);
> +     }
> +}
> +
> +static irqreturn_t crypto_irq_handle(int irq, void *dev_id)
> +{
> +     struct crypto_info_t *dev  = platform_get_drvdata(dev_id);
> +     uint32_t interrupt_status;
> +     int err = 0;
> +
> +     spin_lock(&dev->lock);
> +
> +     if (irq == dev->irq) {
> +             interrupt_status = CRYPTO_READ(dev, RK_CRYPTO_INTSTS);
> +             CRYPTO_WRITE(dev, RK_CRYPTO_INTSTS, interrupt_status);
> +             if (interrupt_status & 0x0a) {
> +                     dev_warn(dev->dev, "DMA Error\n");
> +                     err = -EFAULT;
> +             } else if (interrupt_status & 0x05)
> +                     err = dev->update(dev);
> +
> +             if (err)
> +                     dev->complete(dev, err);
> +     }
> +     spin_unlock(&dev->lock);
> +     return IRQ_HANDLED;
> +}
> +
> +static void rk_crypto_tasklet_cb(unsigned long data)
> +{
> +     struct crypto_info_t *dev = (struct crypto_info_t *)data;
> +     struct crypto_async_request *async_req, *backlog;
> +     struct rk_ahash_reqctx *hash_reqctx;
> +     struct rk_cipher_reqctx *ablk_reqctx;
> +     int err = 0;
> +
> +     spin_lock(&dev->lock);
> +     backlog   = crypto_get_backlog(&dev->queue);
> +     async_req = crypto_dequeue_request(&dev->queue);
> +     spin_unlock(&dev->lock);
> +     if (!async_req) {
> +             dev_err(dev->dev, "async_req is NULL !!\n");
> +             return;
> +     }
> +     if (backlog) {
> +             backlog->complete(backlog, -EINPROGRESS);
> +             backlog = NULL;
> +     }
> +
> +     if (crypto_tfm_alg_type(async_req->tfm) == CRYPTO_ALG_TYPE_AHASH) {
> +             dev->ahash_req = ahash_request_cast(async_req);
> +             hash_reqctx = ahash_request_ctx(dev->ahash_req);
> +     } else {
> +             dev->ablk_req = ablkcipher_request_cast(async_req);
> +             ablk_reqctx   = ablkcipher_request_ctx(dev->ablk_req);
> +     }
> +     err = dev->start(dev);
> +     if (err)
> +             dev->complete(dev, err);
> +}
> +
> +static struct crypto_alg *rk_cipher_algs[] = {
> +     &rk_ecb_aes_alg,
> +     &rk_cbc_aes_alg,
> +     &rk_ecb_des_alg,
> +     &rk_cbc_des_alg,
> +     &rk_ecb_des3_ede_alg,
> +     &rk_cbc_des3_ede_alg,
> +};
> +
> +static int rk_crypto_register(void)
> +{
> +     int i, k;
> +     int err = 0;
> +
> +     for (i = 0; i < ARRAY_SIZE(rk_cipher_algs); i++) {
> +             if (crypto_register_alg(rk_cipher_algs[i]))
> +                     goto err_cipher_algs;
> +     }
> +     return err;
> +
> +err_cipher_algs:
> +     for (k = 0; k < i; k++)
> +             crypto_unregister_alg(rk_cipher_algs[k]);
> +     return err;
> +}
You never set err to anything different from 0, so even in case of error the 
function return 0.

> +
> +static void rk_crypto_unregister(void)
> +{
> +     int i;
> +
> +     for (i = 0; i < ARRAY_SIZE(rk_cipher_algs); i++)
> +             crypto_unregister_alg(rk_cipher_algs[i]);
> +}
You could set i to unsigned for cleaning a possible warning

> +
> +static int rk_crypto_probe(struct platform_device *pdev)
> +{
> +     int err = 0;
> +     struct resource *res;
> +     struct device *dev = &pdev->dev;
> +     struct crypto_info_t *crypto_info;
> +
> +     crypto_info = devm_kzalloc(&pdev->dev,
> +                     sizeof(*crypto_info), GFP_KERNEL);
> +     if (!crypto_info)
> +             return -ENOMEM;
> +
> +     spin_lock_init(&crypto_info->lock);
> +
> +     res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
> +     crypto_info->reg = devm_ioremap_resource(&pdev->dev, res);
> +     if (IS_ERR(crypto_info->reg)) {
> +             dev_warn(crypto_info->dev, "Error on remap reg\n");
Typo, ioremap ?
And you warn about an error ?

> +             err = PTR_ERR(crypto_info->reg);
> +             goto err_ioremap;
> +     }
> +
> +     crypto_info->aclk = clk_get(&pdev->dev, "aclk_crypto");
> +     if (IS_ERR(crypto_info->aclk)) {
> +             dev_err(dev, "failed to find crypto clock source\n");
> +             err = -ENOENT;
The err must be set with PTR_ERR()
And you could also use devm_clk_get()

> +             goto err_ioremap;
> +     }
> +
> +     crypto_info->hclk = clk_get(&pdev->dev, "hclk_crypto");
> +     if (IS_ERR(crypto_info->hclk)) {
> +             dev_err(dev, "failed to find crypto clock source\n");
> +             err = -ENOENT;
Same comment

> +             goto err_ioremap;
> +     }
> +
> +     crypto_info->clk = clk_get(&pdev->dev, "srst_crypto");
> +     if (IS_ERR(crypto_info->clk)) {
> +             dev_err(dev, "failed to find crypto clock source\n");
> +             err = -ENOENT;
Same comment

> +             goto err_ioremap;
> +     }
> +
> +     crypto_info->pclk = clk_get(&pdev->dev, "apb_pclk");
> +     if (IS_ERR(crypto_info->pclk)) {
> +             dev_err(dev, "failed to find crypto clock source\n");
> +             err = -ENOENT;
Same comment

> +             goto err_ioremap;
> +     }
> +
> +     crypto_info->irq = platform_get_irq(pdev, 0);
> +     if (crypto_info->irq < 0) {
> +             dev_warn(crypto_info->dev,
> +                             "control Interrupt is not available.\n");
> +             err = crypto_info->irq;
> +             goto err_ioremap;
> +     }
> +
> +     err = request_irq(crypto_info->irq, crypto_irq_handle, IRQF_SHARED,
> +                       "rk-crypto", pdev);
You could use devm_request_irq()

> +     if (err) {
> +             dev_warn(crypto_info->dev, "irq request failed.\n");
> +             goto err_ioremap;
> +     }
> +
> +     crypto_info->dev = &pdev->dev;
> +     platform_set_drvdata(pdev, crypto_info);
> +     crypto_p = crypto_info;
> +
> +     tasklet_init(&crypto_info->crypto_tasklet,
> +                     rk_crypto_tasklet_cb, (unsigned long)crypto_info);
> +     crypto_init_queue(&crypto_info->queue, 50);
> +
> +     crypto_info->enable_clk = rk_crypto_enable_clk;
> +     crypto_info->disable_clk = rk_crypto_disable_clk;
> +     crypto_info->load_data = rk_load_data;
> +     crypto_info->unload_data = rk_unload_data;
> +
> +     err = rk_crypto_register();
> +     if (err) {
> +             pr_info("err in register alg");
Use dev_err()

> +             goto err_reg_alg;
> +     }
> +
> +     return 0;
> +
> +err_reg_alg:
> +     free_irq(crypto_info->irq, crypto_info);
> +err_ioremap:
> +     kfree(crypto_info);
Wrong, you do not have to free something allocated with devm_kzalloc

> +     crypto_p = NULL;
> +
> +     return err;
> +}
> +
> +static int rk_crypto_remove(struct platform_device *pdev)
> +{
> +     struct crypto_info_t *crypto_tmp = platform_get_drvdata(pdev);
> +
> +     rk_crypto_unregister();
> +     tasklet_kill(&crypto_tmp->crypto_tasklet);
> +     free_irq(crypto_tmp->irq, crypto_tmp);
> +     kfree(crypto_tmp);
> +     crypto_p = NULL;
> +
> +     return 0;
> +}
> +#ifdef CONFIG_OF
> +static const struct of_device_id crypto_of_id_table[] = {
> +     { .compatible = "rockchip,crypto" },
> +     {}
> +};
> +#endif /* CONFIG_OF */
> +
> +static const struct platform_device_id crypto_id_table[] = {
> +     { "rockchip,crypto" },
> +     {}
> +};
> +
> +static struct platform_driver crypto_driver = {
> +     .probe          = rk_crypto_probe,
> +     .remove         = rk_crypto_remove,
> +     .driver         = {
> +             .name   = "rockchip,crypto",
> +             .of_match_table = of_match_ptr(crypto_of_id_table),
> +     },
> +     .id_table       = crypto_id_table,
> +};
> +
> +module_platform_driver(crypto_driver);
> +
> +MODULE_LICENSE("GPL");
> +MODULE_AUTHOR("Zain Wang");
> diff --git a/drivers/crypto/rk_crypto/rk3288_crypto.h 
> b/drivers/crypto/rk_crypto/rk3288_crypto.h
> new file mode 100644
> index 0000000..27022bc
> --- /dev/null
> +++ b/drivers/crypto/rk_crypto/rk3288_crypto.h
> @@ -0,0 +1,291 @@
> +#ifndef __RK3288_CRYPTO_H__
> +#define __RK3288_CRYPTO_H__
> +
> +#include <crypto/sha.h>
> +#include <crypto/internal/hash.h>
> +#include <crypto/aes.h>
> +#include <crypto/des.h>
> +#include <crypto/ctr.h>
> +#include <crypto/algapi.h>
> +#include <linux/interrupt.h>
> +#include <linux/delay.h>
> +
> +#define _SBF(s, v)                   ((v) << (s))
> +#define _BIT(b)                              _SBF(b, 1)
> +
> +#define FLAGS_HASH_SHA1                      _SBF(2, 0x00)
> +#define FLAGS_HASH_MD5                       _SBF(2, 0x01)
> +#define FLAGS_HASH_SHA256            _SBF(2, 0x02)
> +#define FLAGS_HASH_PRNG                      _SBF(2, 0x03)
> +
> +/* Crypto control registers*/
> +#define RK_CRYPTO_INTSTS             0x0000
> +#define RK_CRYPTO_PKA_DONE_INT               _BIT(5)
> +#define RK_CRYPTO_HASH_DONE_INT              _BIT(4)
> +#define RK_CRYPTO_HRDMA_ERR_INT              _BIT(3)
> +#define RK_CRYPTO_HRDMA_DONE_INT     _BIT(2)
> +#define RK_CRYPTO_BCDMA_ERR_INT              _BIT(1)
> +#define RK_CRYPTO_BCDMA_DONE_INT     _BIT(0)
> +
> +#define RK_CRYPTO_INTENA             0x0004
> +#define RK_CRYPTO_PKA_DONE_ENA               _BIT(5)
> +#define RK_CRYPTO_HASH_DONE_ENA              _BIT(4)
> +#define RK_CRYPTO_HRDMA_ERR_ENA              _BIT(3)
> +#define RK_CRYPTO_HRDMA_DONE_ENA     _BIT(2)
> +#define RK_CRYPTO_BCDMA_ERR_ENA              _BIT(1)
> +#define RK_CRYPTO_BCDMA_DONE_ENA     _BIT(0)
> +
> +#define RK_CRYPTO_CTRL                       0x0008
> +#define RK_CRYPTO_WRITE_MASK         (0xFFFF<<16)
> +#define RK_CRYPTO_TRNG_FLUSH         _BIT(9)
> +#define RK_CRYPTO_TRNG_START         _BIT(8)
> +#define RK_CRYPTO_PKA_FLUSH          _BIT(7)
> +#define RK_CRYPTO_HASH_FLUSH         _BIT(6)
> +#define RK_CRYPTO_BLOCK_FLUSH                _BIT(5)
> +#define RK_CRYPTO_PKA_START          _BIT(4)
> +#define RK_CRYPTO_HASH_START         _BIT(3)
> +#define RK_CRYPTO_BLOCK_START                _BIT(2)
> +#define RK_CRYPTO_TDES_START         _BIT(1)
> +#define RK_CRYPTO_AES_START          _BIT(0)
> +
> +#define RK_CRYPTO_CONF                       0x000c
> +/* HASH Receive DMA Address Mode:   fix | increment */
> +#define RK_CRYPTO_HR_ADDR_MODE               _BIT(8)
> +/* Block Transmit DMA Address Mode: fix | increment */
> +#define RK_CRYPTO_BT_ADDR_MODE               _BIT(7)
> +/* Block Receive DMA Address Mode:  fix | increment */
> +#define RK_CRYPTO_BR_ADDR_MODE               _BIT(6)
> +#define RK_CRYPTO_BYTESWAP_HRFIFO    _BIT(5)
> +#define RK_CRYPTO_BYTESWAP_BTFIFO    _BIT(4)
> +#define RK_CRYPTO_BYTESWAP_BRFIFO    _BIT(3)
> +/* AES = 0 OR DES = 1 */
> +#define RK_CRYPTO_DESSEL                             _BIT(2)
> +#define RK_CYYPTO_HASHINSEL_INDEPENDENT_SOURCE               _SBF(0, 0x00)
> +#define RK_CYYPTO_HASHINSEL_BLOCK_CIPHER_INPUT               _SBF(0, 0x01)
> +#define RK_CYYPTO_HASHINSEL_BLOCK_CIPHER_OUTPUT              _SBF(0, 0x02)
> +
> +/* Block Receiving DMA Start Address Register */
> +#define RK_CRYPTO_BRDMAS             0x0010
> +/* Block Transmitting DMA Start Address Register */
> +#define RK_CRYPTO_BTDMAS             0x0014
> +/* Block Receiving DMA Length Register */
> +#define RK_CRYPTO_BRDMAL             0x0018
> +/* Hash Receiving DMA Start Address Register */
> +#define RK_CRYPTO_HRDMAS             0x001c
> +/* Hash Receiving DMA Length Register */
> +#define RK_CRYPTO_HRDMAL             0x0020
> +
> +/* AES registers */
> +#define RK_CRYPTO_AES_CTRL                     0x0080
> +#define RK_CRYPTO_AES_BYTESWAP_CNT   _BIT(11)
> +#define RK_CRYPTO_AES_BYTESWAP_KEY   _BIT(10)
> +#define RK_CRYPTO_AES_BYTESWAP_IV    _BIT(9)
> +#define RK_CRYPTO_AES_BYTESWAP_DO    _BIT(8)
> +#define RK_CRYPTO_AES_BYTESWAP_DI    _BIT(7)
> +#define RK_CRYPTO_AES_KEY_CHANGE     _BIT(6)
> +#define RK_CRYPTO_AES_ECB_MODE               _SBF(4, 0x00)
> +#define RK_CRYPTO_AES_CBC_MODE               _SBF(4, 0x01)
> +#define RK_CRYPTO_AES_CTR_MODE               _SBF(4, 0x02)
> +#define RK_CRYPTO_AES_128_bit_key    _SBF(2, 0x00)
> +#define RK_CRYPTO_AES_192_bit_key    _SBF(2, 0x01)
> +#define RK_CRYPTO_AES_256_bit_key    _SBF(2, 0x02)
> +/* Slave = 0 / fifo = 1 */
> +#define RK_CRYPTO_AES_FIFO_MODE              _BIT(1)
> +/* Encryption = 0 , Decryption = 1 */
> +#define RK_CRYPTO_AES_DEC            _BIT(0)
> +
> +#define RK_CRYPTO_AES_STS            0x0084
> +#define RK_CRYPTO_AES_DONE           _BIT(0)
> +
> +/* AES Input Data 0-3 Register */
> +#define RK_CRYPTO_AES_DIN_0          0x0088
> +#define RK_CRYPTO_AES_DIN_1          0x008c
> +#define RK_CRYPTO_AES_DIN_2          0x0090
> +#define RK_CRYPTO_AES_DIN_3          0x0094
> +
> +/* AES output Data 0-3 Register */
> +#define RK_CRYPTO_AES_DOUT_0         0x0098
> +#define RK_CRYPTO_AES_DOUT_1         0x009c
> +#define RK_CRYPTO_AES_DOUT_2         0x00a0
> +#define RK_CRYPTO_AES_DOUT_3         0x00a4
> +
> +/* AES IV Data 0-3 Register */
> +#define RK_CRYPTO_AES_IV_0           0x00a8
> +#define RK_CRYPTO_AES_IV_1           0x00ac
> +#define RK_CRYPTO_AES_IV_2           0x00b0
> +#define RK_CRYPTO_AES_IV_3           0x00b4
> +
> +/* AES Key Data 0-3 Register */
> +#define RK_CRYPTO_AES_KEY_0          0x00b8
> +#define RK_CRYPTO_AES_KEY_1          0x00bc
> +#define RK_CRYPTO_AES_KEY_2          0x00c0
> +#define RK_CRYPTO_AES_KEY_3          0x00c4
> +#define RK_CRYPTO_AES_KEY_4          0x00c8
> +#define RK_CRYPTO_AES_KEY_5          0x00cc
> +#define RK_CRYPTO_AES_KEY_6          0x00d0
> +#define RK_CRYPTO_AES_KEY_7          0x00d4
> +
> +/* AES Input Counter 0-3 Register */
> +#define RK_CRYPTO_AES_CNT_0          0x00d8
> +#define RK_CRYPTO_AES_CNT_1          0x00dc
> +#define RK_CRYPTO_AES_CNT_2          0x00e0
> +#define RK_CRYPTO_AES_CNT_3          0x00e4
> +
> +/* des/tdes */
> +#define RK_CRYPTO_TDES_CTRL          0x0100
> +#define RK_CRYPTO_TDES_BYTESWAP_KEY  _BIT(8)
> +#define RK_CRYPTO_TDES_BYTESWAP_IV   _BIT(7)
> +#define RK_CRYPTO_TDES_BYTESWAP_DO   _BIT(6)
> +#define RK_CRYPTO_TDES_BYTESWAP_DI   _BIT(5)
> +/* 0: ECB, 1: CBC */
> +#define RK_CRYPTO_TDES_CHAINMODE     _BIT(4)
> +/* TDES Key Mode, 0 : EDE, 1 : EEE */
> +#define RK_CRYPTO_TDES_EEE           _BIT(3)
> +/* 0: DES, 1:TDES */
> +#define RK_CRYPTO_TDES_SELECT                _BIT(2)
> +/* 0: Slave, 1:Fifo */
> +#define RK_CRYPTO_TDES_FIFO_MODE     _BIT(1)
> +/* Encryption = 0 , Decryption = 1 */
> +#define RK_CRYPTO_TDES_DEC           _BIT(0)
> +
> +#define RK_CRYPTO_TDES_STS           0x0104
> +#define RK_CRYPTO_TDES_DONE          _BIT(0)
> +
> +#define RK_CRYPTO_TDES_DIN_0         0x0108
> +#define RK_CRYPTO_TDES_DIN_1         0x010c
> +#define RK_CRYPTO_TDES_DOUT_0                0x0110
> +#define RK_CRYPTO_TDES_DOUT_1                0x0114
> +#define RK_CRYPTO_TDES_IV_0          0x0118
> +#define RK_CRYPTO_TDES_IV_1          0x011c
> +#define RK_CRYPTO_TDES_KEY1_0                0x0120
> +#define RK_CRYPTO_TDES_KEY1_1                0x0124
> +#define RK_CRYPTO_TDES_KEY2_0                0x0128
> +#define RK_CRYPTO_TDES_KEY2_1                0x012c
> +#define RK_CRYPTO_TDES_KEY3_0                0x0130
> +#define RK_CRYPTO_TDES_KEY3_1                0x0134
> +
> +/* HASH */
> +#define RK_CRYPTO_HASH_CTRL          0x0180
> +#define RK_CRYPTO_HASH_SWAP_DO               _BIT(3)
> +#define RK_CRYPTO_HASH_SWAP_DI               _BIT(2)
> +#define RK_CRYPTO_HASH_SHA1          _SBF(0, 0x00)
> +#define RK_CRYPTO_HASH_MD5           _SBF(0, 0x01)
> +#define RK_CRYPTO_HASH_SHA256                _SBF(0, 0x02)
> +#define RK_CRYPTO_HASH_PRNG          _SBF(0, 0x03)
> +
> +#define RK_CRYPTO_HASH_STS           0x0184
> +#define RK_CRYPTO_HASH_DONE          _BIT(0)
> +
> +#define RK_CRYPTO_HASH_MSG_LEN               0x0188
> +#define RK_CRYPTO_HASH_DOUT_0                0x018c
> +#define RK_CRYPTO_HASH_DOUT_1                0x0190
> +#define RK_CRYPTO_HASH_DOUT_2                0x0194
> +#define RK_CRYPTO_HASH_DOUT_3                0x0198
> +#define RK_CRYPTO_HASH_DOUT_4                0x019c
> +#define RK_CRYPTO_HASH_DOUT_5                0x01a0
> +#define RK_CRYPTO_HASH_DOUT_6                0x01a4
> +#define RK_CRYPTO_HASH_DOUT_7                0x01a8
> +#define RK_CRYPTO_HASH_SEED_0                0x01ac
> +#define RK_CRYPTO_HASH_SEED_1                0x01b0
> +#define RK_CRYPTO_HASH_SEED_2                0x01b4
> +#define RK_CRYPTO_HASH_SEED_3                0x01b8
> +#define RK_CRYPTO_HASH_SEED_4                0x01bc
> +
> +/* TRNG */
> +#define RK_CRYPTO_TRNG_CTRL          0x0200
> +#define RK_CRYPTO_OSC_ENABLE         _BIT(16)
> +
> +#define RK_CRYPTO_TRNG_DOUT_0                0x0204
> +#define RK_CRYPTO_TRNG_DOUT_1                0x0208
> +#define RK_CRYPTO_TRNG_DOUT_2                0x020c
> +#define RK_CRYPTO_TRNG_DOUT_3                0x0210
> +#define RK_CRYPTO_TRNG_DOUT_4                0x0214
> +#define RK_CRYPTO_TRNG_DOUT_5                0x0218
> +#define RK_CRYPTO_TRNG_DOUT_6                0x021c
> +#define RK_CRYPTO_TRNG_DOUT_7                0x0220
> +
> +/* PAK OR RSA */
> +#define RK_CRYPTO_PKA_CTRL                   0x0280
> +#define RK_CRYPTO_PKA_BLOCK_SIZE_512BIT              _SBF(0, 0x00)
> +#define RK_CRYPTO_PKA_BLOCK_SIZE_1024BIT     _SBF(0, 0x01)
> +#define RK_CRYPTO_PKA_BLOCK_SIZE_2048BIT     _SBF(0, 0x02)
> +
> +/* result = (M ^ E) mod N */
> +#define RK_CRYPTO_PKA_M                      0x0400
> +/* C = 2 ^ (2n+2) mod N */
> +#define RK_CRYPTO_PKA_C                      0x0500
> +#define RK_CRYPTO_PKA_N                      0x0600
> +#define RK_CRYPTO_PKA_E                      0x0700
> +
> +#define CRYPTO_READ(dev, offset)               \
> +             __raw_readl(((dev)->reg + (offset)))
> +#define CRYPTO_WRITE(dev, offset, val)         \
> +             __raw_writel((val), ((dev)->reg + (offset)))
> +/* get register virt address */
> +#define CRYPTO_GET_REG_VIRT(dev, offset)   ((dev)->reg + (offset))
> +
> +#define MD5_DIGEST_SIZE                      16
#This is already defined in incluude/crypto/md5.h

> +#define RK_ALIGN_MASK                        (sizeof(u32)-1)
> +
> +struct crypto_info_t {
> +     struct device                   *dev;
> +     struct clk                      *aclk;
> +     struct clk                      *hclk;
> +     struct clk                      *clk;
> +     struct clk                      *pclk;
> +     void __iomem                    *reg;
> +     int                             irq;
> +     struct crypto_queue             queue;
> +     struct tasklet_struct           crypto_tasklet;
> +     struct ahash_request            *ahash_req;
> +     struct ablkcipher_request       *ablk_req;
> +     spinlock_t                      lock;
> +
> +     /* the public variable */
> +     struct scatterlist              *sg_src;
> +     struct scatterlist              *sg_dst;
> +     struct scatterlist              sg_tmp;
> +     struct scatterlist              *first;
> +     unsigned int                    left_bytes;
> +     char                            *addr_vir;
> +     int                             aligned;
> +     int                             align_size;
> +     size_t                          nents;
> +     unsigned int                    total;
> +     uint32_t                        count;
> +     uint32_t                        mode;
> +     dma_addr_t                      addr_in;
> +     dma_addr_t                      addr_out;
> +     int (*start)(struct crypto_info_t *dev);
> +     int (*update)(struct crypto_info_t *dev);
> +     void (*complete)(struct crypto_info_t *dev, int err);
> +     int (*enable_clk)(struct crypto_info_t *dev);
> +     void (*disable_clk)(struct crypto_info_t *dev);
> +     int (*load_data)(struct crypto_info_t *dev,
> +                       struct scatterlist *sg_src,
> +                       struct scatterlist *sg_dst);
> +     void (*unload_data)(struct crypto_info_t *dev);
> +};
> +
> +/* the private variable of hash */
> +struct rk_ahash_ctx {
> +     struct crypto_info_t            *dev;
> +     int                             FLAG_FINUP;
> +     int                             first_op;
> +};
> +
> +/* the private variable of cipher */
> +struct rk_cipher_ctx {
> +     struct crypto_info_t            *dev;
> +     int                             keylen;
> +};
> +extern struct crypto_info_t *crypto_p;
> +
> +extern struct crypto_alg rk_ecb_aes_alg;
> +extern struct crypto_alg rk_cbc_aes_alg;
> +extern struct crypto_alg rk_ecb_des_alg;
> +extern struct crypto_alg rk_cbc_des_alg;
> +extern struct crypto_alg rk_ecb_des3_ede_alg;
> +extern struct crypto_alg rk_cbc_des3_ede_alg;
> +
> +#endif
> diff --git a/drivers/crypto/rk_crypto/rk3288_crypto_ablkcipher.c 
> b/drivers/crypto/rk_crypto/rk3288_crypto_ablkcipher.c
> new file mode 100644
> index 0000000..3d08875
> --- /dev/null
> +++ b/drivers/crypto/rk_crypto/rk3288_crypto_ablkcipher.c
> @@ -0,0 +1,502 @@
> +/*
> + *Crypto acceleration support for Rockchip RK3288
> + *
> + * Copyright (c) 2015, Fuzhou Rockchip Electronics Co., Ltd
> + *
> + * Author: Zain Wang <zain.w...@rock-chips.com>
> + *
> + * This program is free software; you can redistribute it and/or modify it
> + * under the terms and conditions of the GNU General Public License,
> + * version 2, as published by the Free Software Foundation.
> + *
> + * Some ideas are from marvell/cesa.c and s5p-sss.c driver.
> + */
> +#include "rk3288_crypto.h"
> +
> +#define RK_CRYPTO_DEC                        _BIT(0)
> +#define AES  0
> +#define TDES _BIT(16)
> +
> +static void rk_crypto_complete(struct crypto_info_t *dev, int err)
> +{
> +     if (dev->ablk_req->base.complete != NULL) {
> +             if (err)
> +                     dev_warn(dev->dev, "[%s:%d] err = %d\n",
> +                                             __func__, __LINE__, err);
> +             dev->ablk_req->base.complete(&dev->ablk_req->base, err);
> +     }
> +}
> +
> +static int rk_handle_req(struct ablkcipher_request *req, int alig_bytes)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +     int err;
> +
> +     if (!IS_ALIGNED(req->nbytes, alig_bytes))
> +             return -EINVAL;
> +
> +     spin_lock(&dev->lock);
> +     err = ablkcipher_enqueue_request(&dev->queue, req);
> +     spin_unlock(&dev->lock);
> +     tasklet_schedule(&dev->crypto_tasklet);
> +     return err;
> +}
> +
> +static void rk_ablk_init(struct crypto_info_t *dev,
> +                             struct ablkcipher_request *req)
> +{
> +     dev->left_bytes = req->nbytes;
> +     dev->total = req->nbytes;
> +     dev->sg_src = req->src;
> +     dev->first = req->src;
> +     dev->nents = sg_nents(req->src);
> +     dev->sg_dst = req->dst;
> +     dev->aligned = 1;
> +     dev->ablk_req = req;
> +}
> +
> +static int rk_aes_setkey(struct crypto_ablkcipher *cipher,
> +                     const uint8_t *key, unsigned int keylen)
> +{
> +     struct crypto_tfm *tfm = crypto_ablkcipher_tfm(cipher);
> +     struct rk_cipher_ctx *ctx = crypto_tfm_ctx(tfm);
> +
> +     if (key == NULL) {
> +             pr_err("[%s:%d] no key error\n", __func__, __LINE__);
> +             return -EINVAL;
> +     }
> +
> +     if (keylen != AES_KEYSIZE_128 && keylen != AES_KEYSIZE_192 &&
> +         keylen != AES_KEYSIZE_256) {
> +             crypto_ablkcipher_set_flags(cipher, CRYPTO_TFM_RES_BAD_KEY_LEN);
> +             pr_err("[%s:%d] expect key len = %d\n",
> +                                             __func__, __LINE__, keylen);
> +             return -EINVAL;
> +     }
> +     ctx->keylen = keylen;
> +     memcpy(ctx->dev->reg + RK_CRYPTO_AES_KEY_0, key, keylen);
> +     return 0;
> +}
> +
> +static int rk_tdes_setkey(struct crypto_ablkcipher *cipher,
> +                     const uint8_t *key, unsigned int keylen)
> +{
> +     struct crypto_tfm *tfm = crypto_ablkcipher_tfm(cipher);
> +     struct rk_cipher_ctx *ctx = crypto_tfm_ctx(tfm);
> +
> +     if (key == NULL) {
> +             pr_err("[%s:%d] no key error\n", __func__, __LINE__);
> +             return -EINVAL;
> +     }
> +
> +     if (keylen != DES_KEY_SIZE && keylen != DES3_EDE_KEY_SIZE) {
> +             crypto_ablkcipher_set_flags(cipher, CRYPTO_TFM_RES_BAD_KEY_LEN);
> +             pr_err("[%s:%d] expect key len = %d\n",
> +                                             __func__, __LINE__, keylen);
> +             return -EINVAL;
> +     }
> +     ctx->keylen = keylen;
> +     memcpy(ctx->dev->reg + RK_CRYPTO_TDES_KEY1_0, key, keylen);
> +     return 0;
> +}
> +
> +static int rk_aes_ecb_encrypt(struct ablkcipher_request *req)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +
> +     dev->mode = RK_CRYPTO_AES_ECB_MODE | AES;
> +     rk_ablk_init(dev, req);
> +     return rk_handle_req(req, dev->align_size);
> +}
> +
> +static int rk_aes_ecb_decrypt(struct ablkcipher_request *req)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +
> +     dev->mode = RK_CRYPTO_AES_ECB_MODE | RK_CRYPTO_DEC | AES;
> +     rk_ablk_init(dev, req);
> +     return rk_handle_req(req, dev->align_size);
> +}
> +
> +static int rk_aes_cbc_encrypt(struct ablkcipher_request *req)
> +{
> +
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +
> +     dev->mode = RK_CRYPTO_AES_CBC_MODE | AES;
> +     rk_ablk_init(dev, req);
> +     return rk_handle_req(req, dev->align_size);
> +}
> +
> +static int rk_aes_cbc_decrypt(struct ablkcipher_request *req)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +
> +     dev->mode = RK_CRYPTO_AES_CBC_MODE | RK_CRYPTO_DEC | AES;
> +     rk_ablk_init(dev, req);
> +     return rk_handle_req(req, dev->align_size);
> +}
> +
> +static int rk_des_ecb_encrypt(struct ablkcipher_request *req)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +
> +     dev->mode = TDES;
> +     rk_ablk_init(dev, req);
> +     return rk_handle_req(req, dev->align_size);
> +}
> +
> +static int rk_des_ecb_decrypt(struct ablkcipher_request *req)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +
> +     dev->mode = RK_CRYPTO_DEC | TDES;
> +     rk_ablk_init(dev, req);
> +     return rk_handle_req(req, dev->align_size);
> +}
> +
> +static int rk_des_cbc_encrypt(struct ablkcipher_request *req)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +
> +     dev->mode = RK_CRYPTO_TDES_CHAINMODE | TDES;
> +     rk_ablk_init(dev, req);
> +     return rk_handle_req(req, dev->align_size);
> +}
> +
> +static int rk_des_cbc_decrypt(struct ablkcipher_request *req)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +
> +     dev->mode = RK_CRYPTO_TDES_CHAINMODE | RK_CRYPTO_DEC | TDES;
> +     rk_ablk_init(dev, req);
> +     return rk_handle_req(req, dev->align_size);
> +}
> +
> +static int rk_des3_ede_ecb_encrypt(struct ablkcipher_request *req)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +
> +     dev->mode = RK_CRYPTO_TDES_SELECT | TDES;
> +     rk_ablk_init(dev, req);
> +     return rk_handle_req(req, dev->align_size);
> +}
> +
> +static int rk_des3_ede_ecb_decrypt(struct ablkcipher_request *req)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +
> +     dev->mode = RK_CRYPTO_TDES_SELECT | RK_CRYPTO_DEC | TDES;
> +     rk_ablk_init(dev, req);
> +     return rk_handle_req(req, dev->align_size);
> +}
> +
> +static int rk_des3_ede_cbc_encrypt(struct ablkcipher_request *req)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +
> +     dev->mode = RK_CRYPTO_TDES_SELECT | RK_CRYPTO_TDES_CHAINMODE | TDES;
> +     rk_ablk_init(dev, req);
> +     return rk_handle_req(req, dev->align_size);
> +}
> +
> +static int rk_des3_ede_cbc_decrypt(struct ablkcipher_request *req)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     struct crypto_info_t *dev = ctx->dev;
> +
> +     dev->mode = RK_CRYPTO_TDES_SELECT | RK_CRYPTO_TDES_CHAINMODE |
> +                 RK_CRYPTO_DEC | TDES;
> +     rk_ablk_init(dev, req);
> +     return rk_handle_req(req, dev->align_size);
> +}
> +
> +static void rk_ablk_hw_init(struct crypto_info_t *dev)
> +{
> +     struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(dev->ablk_req);
> +     struct rk_cipher_ctx *ctx = crypto_ablkcipher_ctx(tfm);
> +     uint32_t conf_reg = 0;
> +
> +     if (dev->mode & TDES) {
> +             dev->mode &= ~TDES;
> +             dev->mode |= RK_CRYPTO_TDES_FIFO_MODE |
> +                          RK_CRYPTO_TDES_BYTESWAP_KEY |
> +                          RK_CRYPTO_TDES_BYTESWAP_IV;
> +             CRYPTO_WRITE(dev, RK_CRYPTO_TDES_CTRL, dev->mode);
> +
> +             memcpy(dev->reg + RK_CRYPTO_TDES_IV_0, dev->ablk_req->info, 8);
> +             conf_reg = RK_CRYPTO_DESSEL;
> +     } else {
> +             dev->mode |= RK_CRYPTO_AES_FIFO_MODE |
> +                          RK_CRYPTO_AES_KEY_CHANGE |
> +                          RK_CRYPTO_AES_BYTESWAP_KEY |
> +                          RK_CRYPTO_AES_BYTESWAP_IV;
> +
> +             if (ctx->keylen == AES_KEYSIZE_192)
> +                     dev->mode |= RK_CRYPTO_AES_192_bit_key;
> +             else if (ctx->keylen == AES_KEYSIZE_256)
> +                     dev->mode |= RK_CRYPTO_AES_256_bit_key;
> +
> +             CRYPTO_WRITE(dev, RK_CRYPTO_AES_CTRL, dev->mode);
> +
> +             memcpy(dev->reg + RK_CRYPTO_AES_IV_0, dev->ablk_req->info, 16);
> +     }
> +     conf_reg |= RK_CRYPTO_BYTESWAP_BTFIFO |
> +                 RK_CRYPTO_BYTESWAP_BRFIFO;
> +     CRYPTO_WRITE(dev, RK_CRYPTO_CONF, conf_reg);
> +     CRYPTO_WRITE(dev, RK_CRYPTO_INTENA,
> +                  RK_CRYPTO_BCDMA_ERR_ENA | RK_CRYPTO_BCDMA_DONE_ENA);
> +}
> +
> +static void crypto_dma_start(struct crypto_info_t *dev)
> +{
> +     CRYPTO_WRITE(dev, RK_CRYPTO_BRDMAS, dev->addr_in);
> +     CRYPTO_WRITE(dev, RK_CRYPTO_BRDMAL, dev->count / 4);
> +     CRYPTO_WRITE(dev, RK_CRYPTO_BTDMAS, dev->addr_out);
> +     CRYPTO_WRITE(dev, RK_CRYPTO_CTRL, RK_CRYPTO_BLOCK_START |
> +                     (RK_CRYPTO_BLOCK_START << 16));
> +}
> +
> +static int rk_set_data_start(struct crypto_info_t *dev)
> +{
> +     int err;
> +
> +     err = dev->load_data(dev, dev->sg_src, dev->sg_dst);
> +     if (!err)
> +             crypto_dma_start(dev);
> +     return err;
> +}
> +
> +static int rk_ablk_start(struct crypto_info_t *dev)
> +{
> +     int err;
> +
> +     spin_lock(&dev->lock);
> +     rk_ablk_hw_init(dev);
> +     err = rk_set_data_start(dev);
> +     spin_unlock(&dev->lock);
> +     return err;
> +}
> +/* return:
> + *   true    some err was occurred
> + *   fault   no err, please continue
> + */
> +static int rk_ablk_rx(struct crypto_info_t *dev)
> +{
> +     int err = 0;
> +
> +     dev->unload_data(dev);
> +     if (!dev->aligned) {
> +             if (!sg_pcopy_from_buffer(dev->ablk_req->dst, dev->nents,
> +                                       dev->addr_vir, dev->count,
> +                                       dev->total - dev->left_bytes -
> +                                       dev->count)) {
> +                     err = -ENOMEM;
I am not sure -ENOMEM is the good error to return

> +                     goto out_rx;
> +             }
> +     }
> +     if (dev->left_bytes) {
> +             if (dev->aligned) {
> +                     if (sg_is_last(dev->sg_src)) {
> +                             dev_warn(dev->dev, "[%s:%d], lack of data\n",
> +                                                     __func__, __LINE__);
> +                             err = -ENOMEM;
> +                             goto out_rx;
> +                     }
> +                     dev->sg_src = sg_next(dev->sg_src);
> +                     dev->sg_dst = sg_next(dev->sg_dst);
> +             }
> +             err = rk_set_data_start(dev);
> +     } else {
> +             /* here show the calculation is over without any err */
> +             dev->complete(dev, 0);
> +     }
> +out_rx:
> +     return err;
> +}
> +
> +static int rk_ablk_cra_init(struct crypto_tfm *tfm)
> +{
> +     struct rk_cipher_ctx *ctx = crypto_tfm_ctx(tfm);
> +
> +     ctx->dev = crypto_p;
> +     ctx->dev->align_size = crypto_tfm_alg_alignmask(tfm) + 1;
> +     ctx->dev->start = rk_ablk_start;
> +     ctx->dev->update = rk_ablk_rx;
> +     ctx->dev->complete = rk_crypto_complete;
> +     ctx->dev->addr_vir = (void *)__get_free_page(GFP_KERNEL);
> +     if (!ctx->dev->addr_vir) {
> +             pr_err("failed to kmalloc for addr_vir\n");
Since you have acces to the dev ptr, you could use all dev_xxx functions.
This comment, apply to whole file.
Moreover, checkpatch said that you do not have to warn about a failed memory 
allocation.

> +             return -ENOMEM;
> +     }
> +     return ctx->dev->enable_clk(ctx->dev);
> +}
> +
> +static void rk_ablk_cra_exit(struct crypto_tfm *tfm)
> +{
> +     struct rk_cipher_ctx *ctx = crypto_tfm_ctx(tfm);
> +
> +     free_page((unsigned long)ctx->dev->addr_vir);
> +     ctx->dev->disable_clk(ctx->dev);
> +}
> +
> +struct crypto_alg rk_ecb_aes_alg = {
> +     .cra_name               = "ecb(aes)",
> +     .cra_driver_name        = "ecb-aes-rk",
> +     .cra_priority           = 100,
> +     .cra_flags              = CRYPTO_ALG_TYPE_ABLKCIPHER |
> +                               CRYPTO_ALG_ASYNC,
> +     .cra_blocksize          = AES_BLOCK_SIZE,
> +     .cra_ctxsize            = sizeof(struct rk_cipher_ctx),
> +     .cra_alignmask          = 0x0f,
> +     .cra_type               = &crypto_ablkcipher_type,
> +     .cra_module             = THIS_MODULE,
> +     .cra_init               = rk_ablk_cra_init,
> +     .cra_exit               = rk_ablk_cra_exit,
> +     .cra_u.ablkcipher       = {
> +             .min_keysize    = AES_MIN_KEY_SIZE,
> +             .max_keysize    = AES_MAX_KEY_SIZE,
> +             .setkey         = rk_aes_setkey,
> +             .encrypt        = rk_aes_ecb_encrypt,
> +             .decrypt        = rk_aes_ecb_decrypt,
> +     }
> +};
> +
> +struct crypto_alg rk_cbc_aes_alg = {
> +     .cra_name               = "cbc(aes)",
> +     .cra_driver_name        = "cbc-aes-rk",
> +     .cra_priority           = 100,
> +     .cra_flags              = CRYPTO_ALG_TYPE_ABLKCIPHER |
> +                               CRYPTO_ALG_ASYNC,
> +     .cra_blocksize          = AES_BLOCK_SIZE,
> +     .cra_ctxsize            = sizeof(struct rk_cipher_ctx),
> +     .cra_alignmask          = 0x0f,
> +     .cra_type               = &crypto_ablkcipher_type,
> +     .cra_module             = THIS_MODULE,
> +     .cra_init               = rk_ablk_cra_init,
> +     .cra_exit               = rk_ablk_cra_exit,
> +     .cra_u.ablkcipher       = {
> +             .min_keysize    = AES_MIN_KEY_SIZE,
> +             .max_keysize    = AES_MAX_KEY_SIZE,
> +             .ivsize         = AES_BLOCK_SIZE,
> +             .setkey         = rk_aes_setkey,
> +             .encrypt        = rk_aes_cbc_encrypt,
> +             .decrypt        = rk_aes_cbc_decrypt,
> +     }
> +};
> +
> +struct crypto_alg rk_ecb_des_alg = {
> +     .cra_name               = "ecb(des)",
> +     .cra_driver_name        = "ecb-des-rk",
> +     .cra_priority           = 100,
> +     .cra_flags              = CRYPTO_ALG_TYPE_ABLKCIPHER |
> +                               CRYPTO_ALG_ASYNC,
> +     .cra_blocksize          = DES_BLOCK_SIZE,
> +     .cra_ctxsize            = sizeof(struct rk_cipher_ctx),
> +     .cra_alignmask          = 0x07,
> +     .cra_type               = &crypto_ablkcipher_type,
> +     .cra_module             = THIS_MODULE,
> +     .cra_init               = rk_ablk_cra_init,
> +     .cra_exit               = rk_ablk_cra_exit,
> +     .cra_u.ablkcipher       = {
> +             .min_keysize    = DES_KEY_SIZE,
> +             .max_keysize    = DES_KEY_SIZE,
> +             .setkey         = rk_tdes_setkey,
> +             .encrypt        = rk_des_ecb_encrypt,
> +             .decrypt        = rk_des_ecb_decrypt,
> +     }
> +};
> +
> +struct crypto_alg rk_cbc_des_alg = {
> +     .cra_name               = "cbc(des)",
> +     .cra_driver_name        = "cbc-des-rk",
> +     .cra_priority           = 100,
> +     .cra_flags              = CRYPTO_ALG_TYPE_ABLKCIPHER |
> +                               CRYPTO_ALG_ASYNC,
> +     .cra_blocksize          = DES_BLOCK_SIZE,
> +     .cra_ctxsize            = sizeof(struct rk_cipher_ctx),
> +     .cra_alignmask          = 0x07,
> +     .cra_type               = &crypto_ablkcipher_type,
> +     .cra_module             = THIS_MODULE,
> +     .cra_init               = rk_ablk_cra_init,
> +     .cra_exit               = rk_ablk_cra_exit,
> +     .cra_u.ablkcipher       = {
> +             .min_keysize    = DES_KEY_SIZE,
> +             .max_keysize    = DES_KEY_SIZE,
> +             .ivsize         = DES_BLOCK_SIZE,
> +             .setkey         = rk_tdes_setkey,
> +             .encrypt        = rk_des_cbc_encrypt,
> +             .decrypt        = rk_des_cbc_decrypt,
> +     }
> +};
> +
> +struct crypto_alg rk_ecb_des3_ede_alg = {
> +     .cra_name               = "ecb(des3_ede)",
> +     .cra_driver_name        = "ecb-des3-ede-rk",
> +     .cra_priority           = 100,
> +     .cra_flags              = CRYPTO_ALG_TYPE_ABLKCIPHER |
> +                               CRYPTO_ALG_ASYNC,
> +     .cra_blocksize          = DES_BLOCK_SIZE,
> +     .cra_ctxsize            = sizeof(struct rk_cipher_ctx),
> +     .cra_alignmask          = 0x07,
> +     .cra_type               = &crypto_ablkcipher_type,
> +     .cra_module             = THIS_MODULE,
> +     .cra_init               = rk_ablk_cra_init,
> +     .cra_exit               = rk_ablk_cra_exit,
> +     .cra_u.ablkcipher       = {
> +             .min_keysize    = DES3_EDE_KEY_SIZE,
> +             .max_keysize    = DES3_EDE_KEY_SIZE,
> +             .ivsize         = DES_BLOCK_SIZE,
> +             .setkey         = rk_tdes_setkey,
> +             .encrypt        = rk_des3_ede_ecb_encrypt,
> +             .decrypt        = rk_des3_ede_ecb_decrypt,
> +     }
> +};
> +
> +struct crypto_alg rk_cbc_des3_ede_alg = {
> +     .cra_name               = "cbc(des3_ede)",
> +     .cra_driver_name        = "cbc-des3-ede-rk",
> +     .cra_priority           = 100,
> +     .cra_flags              = CRYPTO_ALG_TYPE_ABLKCIPHER |
> +                               CRYPTO_ALG_ASYNC,
> +     .cra_blocksize          = DES_BLOCK_SIZE,
> +     .cra_ctxsize            = sizeof(struct rk_cipher_ctx),
> +     .cra_alignmask          = 0x07,
> +     .cra_type               = &crypto_ablkcipher_type,
> +     .cra_module             = THIS_MODULE,
> +     .cra_init               = rk_ablk_cra_init,
> +     .cra_exit               = rk_ablk_cra_exit,
> +     .cra_u.ablkcipher       = {
> +             .min_keysize    = DES3_EDE_KEY_SIZE,
> +             .max_keysize    = DES3_EDE_KEY_SIZE,
> +             .ivsize         = DES_BLOCK_SIZE,
> +             .setkey         = rk_tdes_setkey,
> +             .encrypt        = rk_des3_ede_cbc_encrypt,
> +             .decrypt        = rk_des3_ede_cbc_decrypt,
> +     }
> +};
> -- 
> 1.9.1

Best regards

--
To unsubscribe from this list: send the line "unsubscribe devicetree" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to