From: Eric Biggers <ebigg...@google.com>

Convert the AVX implementation of CAST5 from the (deprecated) ablkcipher
and blkcipher interfaces over to the skcipher interface.  Note that this
includes replacing the use of ablk_helper with crypto_simd.

Signed-off-by: Eric Biggers <ebigg...@google.com>
---
 arch/x86/crypto/cast5_avx_glue.c | 351 ++++++++++++++-------------------------
 crypto/Kconfig                   |   7 +-
 2 files changed, 130 insertions(+), 228 deletions(-)

diff --git a/arch/x86/crypto/cast5_avx_glue.c b/arch/x86/crypto/cast5_avx_glue.c
index 575292a33bdf2..56f2a1b0ccf57 100644
--- a/arch/x86/crypto/cast5_avx_glue.c
+++ b/arch/x86/crypto/cast5_avx_glue.c
@@ -21,18 +21,14 @@
  *
  */
 
-#include <linux/module.h>
-#include <linux/hardirq.h>
-#include <linux/types.h>
-#include <linux/crypto.h>
-#include <linux/err.h>
-#include <crypto/ablk_helper.h>
+#include <asm/crypto/glue_helper.h>
 #include <crypto/algapi.h>
 #include <crypto/cast5.h>
-#include <crypto/cryptd.h>
-#include <crypto/ctr.h>
-#include <asm/fpu/api.h>
-#include <asm/crypto/glue_helper.h>
+#include <crypto/internal/simd.h>
+#include <linux/crypto.h>
+#include <linux/err.h>
+#include <linux/module.h>
+#include <linux/types.h>
 
 #define CAST5_PARALLEL_BLOCKS 16
 
@@ -45,10 +41,17 @@ asmlinkage void cast5_cbc_dec_16way(struct cast5_ctx *ctx, 
u8 *dst,
 asmlinkage void cast5_ctr_16way(struct cast5_ctx *ctx, u8 *dst, const u8 *src,
                                __be64 *iv);
 
-static inline bool cast5_fpu_begin(bool fpu_enabled, unsigned int nbytes)
+static int cast5_setkey_skcipher(struct crypto_skcipher *tfm, const u8 *key,
+                                unsigned int keylen)
 {
-       return glue_fpu_begin(CAST5_BLOCK_SIZE, CAST5_PARALLEL_BLOCKS,
-                             NULL, fpu_enabled, nbytes);
+       return cast5_setkey(&tfm->base, key, keylen);
+}
+
+static inline bool cast5_fpu_begin(bool fpu_enabled, struct skcipher_walk 
*walk,
+                                  unsigned int nbytes)
+{
+       return glue_skwalk_fpu_begin(CAST5_BLOCK_SIZE, CAST5_PARALLEL_BLOCKS,
+                                    walk, fpu_enabled, nbytes);
 }
 
 static inline void cast5_fpu_end(bool fpu_enabled)
@@ -56,24 +59,24 @@ static inline void cast5_fpu_end(bool fpu_enabled)
        return glue_fpu_end(fpu_enabled);
 }
 
-static int ecb_crypt(struct blkcipher_desc *desc, struct blkcipher_walk *walk,
-                    bool enc)
+static int ecb_crypt(struct skcipher_request *req, bool enc)
 {
        bool fpu_enabled = false;
-       struct cast5_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
+       struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
+       struct cast5_ctx *ctx = crypto_skcipher_ctx(tfm);
+       struct skcipher_walk walk;
        const unsigned int bsize = CAST5_BLOCK_SIZE;
        unsigned int nbytes;
        void (*fn)(struct cast5_ctx *ctx, u8 *dst, const u8 *src);
        int err;
 
-       err = blkcipher_walk_virt(desc, walk);
-       desc->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP;
+       err = skcipher_walk_virt(&walk, req, false);
 
-       while ((nbytes = walk->nbytes)) {
-               u8 *wsrc = walk->src.virt.addr;
-               u8 *wdst = walk->dst.virt.addr;
+       while ((nbytes = walk.nbytes)) {
+               u8 *wsrc = walk.src.virt.addr;
+               u8 *wdst = walk.dst.virt.addr;
 
-               fpu_enabled = cast5_fpu_begin(fpu_enabled, nbytes);
+               fpu_enabled = cast5_fpu_begin(fpu_enabled, &walk, nbytes);
 
                /* Process multi-block batch */
                if (nbytes >= bsize * CAST5_PARALLEL_BLOCKS) {
@@ -102,76 +105,58 @@ static int ecb_crypt(struct blkcipher_desc *desc, struct 
blkcipher_walk *walk,
                } while (nbytes >= bsize);
 
 done:
-               err = blkcipher_walk_done(desc, walk, nbytes);
+               err = skcipher_walk_done(&walk, nbytes);
        }
 
        cast5_fpu_end(fpu_enabled);
        return err;
 }
 
-static int ecb_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
-                      struct scatterlist *src, unsigned int nbytes)
+static int ecb_encrypt(struct skcipher_request *req)
 {
-       struct blkcipher_walk walk;
-
-       blkcipher_walk_init(&walk, dst, src, nbytes);
-       return ecb_crypt(desc, &walk, true);
+       return ecb_crypt(req, true);
 }
 
-static int ecb_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
-                      struct scatterlist *src, unsigned int nbytes)
+static int ecb_decrypt(struct skcipher_request *req)
 {
-       struct blkcipher_walk walk;
-
-       blkcipher_walk_init(&walk, dst, src, nbytes);
-       return ecb_crypt(desc, &walk, false);
+       return ecb_crypt(req, false);
 }
 
-static unsigned int __cbc_encrypt(struct blkcipher_desc *desc,
-                                 struct blkcipher_walk *walk)
+static int cbc_encrypt(struct skcipher_request *req)
 {
-       struct cast5_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
        const unsigned int bsize = CAST5_BLOCK_SIZE;
-       unsigned int nbytes = walk->nbytes;
-       u64 *src = (u64 *)walk->src.virt.addr;
-       u64 *dst = (u64 *)walk->dst.virt.addr;
-       u64 *iv = (u64 *)walk->iv;
-
-       do {
-               *dst = *src ^ *iv;
-               __cast5_encrypt(ctx, (u8 *)dst, (u8 *)dst);
-               iv = dst;
-
-               src += 1;
-               dst += 1;
-               nbytes -= bsize;
-       } while (nbytes >= bsize);
-
-       *(u64 *)walk->iv = *iv;
-       return nbytes;
-}
-
-static int cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
-                      struct scatterlist *src, unsigned int nbytes)
-{
-       struct blkcipher_walk walk;
+       struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
+       struct cast5_ctx *ctx = crypto_skcipher_ctx(tfm);
+       struct skcipher_walk walk;
+       unsigned int nbytes;
        int err;
 
-       blkcipher_walk_init(&walk, dst, src, nbytes);
-       err = blkcipher_walk_virt(desc, &walk);
+       err = skcipher_walk_virt(&walk, req, false);
 
        while ((nbytes = walk.nbytes)) {
-               nbytes = __cbc_encrypt(desc, &walk);
-               err = blkcipher_walk_done(desc, &walk, nbytes);
+               u64 *src = (u64 *)walk.src.virt.addr;
+               u64 *dst = (u64 *)walk.dst.virt.addr;
+               u64 *iv = (u64 *)walk.iv;
+
+               do {
+                       *dst = *src ^ *iv;
+                       __cast5_encrypt(ctx, (u8 *)dst, (u8 *)dst);
+                       iv = dst;
+                       src++;
+                       dst++;
+                       nbytes -= bsize;
+               } while (nbytes >= bsize);
+
+               *(u64 *)walk.iv = *iv;
+               err = skcipher_walk_done(&walk, nbytes);
        }
 
        return err;
 }
 
-static unsigned int __cbc_decrypt(struct blkcipher_desc *desc,
-                                 struct blkcipher_walk *walk)
+static unsigned int __cbc_decrypt(struct cast5_ctx *ctx,
+                                 struct skcipher_walk *walk)
 {
-       struct cast5_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
        const unsigned int bsize = CAST5_BLOCK_SIZE;
        unsigned int nbytes = walk->nbytes;
        u64 *src = (u64 *)walk->src.virt.addr;
@@ -223,31 +208,29 @@ static unsigned int __cbc_decrypt(struct blkcipher_desc 
*desc,
        return nbytes;
 }
 
-static int cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
-                      struct scatterlist *src, unsigned int nbytes)
+static int cbc_decrypt(struct skcipher_request *req)
 {
+       struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
+       struct cast5_ctx *ctx = crypto_skcipher_ctx(tfm);
        bool fpu_enabled = false;
-       struct blkcipher_walk walk;
+       struct skcipher_walk walk;
+       unsigned int nbytes;
        int err;
 
-       blkcipher_walk_init(&walk, dst, src, nbytes);
-       err = blkcipher_walk_virt(desc, &walk);
-       desc->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP;
+       err = skcipher_walk_virt(&walk, req, false);
 
        while ((nbytes = walk.nbytes)) {
-               fpu_enabled = cast5_fpu_begin(fpu_enabled, nbytes);
-               nbytes = __cbc_decrypt(desc, &walk);
-               err = blkcipher_walk_done(desc, &walk, nbytes);
+               fpu_enabled = cast5_fpu_begin(fpu_enabled, &walk, nbytes);
+               nbytes = __cbc_decrypt(ctx, &walk);
+               err = skcipher_walk_done(&walk, nbytes);
        }
 
        cast5_fpu_end(fpu_enabled);
        return err;
 }
 
-static void ctr_crypt_final(struct blkcipher_desc *desc,
-                           struct blkcipher_walk *walk)
+static void ctr_crypt_final(struct skcipher_walk *walk, struct cast5_ctx *ctx)
 {
-       struct cast5_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
        u8 *ctrblk = walk->iv;
        u8 keystream[CAST5_BLOCK_SIZE];
        u8 *src = walk->src.virt.addr;
@@ -260,10 +243,9 @@ static void ctr_crypt_final(struct blkcipher_desc *desc,
        crypto_inc(ctrblk, CAST5_BLOCK_SIZE);
 }
 
-static unsigned int __ctr_crypt(struct blkcipher_desc *desc,
-                               struct blkcipher_walk *walk)
+static unsigned int __ctr_crypt(struct skcipher_walk *walk,
+                               struct cast5_ctx *ctx)
 {
-       struct cast5_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
        const unsigned int bsize = CAST5_BLOCK_SIZE;
        unsigned int nbytes = walk->nbytes;
        u64 *src = (u64 *)walk->src.virt.addr;
@@ -306,162 +288,80 @@ static unsigned int __ctr_crypt(struct blkcipher_desc 
*desc,
        return nbytes;
 }
 
-static int ctr_crypt(struct blkcipher_desc *desc, struct scatterlist *dst,
-                    struct scatterlist *src, unsigned int nbytes)
+static int ctr_crypt(struct skcipher_request *req)
 {
+       struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
+       struct cast5_ctx *ctx = crypto_skcipher_ctx(tfm);
        bool fpu_enabled = false;
-       struct blkcipher_walk walk;
+       struct skcipher_walk walk;
+       unsigned int nbytes;
        int err;
 
-       blkcipher_walk_init(&walk, dst, src, nbytes);
-       err = blkcipher_walk_virt_block(desc, &walk, CAST5_BLOCK_SIZE);
-       desc->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP;
+       err = skcipher_walk_virt(&walk, req, false);
 
        while ((nbytes = walk.nbytes) >= CAST5_BLOCK_SIZE) {
-               fpu_enabled = cast5_fpu_begin(fpu_enabled, nbytes);
-               nbytes = __ctr_crypt(desc, &walk);
-               err = blkcipher_walk_done(desc, &walk, nbytes);
+               fpu_enabled = cast5_fpu_begin(fpu_enabled, &walk, nbytes);
+               nbytes = __ctr_crypt(&walk, ctx);
+               err = skcipher_walk_done(&walk, nbytes);
        }
 
        cast5_fpu_end(fpu_enabled);
 
        if (walk.nbytes) {
-               ctr_crypt_final(desc, &walk);
-               err = blkcipher_walk_done(desc, &walk, 0);
+               ctr_crypt_final(&walk, ctx);
+               err = skcipher_walk_done(&walk, 0);
        }
 
        return err;
 }
 
+static struct skcipher_alg cast5_algs[] = {
+       {
+               .base.cra_name          = "__ecb(cast5)",
+               .base.cra_driver_name   = "__ecb-cast5-avx",
+               .base.cra_priority      = 200,
+               .base.cra_flags         = CRYPTO_ALG_INTERNAL,
+               .base.cra_blocksize     = CAST5_BLOCK_SIZE,
+               .base.cra_ctxsize       = sizeof(struct cast5_ctx),
+               .base.cra_module        = THIS_MODULE,
+               .min_keysize            = CAST5_MIN_KEY_SIZE,
+               .max_keysize            = CAST5_MAX_KEY_SIZE,
+               .setkey                 = cast5_setkey_skcipher,
+               .encrypt                = ecb_encrypt,
+               .decrypt                = ecb_decrypt,
+       }, {
+               .base.cra_name          = "__cbc(cast5)",
+               .base.cra_driver_name   = "__cbc-cast5-avx",
+               .base.cra_priority      = 200,
+               .base.cra_flags         = CRYPTO_ALG_INTERNAL,
+               .base.cra_blocksize     = CAST5_BLOCK_SIZE,
+               .base.cra_ctxsize       = sizeof(struct cast5_ctx),
+               .base.cra_module        = THIS_MODULE,
+               .min_keysize            = CAST5_MIN_KEY_SIZE,
+               .max_keysize            = CAST5_MAX_KEY_SIZE,
+               .ivsize                 = CAST5_BLOCK_SIZE,
+               .setkey                 = cast5_setkey_skcipher,
+               .encrypt                = cbc_encrypt,
+               .decrypt                = cbc_decrypt,
+       }, {
+               .base.cra_name          = "__ctr(cast5)",
+               .base.cra_driver_name   = "__ctr-cast5-avx",
+               .base.cra_priority      = 200,
+               .base.cra_flags         = CRYPTO_ALG_INTERNAL,
+               .base.cra_blocksize     = 1,
+               .base.cra_ctxsize       = sizeof(struct cast5_ctx),
+               .base.cra_module        = THIS_MODULE,
+               .min_keysize            = CAST5_MIN_KEY_SIZE,
+               .max_keysize            = CAST5_MAX_KEY_SIZE,
+               .ivsize                 = CAST5_BLOCK_SIZE,
+               .chunksize              = CAST5_BLOCK_SIZE,
+               .setkey                 = cast5_setkey_skcipher,
+               .encrypt                = ctr_crypt,
+               .decrypt                = ctr_crypt,
+       }
+};
 
-static struct crypto_alg cast5_algs[6] = { {
-       .cra_name               = "__ecb-cast5-avx",
-       .cra_driver_name        = "__driver-ecb-cast5-avx",
-       .cra_priority           = 0,
-       .cra_flags              = CRYPTO_ALG_TYPE_BLKCIPHER |
-                                 CRYPTO_ALG_INTERNAL,
-       .cra_blocksize          = CAST5_BLOCK_SIZE,
-       .cra_ctxsize            = sizeof(struct cast5_ctx),
-       .cra_alignmask          = 0,
-       .cra_type               = &crypto_blkcipher_type,
-       .cra_module             = THIS_MODULE,
-       .cra_u = {
-               .blkcipher = {
-                       .min_keysize    = CAST5_MIN_KEY_SIZE,
-                       .max_keysize    = CAST5_MAX_KEY_SIZE,
-                       .setkey         = cast5_setkey,
-                       .encrypt        = ecb_encrypt,
-                       .decrypt        = ecb_decrypt,
-               },
-       },
-}, {
-       .cra_name               = "__cbc-cast5-avx",
-       .cra_driver_name        = "__driver-cbc-cast5-avx",
-       .cra_priority           = 0,
-       .cra_flags              = CRYPTO_ALG_TYPE_BLKCIPHER |
-                                 CRYPTO_ALG_INTERNAL,
-       .cra_blocksize          = CAST5_BLOCK_SIZE,
-       .cra_ctxsize            = sizeof(struct cast5_ctx),
-       .cra_alignmask          = 0,
-       .cra_type               = &crypto_blkcipher_type,
-       .cra_module             = THIS_MODULE,
-       .cra_u = {
-               .blkcipher = {
-                       .min_keysize    = CAST5_MIN_KEY_SIZE,
-                       .max_keysize    = CAST5_MAX_KEY_SIZE,
-                       .setkey         = cast5_setkey,
-                       .encrypt        = cbc_encrypt,
-                       .decrypt        = cbc_decrypt,
-               },
-       },
-}, {
-       .cra_name               = "__ctr-cast5-avx",
-       .cra_driver_name        = "__driver-ctr-cast5-avx",
-       .cra_priority           = 0,
-       .cra_flags              = CRYPTO_ALG_TYPE_BLKCIPHER |
-                                 CRYPTO_ALG_INTERNAL,
-       .cra_blocksize          = 1,
-       .cra_ctxsize            = sizeof(struct cast5_ctx),
-       .cra_alignmask          = 0,
-       .cra_type               = &crypto_blkcipher_type,
-       .cra_module             = THIS_MODULE,
-       .cra_u = {
-               .blkcipher = {
-                       .min_keysize    = CAST5_MIN_KEY_SIZE,
-                       .max_keysize    = CAST5_MAX_KEY_SIZE,
-                       .ivsize         = CAST5_BLOCK_SIZE,
-                       .setkey         = cast5_setkey,
-                       .encrypt        = ctr_crypt,
-                       .decrypt        = ctr_crypt,
-               },
-       },
-}, {
-       .cra_name               = "ecb(cast5)",
-       .cra_driver_name        = "ecb-cast5-avx",
-       .cra_priority           = 200,
-       .cra_flags              = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
-       .cra_blocksize          = CAST5_BLOCK_SIZE,
-       .cra_ctxsize            = sizeof(struct async_helper_ctx),
-       .cra_alignmask          = 0,
-       .cra_type               = &crypto_ablkcipher_type,
-       .cra_module             = THIS_MODULE,
-       .cra_init               = ablk_init,
-       .cra_exit               = ablk_exit,
-       .cra_u = {
-               .ablkcipher = {
-                       .min_keysize    = CAST5_MIN_KEY_SIZE,
-                       .max_keysize    = CAST5_MAX_KEY_SIZE,
-                       .setkey         = ablk_set_key,
-                       .encrypt        = ablk_encrypt,
-                       .decrypt        = ablk_decrypt,
-               },
-       },
-}, {
-       .cra_name               = "cbc(cast5)",
-       .cra_driver_name        = "cbc-cast5-avx",
-       .cra_priority           = 200,
-       .cra_flags              = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
-       .cra_blocksize          = CAST5_BLOCK_SIZE,
-       .cra_ctxsize            = sizeof(struct async_helper_ctx),
-       .cra_alignmask          = 0,
-       .cra_type               = &crypto_ablkcipher_type,
-       .cra_module             = THIS_MODULE,
-       .cra_init               = ablk_init,
-       .cra_exit               = ablk_exit,
-       .cra_u = {
-               .ablkcipher = {
-                       .min_keysize    = CAST5_MIN_KEY_SIZE,
-                       .max_keysize    = CAST5_MAX_KEY_SIZE,
-                       .ivsize         = CAST5_BLOCK_SIZE,
-                       .setkey         = ablk_set_key,
-                       .encrypt        = __ablk_encrypt,
-                       .decrypt        = ablk_decrypt,
-               },
-       },
-}, {
-       .cra_name               = "ctr(cast5)",
-       .cra_driver_name        = "ctr-cast5-avx",
-       .cra_priority           = 200,
-       .cra_flags              = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
-       .cra_blocksize          = 1,
-       .cra_ctxsize            = sizeof(struct async_helper_ctx),
-       .cra_alignmask          = 0,
-       .cra_type               = &crypto_ablkcipher_type,
-       .cra_module             = THIS_MODULE,
-       .cra_init               = ablk_init,
-       .cra_exit               = ablk_exit,
-       .cra_u = {
-               .ablkcipher = {
-                       .min_keysize    = CAST5_MIN_KEY_SIZE,
-                       .max_keysize    = CAST5_MAX_KEY_SIZE,
-                       .ivsize         = CAST5_BLOCK_SIZE,
-                       .setkey         = ablk_set_key,
-                       .encrypt        = ablk_encrypt,
-                       .decrypt        = ablk_encrypt,
-                       .geniv          = "chainiv",
-               },
-       },
-} };
+static struct simd_skcipher_alg *cast5_simd_algs[ARRAY_SIZE(cast5_algs)];
 
 static int __init cast5_init(void)
 {
@@ -473,12 +373,15 @@ static int __init cast5_init(void)
                return -ENODEV;
        }
 
-       return crypto_register_algs(cast5_algs, ARRAY_SIZE(cast5_algs));
+       return simd_register_skciphers_compat(cast5_algs,
+                                             ARRAY_SIZE(cast5_algs),
+                                             cast5_simd_algs);
 }
 
 static void __exit cast5_exit(void)
 {
-       crypto_unregister_algs(cast5_algs, ARRAY_SIZE(cast5_algs));
+       simd_unregister_skciphers(cast5_algs, ARRAY_SIZE(cast5_algs),
+                                 cast5_simd_algs);
 }
 
 module_init(cast5_init);
diff --git a/crypto/Kconfig b/crypto/Kconfig
index 35d84cea7ddf1..20ac64afa02d3 100644
--- a/crypto/Kconfig
+++ b/crypto/Kconfig
@@ -1238,11 +1238,10 @@ config CRYPTO_CAST5
 config CRYPTO_CAST5_AVX_X86_64
        tristate "CAST5 (CAST-128) cipher algorithm (x86_64/AVX)"
        depends on X86 && 64BIT
-       select CRYPTO_ALGAPI
-       select CRYPTO_CRYPTD
-       select CRYPTO_ABLK_HELPER
-       select CRYPTO_CAST_COMMON
+       select CRYPTO_BLKCIPHER
        select CRYPTO_CAST5
+       select CRYPTO_CAST_COMMON
+       select CRYPTO_SIMD
        help
          The CAST5 encryption algorithm (synonymous with CAST-128) is
          described in RFC2144.
-- 
2.16.2

Reply via email to