crypto: engine - Permit to enqueue all async requests
authorCorentin LABBE <clabbe.montjoie@gmail.com>
Fri, 26 Jan 2018 19:15:30 +0000 (20:15 +0100)
committerHerbert Xu <herbert@gondor.apana.org.au>
Thu, 15 Feb 2018 15:26:50 +0000 (23:26 +0800)
The crypto engine could actually only enqueue hash and ablkcipher request.
This patch permit it to enqueue any type of crypto_async_request.

Signed-off-by: Corentin Labbe <clabbe.montjoie@gmail.com>
Tested-by: Fabien Dessenne <fabien.dessenne@st.com>
Tested-by: Fabien Dessenne <fabien.dessenne@st.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
crypto/crypto_engine.c
include/crypto/engine.h

index 61e7c4e02fd264b2228da7f25a13ee317767cc87..992e8d8dcdd91fe91ee26d8b49dfab32daafa5e3 100644 (file)
 #include <linux/err.h>
 #include <linux/delay.h>
 #include <crypto/engine.h>
-#include <crypto/internal/hash.h>
 #include <uapi/linux/sched/types.h>
 #include "internal.h"
 
 #define CRYPTO_ENGINE_MAX_QLEN 10
 
+/**
+ * crypto_finalize_request - finalize one request if the request is done
+ * @engine: the hardware engine
+ * @req: the request need to be finalized
+ * @err: error number
+ */
+static void crypto_finalize_request(struct crypto_engine *engine,
+                            struct crypto_async_request *req, int err)
+{
+       unsigned long flags;
+       bool finalize_cur_req = false;
+       int ret;
+       struct crypto_engine_ctx *enginectx;
+
+       spin_lock_irqsave(&engine->queue_lock, flags);
+       if (engine->cur_req == req)
+               finalize_cur_req = true;
+       spin_unlock_irqrestore(&engine->queue_lock, flags);
+
+       if (finalize_cur_req) {
+               enginectx = crypto_tfm_ctx(req->tfm);
+               if (engine->cur_req_prepared &&
+                   enginectx->op.unprepare_request) {
+                       ret = enginectx->op.unprepare_request(engine, req);
+                       if (ret)
+                               dev_err(engine->dev, "failed to unprepare request\n");
+               }
+               spin_lock_irqsave(&engine->queue_lock, flags);
+               engine->cur_req = NULL;
+               engine->cur_req_prepared = false;
+               spin_unlock_irqrestore(&engine->queue_lock, flags);
+       }
+
+       req->complete(req, err);
+
+       kthread_queue_work(engine->kworker, &engine->pump_requests);
+}
+
 /**
  * crypto_pump_requests - dequeue one request from engine queue to process
  * @engine: the hardware engine
@@ -34,11 +71,10 @@ static void crypto_pump_requests(struct crypto_engine *engine,
                                 bool in_kthread)
 {
        struct crypto_async_request *async_req, *backlog;
-       struct ahash_request *hreq;
-       struct ablkcipher_request *breq;
        unsigned long flags;
        bool was_busy = false;
-       int ret, rtype;
+       int ret;
+       struct crypto_engine_ctx *enginectx;
 
        spin_lock_irqsave(&engine->queue_lock, flags);
 
@@ -94,7 +130,6 @@ static void crypto_pump_requests(struct crypto_engine *engine,
 
        spin_unlock_irqrestore(&engine->queue_lock, flags);
 
-       rtype = crypto_tfm_alg_type(engine->cur_req->tfm);
        /* Until here we get the request need to be encrypted successfully */
        if (!was_busy && engine->prepare_crypt_hardware) {
                ret = engine->prepare_crypt_hardware(engine);
@@ -104,57 +139,31 @@ static void crypto_pump_requests(struct crypto_engine *engine,
                }
        }
 
-       switch (rtype) {
-       case CRYPTO_ALG_TYPE_AHASH:
-               hreq = ahash_request_cast(engine->cur_req);
-               if (engine->prepare_hash_request) {
-                       ret = engine->prepare_hash_request(engine, hreq);
-                       if (ret) {
-                               dev_err(engine->dev, "failed to prepare request: %d\n",
-                                       ret);
-                               goto req_err;
-                       }
-                       engine->cur_req_prepared = true;
-               }
-               ret = engine->hash_one_request(engine, hreq);
-               if (ret) {
-                       dev_err(engine->dev, "failed to hash one request from queue\n");
-                       goto req_err;
-               }
-               return;
-       case CRYPTO_ALG_TYPE_ABLKCIPHER:
-               breq = ablkcipher_request_cast(engine->cur_req);
-               if (engine->prepare_cipher_request) {
-                       ret = engine->prepare_cipher_request(engine, breq);
-                       if (ret) {
-                               dev_err(engine->dev, "failed to prepare request: %d\n",
-                                       ret);
-                               goto req_err;
-                       }
-                       engine->cur_req_prepared = true;
-               }
-               ret = engine->cipher_one_request(engine, breq);
+       enginectx = crypto_tfm_ctx(async_req->tfm);
+
+       if (enginectx->op.prepare_request) {
+               ret = enginectx->op.prepare_request(engine, async_req);
                if (ret) {
-                       dev_err(engine->dev, "failed to cipher one request from queue\n");
+                       dev_err(engine->dev, "failed to prepare request: %d\n",
+                               ret);
                        goto req_err;
                }
-               return;
-       default:
-               dev_err(engine->dev, "failed to prepare request of unknown type\n");
-               return;
+               engine->cur_req_prepared = true;
+       }
+       if (!enginectx->op.do_one_request) {
+               dev_err(engine->dev, "failed to do request\n");
+               ret = -EINVAL;
+               goto req_err;
        }
+       ret = enginectx->op.do_one_request(engine, async_req);
+       if (ret) {
+               dev_err(engine->dev, "Failed to do one request from queue: %d\n", ret);
+               goto req_err;
+       }
+       return;
 
 req_err:
-       switch (rtype) {
-       case CRYPTO_ALG_TYPE_AHASH:
-               hreq = ahash_request_cast(engine->cur_req);
-               crypto_finalize_hash_request(engine, hreq, ret);
-               break;
-       case CRYPTO_ALG_TYPE_ABLKCIPHER:
-               breq = ablkcipher_request_cast(engine->cur_req);
-               crypto_finalize_cipher_request(engine, breq, ret);
-               break;
-       }
+       crypto_finalize_request(engine, async_req, ret);
        return;
 
 out:
@@ -170,13 +179,12 @@ static void crypto_pump_work(struct kthread_work *work)
 }
 
 /**
- * crypto_transfer_cipher_request - transfer the new request into the
- * enginequeue
+ * crypto_transfer_request - transfer the new request into the engine queue
  * @engine: the hardware engine
  * @req: the request need to be listed into the engine queue
  */
-int crypto_transfer_cipher_request(struct crypto_engine *engine,
-                                  struct ablkcipher_request *req,
+static int crypto_transfer_request(struct crypto_engine *engine,
+                                  struct crypto_async_request *req,
                                   bool need_pump)
 {
        unsigned long flags;
@@ -189,7 +197,7 @@ int crypto_transfer_cipher_request(struct crypto_engine *engine,
                return -ESHUTDOWN;
        }
 
-       ret = ablkcipher_enqueue_request(&engine->queue, req);
+       ret = crypto_enqueue_request(&engine->queue, req);
 
        if (!engine->busy && need_pump)
                kthread_queue_work(engine->kworker, &engine->pump_requests);
@@ -197,102 +205,131 @@ int crypto_transfer_cipher_request(struct crypto_engine *engine,
        spin_unlock_irqrestore(&engine->queue_lock, flags);
        return ret;
 }
-EXPORT_SYMBOL_GPL(crypto_transfer_cipher_request);
 
 /**
- * crypto_transfer_cipher_request_to_engine - transfer one request to list
+ * crypto_transfer_request_to_engine - transfer one request to list
  * into the engine queue
  * @engine: the hardware engine
  * @req: the request need to be listed into the engine queue
  */
-int crypto_transfer_cipher_request_to_engine(struct crypto_engine *engine,
-                                            struct ablkcipher_request *req)
+static int crypto_transfer_request_to_engine(struct crypto_engine *engine,
+                                            struct crypto_async_request *req)
 {
-       return crypto_transfer_cipher_request(engine, req, true);
+       return crypto_transfer_request(engine, req, true);
 }
-EXPORT_SYMBOL_GPL(crypto_transfer_cipher_request_to_engine);
 
 /**
- * crypto_transfer_hash_request - transfer the new request into the
- * enginequeue
+ * crypto_transfer_ablkcipher_request_to_engine - transfer one ablkcipher_request
+ * to list into the engine queue
  * @engine: the hardware engine
  * @req: the request need to be listed into the engine queue
+ * TODO: Remove this function when skcipher conversion is finished
  */
-int crypto_transfer_hash_request(struct crypto_engine *engine,
-                                struct ahash_request *req, bool need_pump)
+int crypto_transfer_ablkcipher_request_to_engine(struct crypto_engine *engine,
+                                                struct ablkcipher_request *req)
 {
-       unsigned long flags;
-       int ret;
-
-       spin_lock_irqsave(&engine->queue_lock, flags);
-
-       if (!engine->running) {
-               spin_unlock_irqrestore(&engine->queue_lock, flags);
-               return -ESHUTDOWN;
-       }
-
-       ret = ahash_enqueue_request(&engine->queue, req);
+       return crypto_transfer_request_to_engine(engine, &req->base);
+}
+EXPORT_SYMBOL_GPL(crypto_transfer_ablkcipher_request_to_engine);
 
-       if (!engine->busy && need_pump)
-               kthread_queue_work(engine->kworker, &engine->pump_requests);
+/**
+ * crypto_transfer_aead_request_to_engine - transfer one aead_request
+ * to list into the engine queue
+ * @engine: the hardware engine
+ * @req: the request need to be listed into the engine queue
+ */
+int crypto_transfer_aead_request_to_engine(struct crypto_engine *engine,
+                                          struct aead_request *req)
+{
+       return crypto_transfer_request_to_engine(engine, &req->base);
+}
+EXPORT_SYMBOL_GPL(crypto_transfer_aead_request_to_engine);
 
-       spin_unlock_irqrestore(&engine->queue_lock, flags);
-       return ret;
+/**
+ * crypto_transfer_akcipher_request_to_engine - transfer one akcipher_request
+ * to list into the engine queue
+ * @engine: the hardware engine
+ * @req: the request need to be listed into the engine queue
+ */
+int crypto_transfer_akcipher_request_to_engine(struct crypto_engine *engine,
+                                              struct akcipher_request *req)
+{
+       return crypto_transfer_request_to_engine(engine, &req->base);
 }
-EXPORT_SYMBOL_GPL(crypto_transfer_hash_request);
+EXPORT_SYMBOL_GPL(crypto_transfer_akcipher_request_to_engine);
 
 /**
- * crypto_transfer_hash_request_to_engine - transfer one request to list
- * into the engine queue
+ * crypto_transfer_hash_request_to_engine - transfer one ahash_request
+ * to list into the engine queue
  * @engine: the hardware engine
  * @req: the request need to be listed into the engine queue
  */
 int crypto_transfer_hash_request_to_engine(struct crypto_engine *engine,
                                           struct ahash_request *req)
 {
-       return crypto_transfer_hash_request(engine, req, true);
+       return crypto_transfer_request_to_engine(engine, &req->base);
 }
 EXPORT_SYMBOL_GPL(crypto_transfer_hash_request_to_engine);
 
 /**
- * crypto_finalize_cipher_request - finalize one request if the request is done
+ * crypto_transfer_skcipher_request_to_engine - transfer one skcipher_request
+ * to list into the engine queue
+ * @engine: the hardware engine
+ * @req: the request need to be listed into the engine queue
+ */
+int crypto_transfer_skcipher_request_to_engine(struct crypto_engine *engine,
+                                              struct skcipher_request *req)
+{
+       return crypto_transfer_request_to_engine(engine, &req->base);
+}
+EXPORT_SYMBOL_GPL(crypto_transfer_skcipher_request_to_engine);
+
+/**
+ * crypto_finalize_ablkcipher_request - finalize one ablkcipher_request if
+ * the request is done
  * @engine: the hardware engine
  * @req: the request need to be finalized
  * @err: error number
+ * TODO: Remove this function when skcipher conversion is finished
  */
-void crypto_finalize_cipher_request(struct crypto_engine *engine,
-                                   struct ablkcipher_request *req, int err)
+void crypto_finalize_ablkcipher_request(struct crypto_engine *engine,
+                                       struct ablkcipher_request *req, int err)
 {
-       unsigned long flags;
-       bool finalize_cur_req = false;
-       int ret;
-
-       spin_lock_irqsave(&engine->queue_lock, flags);
-       if (engine->cur_req == &req->base)
-               finalize_cur_req = true;
-       spin_unlock_irqrestore(&engine->queue_lock, flags);
-
-       if (finalize_cur_req) {
-               if (engine->cur_req_prepared &&
-                   engine->unprepare_cipher_request) {
-                       ret = engine->unprepare_cipher_request(engine, req);
-                       if (ret)
-                               dev_err(engine->dev, "failed to unprepare request\n");
-               }
-               spin_lock_irqsave(&engine->queue_lock, flags);
-               engine->cur_req = NULL;
-               engine->cur_req_prepared = false;
-               spin_unlock_irqrestore(&engine->queue_lock, flags);
-       }
+       return crypto_finalize_request(engine, &req->base, err);
+}
+EXPORT_SYMBOL_GPL(crypto_finalize_ablkcipher_request);
 
-       req->base.complete(&req->base, err);
+/**
+ * crypto_finalize_aead_request - finalize one aead_request if
+ * the request is done
+ * @engine: the hardware engine
+ * @req: the request need to be finalized
+ * @err: error number
+ */
+void crypto_finalize_aead_request(struct crypto_engine *engine,
+                                 struct aead_request *req, int err)
+{
+       return crypto_finalize_request(engine, &req->base, err);
+}
+EXPORT_SYMBOL_GPL(crypto_finalize_aead_request);
 
-       kthread_queue_work(engine->kworker, &engine->pump_requests);
+/**
+ * crypto_finalize_akcipher_request - finalize one akcipher_request if
+ * the request is done
+ * @engine: the hardware engine
+ * @req: the request need to be finalized
+ * @err: error number
+ */
+void crypto_finalize_akcipher_request(struct crypto_engine *engine,
+                                     struct akcipher_request *req, int err)
+{
+       return crypto_finalize_request(engine, &req->base, err);
 }
-EXPORT_SYMBOL_GPL(crypto_finalize_cipher_request);
+EXPORT_SYMBOL_GPL(crypto_finalize_akcipher_request);
 
 /**
- * crypto_finalize_hash_request - finalize one request if the request is done
+ * crypto_finalize_hash_request - finalize one ahash_request if
+ * the request is done
  * @engine: the hardware engine
  * @req: the request need to be finalized
  * @err: error number
@@ -300,34 +337,24 @@ EXPORT_SYMBOL_GPL(crypto_finalize_cipher_request);
 void crypto_finalize_hash_request(struct crypto_engine *engine,
                                  struct ahash_request *req, int err)
 {
-       unsigned long flags;
-       bool finalize_cur_req = false;
-       int ret;
-
-       spin_lock_irqsave(&engine->queue_lock, flags);
-       if (engine->cur_req == &req->base)
-               finalize_cur_req = true;
-       spin_unlock_irqrestore(&engine->queue_lock, flags);
-
-       if (finalize_cur_req) {
-               if (engine->cur_req_prepared &&
-                   engine->unprepare_hash_request) {
-                       ret = engine->unprepare_hash_request(engine, req);
-                       if (ret)
-                               dev_err(engine->dev, "failed to unprepare request\n");
-               }
-               spin_lock_irqsave(&engine->queue_lock, flags);
-               engine->cur_req = NULL;
-               engine->cur_req_prepared = false;
-               spin_unlock_irqrestore(&engine->queue_lock, flags);
-       }
-
-       req->base.complete(&req->base, err);
-
-       kthread_queue_work(engine->kworker, &engine->pump_requests);
+       return crypto_finalize_request(engine, &req->base, err);
 }
 EXPORT_SYMBOL_GPL(crypto_finalize_hash_request);
 
+/**
+ * crypto_finalize_skcipher_request - finalize one skcipher_request if
+ * the request is done
+ * @engine: the hardware engine
+ * @req: the request need to be finalized
+ * @err: error number
+ */
+void crypto_finalize_skcipher_request(struct crypto_engine *engine,
+                                     struct skcipher_request *req, int err)
+{
+       return crypto_finalize_request(engine, &req->base, err);
+}
+EXPORT_SYMBOL_GPL(crypto_finalize_skcipher_request);
+
 /**
  * crypto_engine_start - start the hardware engine
  * @engine: the hardware engine need to be started
index dd04c1699b5186cbf8331b0f34006977b5ccde57..1cbec29af3d6806f746d23edae3ed42f7f8a68ee 100644 (file)
 #include <linux/kernel.h>
 #include <linux/kthread.h>
 #include <crypto/algapi.h>
+#include <crypto/aead.h>
+#include <crypto/akcipher.h>
 #include <crypto/hash.h>
+#include <crypto/skcipher.h>
 
 #define ENGINE_NAME_LEN        30
 /*
  * @unprepare_crypt_hardware: there are currently no more requests on the
  * queue so the subsystem notifies the driver that it may relax the
  * hardware by issuing this call
- * @prepare_cipher_request: do some prepare if need before handle the current request
- * @unprepare_cipher_request: undo any work done by prepare_cipher_request()
- * @cipher_one_request: do encryption for current request
- * @prepare_hash_request: do some prepare if need before handle the current request
- * @unprepare_hash_request: undo any work done by prepare_hash_request()
- * @hash_one_request: do hash for current request
  * @kworker: kthread worker struct for request pump
  * @pump_requests: work struct for scheduling work to the request pump
  * @priv_data: the engine private data
@@ -65,19 +62,6 @@ struct crypto_engine {
        int (*prepare_crypt_hardware)(struct crypto_engine *engine);
        int (*unprepare_crypt_hardware)(struct crypto_engine *engine);
 
-       int (*prepare_cipher_request)(struct crypto_engine *engine,
-                                     struct ablkcipher_request *req);
-       int (*unprepare_cipher_request)(struct crypto_engine *engine,
-                                       struct ablkcipher_request *req);
-       int (*prepare_hash_request)(struct crypto_engine *engine,
-                                   struct ahash_request *req);
-       int (*unprepare_hash_request)(struct crypto_engine *engine,
-                                     struct ahash_request *req);
-       int (*cipher_one_request)(struct crypto_engine *engine,
-                                 struct ablkcipher_request *req);
-       int (*hash_one_request)(struct crypto_engine *engine,
-                               struct ahash_request *req);
-
        struct kthread_worker           *kworker;
        struct kthread_work             pump_requests;
 
@@ -85,19 +69,45 @@ struct crypto_engine {
        struct crypto_async_request     *cur_req;
 };
 
-int crypto_transfer_cipher_request(struct crypto_engine *engine,
-                                  struct ablkcipher_request *req,
-                                  bool need_pump);
-int crypto_transfer_cipher_request_to_engine(struct crypto_engine *engine,
-                                            struct ablkcipher_request *req);
-int crypto_transfer_hash_request(struct crypto_engine *engine,
-                                struct ahash_request *req, bool need_pump);
+/*
+ * struct crypto_engine_op - crypto hardware engine operations
+ * @prepare__request: do some prepare if need before handle the current request
+ * @unprepare_request: undo any work done by prepare_request()
+ * @do_one_request: do encryption for current request
+ */
+struct crypto_engine_op {
+       int (*prepare_request)(struct crypto_engine *engine,
+                              void *areq);
+       int (*unprepare_request)(struct crypto_engine *engine,
+                                void *areq);
+       int (*do_one_request)(struct crypto_engine *engine,
+                             void *areq);
+};
+
+struct crypto_engine_ctx {
+       struct crypto_engine_op op;
+};
+
+int crypto_transfer_ablkcipher_request_to_engine(struct crypto_engine *engine,
+                                                struct ablkcipher_request *req);
+int crypto_transfer_aead_request_to_engine(struct crypto_engine *engine,
+                                          struct aead_request *req);
+int crypto_transfer_akcipher_request_to_engine(struct crypto_engine *engine,
+                                              struct akcipher_request *req);
 int crypto_transfer_hash_request_to_engine(struct crypto_engine *engine,
-                                          struct ahash_request *req);
-void crypto_finalize_cipher_request(struct crypto_engine *engine,
-                                   struct ablkcipher_request *req, int err);
+                                              struct ahash_request *req);
+int crypto_transfer_skcipher_request_to_engine(struct crypto_engine *engine,
+                                              struct skcipher_request *req);
+void crypto_finalize_ablkcipher_request(struct crypto_engine *engine,
+                                       struct ablkcipher_request *req, int err);
+void crypto_finalize_aead_request(struct crypto_engine *engine,
+                                 struct aead_request *req, int err);
+void crypto_finalize_akcipher_request(struct crypto_engine *engine,
+                                     struct akcipher_request *req, int err);
 void crypto_finalize_hash_request(struct crypto_engine *engine,
                                  struct ahash_request *req, int err);
+void crypto_finalize_skcipher_request(struct crypto_engine *engine,
+                                     struct skcipher_request *req, int err);
 int crypto_engine_start(struct crypto_engine *engine);
 int crypto_engine_stop(struct crypto_engine *engine);
 struct crypto_engine *crypto_engine_alloc_init(struct device *dev, bool rt);