struct skcipher_async_req *sreq;
struct ablkcipher_request *req;
struct skcipher_async_rsgl *last_rsgl = NULL;
- unsigned int len = 0, tx_nents = skcipher_all_sg_nents(ctx);
+ unsigned int txbufs = 0, len = 0, tx_nents = skcipher_all_sg_nents(ctx);
unsigned int reqlen = sizeof(struct skcipher_async_req) +
GET_REQ_SIZE(ctx) + GET_IV_SIZE(ctx);
- int i = 0;
int err = -ENOMEM;
+ bool mark = false;
lock_sock(sk);
req = kmalloc(reqlen, GFP_KERNEL);
iov_iter_count(&msg->msg_iter));
used = min_t(unsigned long, used, sg->length);
- if (i == tx_nents) {
+ if (txbufs == tx_nents) {
struct scatterlist *tmp;
int x;
/* Ran out of tx slots in async request
kfree(sreq->tsg);
sreq->tsg = tmp;
tx_nents *= 2;
+ mark = true;
}
/* Need to take over the tx sgl from ctx
* to the asynch req - these sgls will be freed later */
- sg_set_page(sreq->tsg + i++, sg_page(sg), sg->length,
+ sg_set_page(sreq->tsg + txbufs++, sg_page(sg), sg->length,
sg->offset);
if (list_empty(&sreq->list)) {
iov_iter_advance(&msg->msg_iter, used);
}
+ if (mark)
+ sg_mark_end(sreq->tsg + txbufs - 1);
+
ablkcipher_request_set_crypt(req, sreq->tsg, sreq->first_sgl.sgl.sg,
len, sreq->iv);
err = ctx->enc ? crypto_ablkcipher_encrypt(req) :