From patchwork Fri Jun 26 16:16:33 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Mikulas Patocka X-Patchwork-Id: 11627957 X-Patchwork-Delegate: herbert@gondor.apana.org.au Return-Path: Received: from mail.kernel.org (pdx-korg-mail-1.web.codeaurora.org [172.30.200.123]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id 8748A92A for ; Fri, 26 Jun 2020 16:16:51 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 5E736207E8 for ; Fri, 26 Jun 2020 16:16:51 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=redhat.com header.i=@redhat.com header.b="VjNRXH9T" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1725890AbgFZQQu (ORCPT ); Fri, 26 Jun 2020 12:16:50 -0400 Received: from us-smtp-delivery-1.mimecast.com ([205.139.110.120]:35353 "EHLO us-smtp-1.mimecast.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1725780AbgFZQQt (ORCPT ); Fri, 26 Jun 2020 12:16:49 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1593188205; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references; bh=00/Af/+XCMPgVUCYVD4QlX7H1DxjWBahFVO8HcgrFm4=; b=VjNRXH9T5nQ++H9zQzLRvySnTZLYSXR3h5zY33bobW1muNpbmj9tPpNsPjiMI922jdJrdY g+Xitsu9JTouQfoZ52sbT5RF0Spt8J92LeG1DYejszPxDLmUKdClEbiMyx8y9yzywTa9x1 f03OMFoCd+SC8tzyL1t8ka8BK0rvfzc= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-350-dwb4F9t_NdOGtz9NM85NHA-1; Fri, 26 Jun 2020 12:16:41 -0400 X-MC-Unique: dwb4F9t_NdOGtz9NM85NHA-1 Received: from smtp.corp.redhat.com (int-mx02.intmail.prod.int.phx2.redhat.com [10.5.11.12]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id A2C3B107ACCA; Fri, 26 Jun 2020 16:16:39 +0000 (UTC) Received: from file01.intranet.prod.int.rdu2.redhat.com (file01.intranet.prod.int.rdu2.redhat.com [10.11.5.7]) by smtp.corp.redhat.com (Postfix) with ESMTPS id 0205560E1C; Fri, 26 Jun 2020 16:16:35 +0000 (UTC) Received: from file01.intranet.prod.int.rdu2.redhat.com (localhost [127.0.0.1]) by file01.intranet.prod.int.rdu2.redhat.com (8.14.4/8.14.4) with ESMTP id 05QGGZYk018593; Fri, 26 Jun 2020 12:16:35 -0400 Received: from localhost (mpatocka@localhost) by file01.intranet.prod.int.rdu2.redhat.com (8.14.4/8.14.4/Submit) with ESMTP id 05QGGXq4018589; Fri, 26 Jun 2020 12:16:33 -0400 X-Authentication-Warning: file01.intranet.prod.int.rdu2.redhat.com: mpatocka owned process doing -bs Date: Fri, 26 Jun 2020 12:16:33 -0400 (EDT) From: Mikulas Patocka X-X-Sender: mpatocka@file01.intranet.prod.int.rdu2.redhat.com To: Herbert Xu cc: Eric Biggers , Mike Snitzer , linux-kernel@vger.kernel.org, dm-devel@redhat.com, linux-crypto@vger.kernel.org, "David S. Miller" , Milan Broz , Jonathan Cameron , George Cherian , Wei Xu , Zaibo Xu Subject: [PATCH 1/3 v2] crypto: introduce the flag CRYPTO_ALG_ALLOCATES_MEMORY In-Reply-To: Message-ID: References: <20200610010450.GA6449@gondor.apana.org.au> <20200610121106.GA23137@gondor.apana.org.au> <20200616173620.GA207319@gmail.com> <20200626044534.GA2870@gondor.apana.org.au> User-Agent: Alpine 2.02 (LRH 1266 2009-07-14) MIME-Version: 1.0 X-Scanned-By: MIMEDefang 2.79 on 10.5.11.12 Sender: linux-crypto-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-crypto@vger.kernel.org Introduce a new flag CRYPTO_ALG_ALLOCATES_MEMORY and pass it down the crypto stack. If the flag is set, then the crypto driver allocates memory in its request routine. Such drivers are not suitable for disk encryption because GFP_ATOMIC allocation can fail anytime (causing random I/O errors) and GFP_KERNEL allocation can recurse into the block layer, causing a deadlock. Pass the flag CRYPTO_ALG_ALLOCATES_MEMORY down through the crypto API. Signed-off-by: Mikulas Patocka --- crypto/adiantum.c | 3 ++- crypto/authenc.c | 5 +++-- crypto/authencesn.c | 5 +++-- crypto/ccm.c | 7 ++++--- crypto/chacha20poly1305.c | 5 +++-- crypto/cryptd.c | 9 ++++++--- crypto/ctr.c | 3 ++- crypto/cts.c | 5 +++-- crypto/essiv.c | 5 +++-- crypto/gcm.c | 15 +++++++++------ crypto/geniv.c | 3 ++- crypto/lrw.c | 5 +++-- crypto/rsa-pkcs1pad.c | 5 +++-- crypto/xts.c | 2 +- include/crypto/algapi.h | 9 +++++++++ include/linux/crypto.h | 12 ++++++++++++ 16 files changed, 68 insertions(+), 30 deletions(-) Index: linux-2.6/include/linux/crypto.h =================================================================== --- linux-2.6.orig/include/linux/crypto.h 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/include/linux/crypto.h 2020-06-26 17:25:28.066417000 +0200 @@ -102,6 +102,18 @@ #define CRYPTO_NOLOAD 0x00008000 /* + * The driver may allocate memory during request processing, so it shouldn't be + * used in cases where memory allocation failures aren't acceptable, such as + * during block device encryption. + */ +#define CRYPTO_ALG_ALLOCATES_MEMORY 0x00010000 + +/* + * Pass these flags down through the crypto API. + */ +#define CRYPTO_ALG_INHERITED_FLAGS (CRYPTO_ALG_ASYNC | CRYPTO_ALG_ALLOCATES_MEMORY) + +/* * Transform masks and values (for crt_flags). */ #define CRYPTO_TFM_NEED_KEY 0x00000001 Index: linux-2.6/crypto/authenc.c =================================================================== --- linux-2.6.orig/crypto/authenc.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/authenc.c 2020-06-26 17:24:03.566417000 +0200 @@ -388,7 +388,8 @@ static int crypto_authenc_create(struct if ((algt->type ^ CRYPTO_ALG_TYPE_AEAD) & algt->mask) return -EINVAL; - mask = crypto_requires_sync(algt->type, algt->mask); + mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask); inst = kzalloc(sizeof(*inst) + sizeof(*ctx), GFP_KERNEL); if (!inst) @@ -424,7 +425,7 @@ static int crypto_authenc_create(struct goto err_free_inst; inst->alg.base.cra_flags = (auth_base->cra_flags | - enc->base.cra_flags) & CRYPTO_ALG_ASYNC; + enc->base.cra_flags) & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = enc->base.cra_priority * 10 + auth_base->cra_priority; inst->alg.base.cra_blocksize = enc->base.cra_blocksize; Index: linux-2.6/crypto/authencesn.c =================================================================== --- linux-2.6.orig/crypto/authencesn.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/authencesn.c 2020-06-26 17:24:03.566417000 +0200 @@ -406,7 +406,8 @@ static int crypto_authenc_esn_create(str if ((algt->type ^ CRYPTO_ALG_TYPE_AEAD) & algt->mask) return -EINVAL; - mask = crypto_requires_sync(algt->type, algt->mask); + mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask); inst = kzalloc(sizeof(*inst) + sizeof(*ctx), GFP_KERNEL); if (!inst) @@ -438,7 +439,7 @@ static int crypto_authenc_esn_create(str goto err_free_inst; inst->alg.base.cra_flags = (auth_base->cra_flags | - enc->base.cra_flags) & CRYPTO_ALG_ASYNC; + enc->base.cra_flags) & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = enc->base.cra_priority * 10 + auth_base->cra_priority; inst->alg.base.cra_blocksize = enc->base.cra_blocksize; Index: linux-2.6/crypto/ccm.c =================================================================== --- linux-2.6.orig/crypto/ccm.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/ccm.c 2020-06-26 17:24:03.566417000 +0200 @@ -462,7 +462,8 @@ static int crypto_ccm_create_common(stru if ((algt->type ^ CRYPTO_ALG_TYPE_AEAD) & algt->mask) return -EINVAL; - mask = crypto_requires_sync(algt->type, algt->mask); + mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask); inst = kzalloc(sizeof(*inst) + sizeof(*ictx), GFP_KERNEL); if (!inst) @@ -507,7 +508,7 @@ static int crypto_ccm_create_common(stru mac->base.cra_driver_name) >= CRYPTO_MAX_ALG_NAME) goto err_free_inst; - inst->alg.base.cra_flags = ctr->base.cra_flags & CRYPTO_ALG_ASYNC; + inst->alg.base.cra_flags = ctr->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = (mac->base.cra_priority + ctr->base.cra_priority) / 2; inst->alg.base.cra_blocksize = 1; @@ -759,7 +760,7 @@ static int crypto_rfc4309_create(struct CRYPTO_MAX_ALG_NAME) goto err_free_inst; - inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_ASYNC; + inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = alg->base.cra_priority; inst->alg.base.cra_blocksize = 1; inst->alg.base.cra_alignmask = alg->base.cra_alignmask; Index: linux-2.6/crypto/chacha20poly1305.c =================================================================== --- linux-2.6.orig/crypto/chacha20poly1305.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/chacha20poly1305.c 2020-06-26 17:24:03.566417000 +0200 @@ -573,7 +573,8 @@ static int chachapoly_create(struct cryp if ((algt->type ^ CRYPTO_ALG_TYPE_AEAD) & algt->mask) return -EINVAL; - mask = crypto_requires_sync(algt->type, algt->mask); + mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask); inst = kzalloc(sizeof(*inst) + sizeof(*ctx), GFP_KERNEL); if (!inst) @@ -614,7 +615,7 @@ static int chachapoly_create(struct cryp goto err_free_inst; inst->alg.base.cra_flags = (chacha->base.cra_flags | - poly->base.cra_flags) & CRYPTO_ALG_ASYNC; + poly->base.cra_flags) & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = (chacha->base.cra_priority + poly->base.cra_priority) / 2; inst->alg.base.cra_blocksize = 1; Index: linux-2.6/crypto/cryptd.c =================================================================== --- linux-2.6.orig/crypto/cryptd.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/cryptd.c 2020-06-26 17:24:03.566417000 +0200 @@ -396,7 +396,8 @@ static int cryptd_create_skcipher(struct goto err_free_inst; inst->alg.base.cra_flags = CRYPTO_ALG_ASYNC | - (alg->base.cra_flags & CRYPTO_ALG_INTERNAL); + (alg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS) | + (alg->base.cra_flags & CRYPTO_ALG_INTERNAL); inst->alg.ivsize = crypto_skcipher_alg_ivsize(alg); inst->alg.chunksize = crypto_skcipher_alg_chunksize(alg); @@ -663,7 +664,8 @@ static int cryptd_create_hash(struct cry inst->alg.halg.base.cra_flags = CRYPTO_ALG_ASYNC | (alg->base.cra_flags & (CRYPTO_ALG_INTERNAL | - CRYPTO_ALG_OPTIONAL_KEY)); + CRYPTO_ALG_OPTIONAL_KEY | + CRYPTO_ALG_INHERITED_FLAGS)); inst->alg.halg.digestsize = alg->digestsize; inst->alg.halg.statesize = alg->statesize; @@ -849,7 +851,8 @@ static int cryptd_create_aead(struct cry goto err_free_inst; inst->alg.base.cra_flags = CRYPTO_ALG_ASYNC | - (alg->base.cra_flags & CRYPTO_ALG_INTERNAL); + (alg->base.cra_flags & CRYPTO_ALG_INTERNAL) | + (alg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS); inst->alg.base.cra_ctxsize = sizeof(struct cryptd_aead_ctx); inst->alg.ivsize = crypto_aead_alg_ivsize(alg); Index: linux-2.6/crypto/ctr.c =================================================================== --- linux-2.6.orig/crypto/ctr.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/ctr.c 2020-06-26 17:24:03.566417000 +0200 @@ -276,6 +276,7 @@ static int crypto_rfc3686_create(struct return -ENOMEM; mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask) | crypto_requires_off(algt->type, algt->mask, CRYPTO_ALG_NEED_FALLBACK); @@ -310,7 +311,7 @@ static int crypto_rfc3686_create(struct inst->alg.base.cra_blocksize = 1; inst->alg.base.cra_alignmask = alg->base.cra_alignmask; - inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_ASYNC; + inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.ivsize = CTR_RFC3686_IV_SIZE; inst->alg.chunksize = crypto_skcipher_alg_chunksize(alg); Index: linux-2.6/crypto/cts.c =================================================================== --- linux-2.6.orig/crypto/cts.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/cts.c 2020-06-26 17:24:03.566417000 +0200 @@ -337,7 +337,8 @@ static int crypto_cts_create(struct cryp if ((algt->type ^ CRYPTO_ALG_TYPE_SKCIPHER) & algt->mask) return -EINVAL; - mask = crypto_requires_sync(algt->type, algt->mask); + mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask); inst = kzalloc(sizeof(*inst) + sizeof(*spawn), GFP_KERNEL); if (!inst) @@ -364,7 +365,7 @@ static int crypto_cts_create(struct cryp if (err) goto err_free_inst; - inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_ASYNC; + inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = alg->base.cra_priority; inst->alg.base.cra_blocksize = alg->base.cra_blocksize; inst->alg.base.cra_alignmask = alg->base.cra_alignmask; Index: linux-2.6/crypto/essiv.c =================================================================== --- linux-2.6.orig/crypto/essiv.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/essiv.c 2020-06-26 17:24:03.566417000 +0200 @@ -466,7 +466,8 @@ static int essiv_create(struct crypto_te return PTR_ERR(shash_name); type = algt->type & algt->mask; - mask = crypto_requires_sync(algt->type, algt->mask); + mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask); switch (type) { case CRYPTO_ALG_TYPE_SKCIPHER: @@ -557,7 +558,7 @@ static int essiv_create(struct crypto_te hash_alg->base.cra_driver_name) >= CRYPTO_MAX_ALG_NAME) goto out_free_hash; - base->cra_flags = block_base->cra_flags & CRYPTO_ALG_ASYNC; + base->cra_flags = block_base->cra_flags & CRYPTO_ALG_INHERITED_FLAGS; base->cra_blocksize = block_base->cra_blocksize; base->cra_ctxsize = sizeof(struct essiv_tfm_ctx); base->cra_alignmask = block_base->cra_alignmask; Index: linux-2.6/crypto/gcm.c =================================================================== --- linux-2.6.orig/crypto/gcm.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/gcm.c 2020-06-26 17:24:03.566417000 +0200 @@ -593,7 +593,8 @@ static int crypto_gcm_create_common(stru if ((algt->type ^ CRYPTO_ALG_TYPE_AEAD) & algt->mask) return -EINVAL; - mask = crypto_requires_sync(algt->type, algt->mask); + mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask); inst = kzalloc(sizeof(*inst) + sizeof(*ctx), GFP_KERNEL); if (!inst) @@ -636,7 +637,7 @@ static int crypto_gcm_create_common(stru goto err_free_inst; inst->alg.base.cra_flags = (ghash->base.cra_flags | - ctr->base.cra_flags) & CRYPTO_ALG_ASYNC; + ctr->base.cra_flags) & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = (ghash->base.cra_priority + ctr->base.cra_priority) / 2; inst->alg.base.cra_blocksize = 1; @@ -849,7 +850,8 @@ static int crypto_rfc4106_create(struct if ((algt->type ^ CRYPTO_ALG_TYPE_AEAD) & algt->mask) return -EINVAL; - mask = crypto_requires_sync(algt->type, algt->mask); + mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask); inst = kzalloc(sizeof(*inst) + sizeof(*spawn), GFP_KERNEL); if (!inst) @@ -882,7 +884,7 @@ static int crypto_rfc4106_create(struct CRYPTO_MAX_ALG_NAME) goto err_free_inst; - inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_ASYNC; + inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = alg->base.cra_priority; inst->alg.base.cra_blocksize = 1; inst->alg.base.cra_alignmask = alg->base.cra_alignmask; @@ -1071,7 +1073,8 @@ static int crypto_rfc4543_create(struct if ((algt->type ^ CRYPTO_ALG_TYPE_AEAD) & algt->mask) return -EINVAL; - mask = crypto_requires_sync(algt->type, algt->mask); + mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask); inst = kzalloc(sizeof(*inst) + sizeof(*ctx), GFP_KERNEL); if (!inst) @@ -1104,7 +1107,7 @@ static int crypto_rfc4543_create(struct CRYPTO_MAX_ALG_NAME) goto err_free_inst; - inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_ASYNC; + inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = alg->base.cra_priority; inst->alg.base.cra_blocksize = 1; inst->alg.base.cra_alignmask = alg->base.cra_alignmask; Index: linux-2.6/crypto/geniv.c =================================================================== --- linux-2.6.orig/crypto/geniv.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/geniv.c 2020-06-26 17:24:03.566417000 +0200 @@ -64,6 +64,7 @@ struct aead_instance *aead_geniv_alloc(s /* Ignore async algorithms if necessary. */ mask |= crypto_requires_sync(algt->type, algt->mask); + mask |= crypto_requires_nomem(algt->type, algt->mask); err = crypto_grab_aead(spawn, aead_crypto_instance(inst), crypto_attr_alg_name(tb[1]), type, mask); @@ -89,7 +90,7 @@ struct aead_instance *aead_geniv_alloc(s CRYPTO_MAX_ALG_NAME) goto err_free_inst; - inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_ASYNC; + inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = alg->base.cra_priority; inst->alg.base.cra_blocksize = alg->base.cra_blocksize; inst->alg.base.cra_alignmask = alg->base.cra_alignmask; Index: linux-2.6/crypto/lrw.c =================================================================== --- linux-2.6.orig/crypto/lrw.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/lrw.c 2020-06-26 17:24:03.566417000 +0200 @@ -311,7 +311,8 @@ static int create(struct crypto_template if ((algt->type ^ CRYPTO_ALG_TYPE_SKCIPHER) & algt->mask) return -EINVAL; - mask = crypto_requires_sync(algt->type, algt->mask); + mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask); cipher_name = crypto_attr_alg_name(tb[1]); if (IS_ERR(cipher_name)) @@ -379,7 +380,7 @@ static int create(struct crypto_template } else goto err_free_inst; - inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_ASYNC; + inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = alg->base.cra_priority; inst->alg.base.cra_blocksize = LRW_BLOCK_SIZE; inst->alg.base.cra_alignmask = alg->base.cra_alignmask | Index: linux-2.6/crypto/rsa-pkcs1pad.c =================================================================== --- linux-2.6.orig/crypto/rsa-pkcs1pad.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/rsa-pkcs1pad.c 2020-06-26 17:24:03.566417000 +0200 @@ -611,7 +611,8 @@ static int pkcs1pad_create(struct crypto if ((algt->type ^ CRYPTO_ALG_TYPE_AKCIPHER) & algt->mask) return -EINVAL; - mask = crypto_requires_sync(algt->type, algt->mask); + mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask); inst = kzalloc(sizeof(*inst) + sizeof(*ctx), GFP_KERNEL); if (!inst) @@ -658,7 +659,7 @@ static int pkcs1pad_create(struct crypto goto err_free_inst; } - inst->alg.base.cra_flags = rsa_alg->base.cra_flags & CRYPTO_ALG_ASYNC; + inst->alg.base.cra_flags = rsa_alg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = rsa_alg->base.cra_priority; inst->alg.base.cra_ctxsize = sizeof(struct pkcs1pad_ctx); Index: linux-2.6/crypto/xts.c =================================================================== --- linux-2.6.orig/crypto/xts.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/xts.c 2020-06-26 17:24:03.566417000 +0200 @@ -415,7 +415,7 @@ static int create(struct crypto_template } else goto err_free_inst; - inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_ASYNC; + inst->alg.base.cra_flags = alg->base.cra_flags & CRYPTO_ALG_INHERITED_FLAGS; inst->alg.base.cra_priority = alg->base.cra_priority; inst->alg.base.cra_blocksize = XTS_BLOCK_SIZE; inst->alg.base.cra_alignmask = alg->base.cra_alignmask | Index: linux-2.6/include/crypto/algapi.h =================================================================== --- linux-2.6.orig/include/crypto/algapi.h 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/include/crypto/algapi.h 2020-06-26 17:24:03.566417000 +0200 @@ -249,6 +249,15 @@ static inline int crypto_requires_sync(u return crypto_requires_off(type, mask, CRYPTO_ALG_ASYNC); } +/* + * Returns CRYPTO_ALG_ALLOCATES_MEMORY if type/mask requires the use of drivers + * that don't allocate memory. Otherwise returns zero. + */ +static inline int crypto_requires_nomem(u32 type, u32 mask) +{ + return crypto_requires_off(type, mask, CRYPTO_ALG_ALLOCATES_MEMORY); +} + noinline unsigned long __crypto_memneq(const void *a, const void *b, size_t size); /** Index: linux-2.6/crypto/adiantum.c =================================================================== --- linux-2.6.orig/crypto/adiantum.c 2020-06-26 17:24:03.566417000 +0200 +++ linux-2.6/crypto/adiantum.c 2020-06-26 17:24:03.566417000 +0200 @@ -507,7 +507,8 @@ static int adiantum_create(struct crypto if ((algt->type ^ CRYPTO_ALG_TYPE_SKCIPHER) & algt->mask) return -EINVAL; - mask = crypto_requires_sync(algt->type, algt->mask); + mask = crypto_requires_sync(algt->type, algt->mask) | + crypto_requires_nomem(algt->type, algt->mask); inst = kzalloc(sizeof(*inst) + sizeof(*ictx), GFP_KERNEL); if (!inst)