|
Message-ID: <CAJHCu1JARUv+8QU1Vy38Cswovvi1cAE_jYQm+QAic9rqhgw_bg@mail.gmail.com> Date: Wed, 14 Mar 2018 20:25:17 +0100 From: Salvatore Mesoraca <s.mesoraca16@...il.com> To: Eric Biggers <ebiggers3@...il.com> Cc: linux-kernel@...r.kernel.org, Kernel Hardening <kernel-hardening@...ts.openwall.com>, linux-crypto@...r.kernel.org, "David S. Miller" <davem@...emloft.net>, Herbert Xu <herbert@...dor.apana.org.au>, Kees Cook <keescook@...omium.org> Subject: Re: [PATCH] crypto: ctr: avoid VLA use 2018-03-14 19:31 GMT+01:00 Eric Biggers <ebiggers3@...il.com>: > On Wed, Mar 14, 2018 at 02:17:30PM +0100, Salvatore Mesoraca wrote: >> All ciphers implemented in Linux have a block size less than or >> equal to 16 bytes and the most demanding hw require 16 bits >> alignment for the block buffer. >> We avoid 2 VLAs[1] by always allocating 16 bytes with 16 bits >> alignment, unless the architecture support efficient unaligned >> accesses. >> We also check, at runtime, that our assumptions still stand, >> possibly dynamically allocating a new buffer, just in case >> something changes in the future. >> >> [1] https://lkml.org/lkml/2018/3/7/621 >> >> Signed-off-by: Salvatore Mesoraca <s.mesoraca16@...il.com> >> --- >> >> Notes: >> Can we maybe skip the runtime check? >> >> crypto/ctr.c | 50 ++++++++++++++++++++++++++++++++++++++++++-------- >> 1 file changed, 42 insertions(+), 8 deletions(-) >> >> diff --git a/crypto/ctr.c b/crypto/ctr.c >> index 854d924..f37adf0 100644 >> --- a/crypto/ctr.c >> +++ b/crypto/ctr.c >> @@ -35,6 +35,16 @@ struct crypto_rfc3686_req_ctx { >> struct skcipher_request subreq CRYPTO_MINALIGN_ATTR; >> }; >> >> +#ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS >> +#define DECLARE_CIPHER_BUFFER(name) u8 name[16] >> +#else >> +#define DECLARE_CIPHER_BUFFER(name) u8 __aligned(16) name[16] >> +#endif >> + >> +#define CHECK_CIPHER_BUFFER(name, size, align) \ >> + likely(size <= sizeof(name) && \ >> + name == PTR_ALIGN(((u8 *) name), align + 1)) >> + >> static int crypto_ctr_setkey(struct crypto_tfm *parent, const u8 *key, >> unsigned int keylen) >> { >> @@ -52,22 +62,35 @@ static int crypto_ctr_setkey(struct crypto_tfm *parent, const u8 *key, >> return err; >> } >> >> -static void crypto_ctr_crypt_final(struct blkcipher_walk *walk, >> - struct crypto_cipher *tfm) >> +static int crypto_ctr_crypt_final(struct blkcipher_walk *walk, >> + struct crypto_cipher *tfm) >> { >> unsigned int bsize = crypto_cipher_blocksize(tfm); >> unsigned long alignmask = crypto_cipher_alignmask(tfm); >> u8 *ctrblk = walk->iv; >> - u8 tmp[bsize + alignmask]; >> - u8 *keystream = PTR_ALIGN(tmp + 0, alignmask + 1); >> u8 *src = walk->src.virt.addr; >> u8 *dst = walk->dst.virt.addr; >> unsigned int nbytes = walk->nbytes; >> + DECLARE_CIPHER_BUFFER(tmp); >> + u8 *keystream, *tmp2; >> + >> + if (CHECK_CIPHER_BUFFER(tmp, bsize, alignmask)) >> + keystream = tmp; >> + else { >> + tmp2 = kmalloc(bsize + alignmask, GFP_ATOMIC); >> + if (!tmp2) >> + return -ENOMEM; >> + keystream = PTR_ALIGN(tmp2 + 0, alignmask + 1); >> + } >> >> crypto_cipher_encrypt_one(tfm, keystream, ctrblk); >> crypto_xor_cpy(dst, keystream, src, nbytes); >> >> crypto_inc(ctrblk, bsize); >> + >> + if (unlikely(keystream != tmp)) >> + kfree(tmp2); >> + return 0; >> } > > This seems silly; isn't the !CHECK_CIPHER_BUFFER() case unreachable? Did you > even test it? If there's going to be limits, the crypto API ought to enforce > them when registering an algorithm. Yes, as I wrote in the commit log, I put that code just in case something changes (i.e. someone adds a cipher with a bigger block size), so that it won't fail but just work as is. Although I didn't really like it, hence the note. > A better alternative may be to move the keystream buffer into the request > context, which is allowed to be variable length. It looks like that would > require converting the ctr template over to the skcipher API, since the > blkcipher API doesn't have a request context. But my understanding is that that > will need to be done eventually anyway, since the blkcipher (and ablkcipher) API > is going away. I converted a bunch of algorithms recently and I can look at the > remaining ones in crypto/*.c if no one else gets to it first, but it may be a > little while until I have time. This seems much better. I don't think that removing these VLAs is urgent, after all their sizes are limited and not under user control: we can just wait. I might help porting some crypto/*.c to skcipher API. > Also, I recall there being a long discussion a while back about how > __aligned(16) doesn't work on local variables because the kernel's stack pointer > isn't guaranteed to maintain the alignment assumed by the compiler (see commit > b8fbe71f7535)... Oh... didn't know this! Interesting... Thank you for your time, Salvatore
Powered by blists - more mailing lists
Confused about mailing lists and their use? Read about mailing lists on Wikipedia and check out these guidelines on proper formatting of your messages.