|
Message-ID: <CAG48ez1TNd5eJCE=x6LeqGO3x2E1rn244xqc2ccnZ3wmeP05XA@mail.gmail.com> Date: Fri, 12 Oct 2018 00:11:19 +0200 From: Jann Horn <jannh@...gle.com> To: Andy Lutomirski <luto@...nel.org> Cc: Kees Cook <keescook@...omium.org>, kristen@...ux.intel.com, Kernel Hardening <kernel-hardening@...ts.openwall.com>, Thomas Gleixner <tglx@...utronix.de>, Ingo Molnar <mingo@...hat.com>, Borislav Petkov <bp@...en8.de>, "H . Peter Anvin" <hpa@...or.com>, "the arch/x86 maintainers" <x86@...nel.org>, kernel list <linux-kernel@...r.kernel.org> Subject: Re: [PATCH] x86: entry: flush the cache if syscall error On Thu, Oct 11, 2018 at 11:17 PM Andy Lutomirski <luto@...nel.org> wrote: > On Thu, Oct 11, 2018 at 1:55 PM Kees Cook <keescook@...omium.org> wrote: > > On Thu, Oct 11, 2018 at 1:48 PM, Andy Lutomirski <luto@...nel.org> wrote: > > > On Thu, Oct 11, 2018 at 11:55 AM Kristen Carlson Accardi > > > <kristen@...ux.intel.com> wrote: > > >> > > >> This patch aims to make it harder to perform cache timing attacks on data > > >> left behind by system calls. If we have an error returned from a syscall, > > >> flush the L1 cache. > > >> > > >> It's important to note that this patch is not addressing any specific > > >> exploit, nor is it intended to be a complete defense against anything. > > >> It is intended to be a low cost way of eliminating some of side effects > > >> of a failed system call. > > >> > > >> A performance test using sysbench on one hyperthread and a script which > > >> attempts to repeatedly access files it does not have permission to access > > >> on the other hyperthread found no significant performance impact. > > >> > > >> Suggested-by: Alan Cox <alan@...ux.intel.com> > > >> Signed-off-by: Kristen Carlson Accardi <kristen@...ux.intel.com> > > >> --- > > >> arch/x86/Kconfig | 9 +++++++++ > > >> arch/x86/entry/common.c | 18 ++++++++++++++++++ > > >> 2 files changed, 27 insertions(+) > > >> > > >> diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig > > >> index 1a0be022f91d..bde978eb3b4e 100644 > > >> --- a/arch/x86/Kconfig > > >> +++ b/arch/x86/Kconfig > > >> @@ -445,6 +445,15 @@ config RETPOLINE > > >> code are eliminated. Since this includes the syscall entry path, > > >> it is not entirely pointless. > > >> > > >> +config SYSCALL_FLUSH > > >> + bool "Clear L1 Cache on syscall errors" > > >> + default n > > >> + help > > >> + Selecting 'y' allows the L1 cache to be cleared upon return of > > >> + an error code from a syscall if the CPU supports "flush_l1d". > > >> + This may reduce the likelyhood of speculative execution style > > >> + attacks on syscalls. > > >> + > > >> config INTEL_RDT > > >> bool "Intel Resource Director Technology support" > > >> default n > > >> diff --git a/arch/x86/entry/common.c b/arch/x86/entry/common.c > > >> index 3b2490b81918..26de8ea71293 100644 > > >> --- a/arch/x86/entry/common.c > > >> +++ b/arch/x86/entry/common.c > > >> @@ -268,6 +268,20 @@ __visible inline void syscall_return_slowpath(struct pt_regs *regs) > > >> prepare_exit_to_usermode(regs); > > >> } > > >> > > >> +__visible inline void l1_cache_flush(struct pt_regs *regs) > > >> +{ > > >> + if (IS_ENABLED(CONFIG_SYSCALL_FLUSH) && > > >> + static_cpu_has(X86_FEATURE_FLUSH_L1D)) { > > >> + if (regs->ax == 0 || regs->ax == -EAGAIN || > > >> + regs->ax == -EEXIST || regs->ax == -ENOENT || > > >> + regs->ax == -EXDEV || regs->ax == -ETIMEDOUT || > > >> + regs->ax == -ENOTCONN || regs->ax == -EINPROGRESS) > > > > > > What about ax > 0? (Or more generally, any ax outside the range of -1 > > > .. -4095 or whatever the error range is.) As it stands, it looks like > > > you'll flush on successful read(), write(), recv(), etc, and that > > > could seriously hurt performance on real workloads. > > > > Seems like just changing this with "ax == 0" into "ax >= 0" would solve that? > > I can easily imagine that there are other errors for which performance > matters. EBUSY comes to mind. > > > > > I think this looks like a good idea. It might be worth adding a > > comment about the checks to explain why those errors are whitelisted. > > It's a cheap and effective mitigation for "unknown future problems" > > that doesn't degrade normal workloads. > > I still want to see two pieces of information before I ack a patch like this: > > - How long does L1D_FLUSH take, roughly? (Especially if L1D is dirty and you can't just wipe it all.) > - An example of a type of attack that would be mitigated. > > For the latter, I assume that the goal is to mitigate against attacks > where a syscall speculatively loads something sensitive and then > fails. But, before it fails, it leaks the information it > speculatively loaded, and that leak ended up in L1D but *not* in other > cache levels. And somehow the L1D can't be probed quickly enough in a > parallel thread to meaningfully get the information out of L1D. And the attacker can't delay the syscall return somehow, e.g. on a fully-preemptible kernel by preempting the syscall, or by tarpitting a userspace memory access in the error path. > Or > maybe it's trying to mitigate against the use of failing syscalls to > get some sensitive data into L1D and then using something like L1TF to > read it out. > > But this really needs to be clarified. Alan said that a bunch of the > "yet another Spectre variant" attacks would have been mitigated by > this patch. An explanation of *how* would be in order. > > And we should seriously consider putting this kind of thing under > CONFIG_SPECULATIVE_MITIGATIONS or similar. The idea being that it > doesn't mitigate a clear known attack family.
Powered by blists - more mailing lists
Confused about mailing lists and their use? Read about mailing lists on Wikipedia and check out these guidelines on proper formatting of your messages.