|
Message-ID: <20180304034704.GB20725@bombadil.infradead.org> Date: Sat, 3 Mar 2018 19:47:04 -0800 From: Matthew Wilcox <willy@...radead.org> To: Daniel Micay <danielmicay@...il.com> Cc: Ilya Smith <blackzert@...il.com>, Kees Cook <keescook@...omium.org>, Andrew Morton <akpm@...ux-foundation.org>, Dan Williams <dan.j.williams@...el.com>, Michal Hocko <mhocko@...e.com>, "Kirill A. Shutemov" <kirill.shutemov@...ux.intel.com>, Jan Kara <jack@...e.cz>, Jerome Glisse <jglisse@...hat.com>, Hugh Dickins <hughd@...gle.com>, Helge Deller <deller@....de>, Andrea Arcangeli <aarcange@...hat.com>, Oleg Nesterov <oleg@...hat.com>, Linux-MM <linux-mm@...ck.org>, LKML <linux-kernel@...r.kernel.org>, Kernel Hardening <kernel-hardening@...ts.openwall.com> Subject: Re: [RFC PATCH] Randomization of address chosen by mmap. On Sat, Mar 03, 2018 at 04:00:45PM -0500, Daniel Micay wrote: > The main thing I'd like to see is just the option to get a guarantee > of enforced gaps around mappings, without necessarily even having > randomization of the gap size. It's possible to add guard pages in > userspace but it adds overhead by doubling the number of system calls > to map memory (mmap PROT_NONE region, mprotect the inner portion to > PROT_READ|PROT_WRITE) and *everything* using mmap would need to > cooperate which is unrealistic. So something like this? To use it, OR in PROT_GUARD(n) to the PROT flags of mmap, and it should pad the map by n pages. I haven't tested it, so I'm sure it's buggy, but it seems like a fairly cheap way to give us padding after every mapping. Running it on an old kernel will result in no padding, so to see if it worked or not, try mapping something immediately after it. diff --git a/include/linux/mm.h b/include/linux/mm.h index 4ef7fb1726ab..9da6df7f62fc 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h @@ -2183,8 +2183,8 @@ extern int install_special_mapping(struct mm_struct *mm, extern unsigned long get_unmapped_area(struct file *, unsigned long, unsigned long, unsigned long, unsigned long); extern unsigned long mmap_region(struct file *file, unsigned long addr, - unsigned long len, vm_flags_t vm_flags, unsigned long pgoff, - struct list_head *uf); + unsigned long len, unsigned long pad_len, vm_flags_t vm_flags, + unsigned long pgoff, struct list_head *uf); extern unsigned long do_mmap(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, vm_flags_t vm_flags, unsigned long pgoff, unsigned long *populate, diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h index 1c5dea402501..9c2b66fa0561 100644 --- a/include/linux/mm_types.h +++ b/include/linux/mm_types.h @@ -299,6 +299,7 @@ struct vm_area_struct { struct mm_struct *vm_mm; /* The address space we belong to. */ pgprot_t vm_page_prot; /* Access permissions of this VMA. */ unsigned long vm_flags; /* Flags, see mm.h. */ + unsigned int vm_guard; /* Number of trailing guard pages */ /* * For areas with an address space and backing store, diff --git a/include/uapi/asm-generic/mman-common.h b/include/uapi/asm-generic/mman-common.h index f8b134f5608f..d88babdf97f9 100644 --- a/include/uapi/asm-generic/mman-common.h +++ b/include/uapi/asm-generic/mman-common.h @@ -12,6 +12,7 @@ #define PROT_EXEC 0x4 /* page can be executed */ #define PROT_SEM 0x8 /* page may be used for atomic ops */ #define PROT_NONE 0x0 /* page can not be accessed */ +#define PROT_GUARD(x) ((x) & 0xffff) << 4 /* guard pages */ #define PROT_GROWSDOWN 0x01000000 /* mprotect flag: extend change to start of growsdown vma */ #define PROT_GROWSUP 0x02000000 /* mprotect flag: extend change to end of growsup vma */ diff --git a/mm/memory.c b/mm/memory.c index 1cfc4699db42..5b0f87afa0af 100644 --- a/mm/memory.c +++ b/mm/memory.c @@ -4125,6 +4125,9 @@ int handle_mm_fault(struct vm_area_struct *vma, unsigned long address, flags & FAULT_FLAG_REMOTE)) return VM_FAULT_SIGSEGV; + if (DIV_ROUND_UP(vma->vm_end - address, PAGE_SIZE) < vma->vm_guard) + return VM_FAULT_SIGSEGV; + /* * Enable the memcg OOM handling for faults triggered in user * space. Kernel faults are handled more gracefully. diff --git a/mm/mmap.c b/mm/mmap.c index 575766ec02f8..b9844b810ee7 100644 --- a/mm/mmap.c +++ b/mm/mmap.c @@ -1433,6 +1433,7 @@ unsigned long do_mmap(struct file *file, unsigned long addr, unsigned long pgoff, unsigned long *populate, struct list_head *uf) { + unsigned int guard_len = ((prot >> 4) & 0xffff) << PAGE_SHIFT; struct mm_struct *mm = current->mm; int pkey = 0; @@ -1458,6 +1459,8 @@ unsigned long do_mmap(struct file *file, unsigned long addr, len = PAGE_ALIGN(len); if (!len) return -ENOMEM; + if (len + guard_len < len) + return -ENOMEM; /* offset overflow? */ if ((pgoff + (len >> PAGE_SHIFT)) < pgoff) @@ -1472,7 +1475,7 @@ unsigned long do_mmap(struct file *file, unsigned long addr, /* Obtain the address to map to. we verify (or select) it and ensure * that it represents a valid section of the address space. */ - addr = get_unmapped_area(file, addr, len, pgoff, flags); + addr = get_unmapped_area(file, addr, len + guard_len, pgoff, flags); if (offset_in_page(addr)) return addr; @@ -1591,7 +1594,7 @@ unsigned long do_mmap(struct file *file, unsigned long addr, vm_flags |= VM_NORESERVE; } - addr = mmap_region(file, addr, len, vm_flags, pgoff, uf); + addr = mmap_region(file, addr, len, len + guard_len, vm_flags, pgoff, uf); if (!IS_ERR_VALUE(addr) && ((vm_flags & VM_LOCKED) || (flags & (MAP_POPULATE | MAP_NONBLOCK)) == MAP_POPULATE)) @@ -1727,8 +1730,8 @@ static inline int accountable_mapping(struct file *file, vm_flags_t vm_flags) } unsigned long mmap_region(struct file *file, unsigned long addr, - unsigned long len, vm_flags_t vm_flags, unsigned long pgoff, - struct list_head *uf) + unsigned long len, unsigned long pad_len, vm_flags_t vm_flags, + unsigned long pgoff, struct list_head *uf) { struct mm_struct *mm = current->mm; struct vm_area_struct *vma, *prev; @@ -1737,24 +1740,24 @@ unsigned long mmap_region(struct file *file, unsigned long addr, unsigned long charged = 0; /* Check against address space limit. */ - if (!may_expand_vm(mm, vm_flags, len >> PAGE_SHIFT)) { + if (!may_expand_vm(mm, vm_flags, pad_len >> PAGE_SHIFT)) { unsigned long nr_pages; /* * MAP_FIXED may remove pages of mappings that intersects with * requested mapping. Account for the pages it would unmap. */ - nr_pages = count_vma_pages_range(mm, addr, addr + len); + nr_pages = count_vma_pages_range(mm, addr, addr + pad_len); if (!may_expand_vm(mm, vm_flags, - (len >> PAGE_SHIFT) - nr_pages)) + (pad_len >> PAGE_SHIFT) - nr_pages)) return -ENOMEM; } /* Clear old maps */ - while (find_vma_links(mm, addr, addr + len, &prev, &rb_link, + while (find_vma_links(mm, addr, addr + pad_len, &prev, &rb_link, &rb_parent)) { - if (do_munmap(mm, addr, len, uf)) + if (do_munmap(mm, addr, pad_len, uf)) return -ENOMEM; } @@ -1771,7 +1774,7 @@ unsigned long mmap_region(struct file *file, unsigned long addr, /* * Can we just expand an old mapping? */ - vma = vma_merge(mm, prev, addr, addr + len, vm_flags, + vma = vma_merge(mm, prev, addr, addr + pad_len, vm_flags, NULL, file, pgoff, NULL, NULL_VM_UFFD_CTX); if (vma) goto out; @@ -1789,9 +1792,10 @@ unsigned long mmap_region(struct file *file, unsigned long addr, vma->vm_mm = mm; vma->vm_start = addr; - vma->vm_end = addr + len; + vma->vm_end = addr + pad_len; vma->vm_flags = vm_flags; vma->vm_page_prot = vm_get_page_prot(vm_flags); + vma->vm_guard = (pad_len - len) >> PAGE_SHIFT; vma->vm_pgoff = pgoff; INIT_LIST_HEAD(&vma->anon_vma_chain);
Powered by blists - more mailing lists
Confused about mailing lists and their use? Read about mailing lists on Wikipedia and check out these guidelines on proper formatting of your messages.