|
Message-ID: <1538760003.3560.136.camel@intel.com> Date: Fri, 5 Oct 2018 17:18:09 +0000 From: "Edgecombe, Rick P" <rick.p.edgecombe@...el.com> To: "linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>, "daniel@...earbox.net" <daniel@...earbox.net>, "jannh@...gle.com" <jannh@...gle.com>, "keescook@...omium.org" <keescook@...omium.org>, "willy@...radead.org" <willy@...radead.org>, "tglx@...utronix.de" <tglx@...utronix.de>, "linux-mm@...ck.org" <linux-mm@...ck.org>, "x86@...nel.org" <x86@...nel.org>, "akpm@...ux-foundation.org" <akpm@...ux-foundation.org>, "hpa@...or.com" <hpa@...or.com>, "mingo@...hat.com" <mingo@...hat.com>, "kernel-hardening@...ts.openwall.com" <kernel-hardening@...ts.openwall.com> CC: "kristen@...ux.intel.com" <kristen@...ux.intel.com>, "arjan@...ux.intel.com" <arjan@...ux.intel.com>, "Hansen, Dave" <dave.hansen@...el.com> Subject: Re: [PATCH v7 1/4] vmalloc: Add __vmalloc_node_try_addr function Forgot to include this: Reviewed-by: Kees Cook <keescook@...omium.org> On Mon, 2018-10-01 at 14:38 -0700, Rick Edgecombe wrote: > Create __vmalloc_node_try_addr function that tries to allocate at a specific > address and supports caller specified behavior for whether any lazy purging > happens if there is a collision. > > This new function draws from the __vmalloc_node_range implementation. Attempts > to merge the two into a single allocator resulted in logic that was difficult > to follow, so they are left separate. > > Signed-off-by: Rick Edgecombe <rick.p.edgecombe@...el.com> > --- > include/linux/vmalloc.h | 3 + > mm/vmalloc.c | 177 > +++++++++++++++++++++++++++++++++++++++++++++++- > 2 files changed, 179 insertions(+), 1 deletion(-) > > diff --git a/include/linux/vmalloc.h b/include/linux/vmalloc.h > index 398e9c9..c7712c8 100644 > --- a/include/linux/vmalloc.h > +++ b/include/linux/vmalloc.h > @@ -82,6 +82,9 @@ extern void *__vmalloc_node_range(unsigned long size, > unsigned long align, > unsigned long start, unsigned long end, gfp_t > gfp_mask, > pgprot_t prot, unsigned long vm_flags, int node, > const void *caller); > +extern void *__vmalloc_node_try_addr(unsigned long addr, unsigned long size, > + gfp_t gfp_mask, pgprot_t prot, unsigned long > vm_flags, > + int node, int try_purge, const void *caller); > #ifndef CONFIG_MMU > extern void *__vmalloc_node_flags(unsigned long size, int node, gfp_t flags); > static inline void *__vmalloc_node_flags_caller(unsigned long size, int node, > diff --git a/mm/vmalloc.c b/mm/vmalloc.c > index a728fc4..1954458 100644 > --- a/mm/vmalloc.c > +++ b/mm/vmalloc.c > @@ -1709,6 +1709,181 @@ static void *__vmalloc_area_node(struct vm_struct > *area, gfp_t gfp_mask, > return NULL; > } > > +static bool pvm_find_next_prev(unsigned long end, > + struct vmap_area **pnext, > + struct vmap_area **pprev); > + > +/* Try to allocate a region of KVA of the specified address and size. */ > +static struct vmap_area *try_alloc_vmap_area(unsigned long addr, > + unsigned long size, int node, gfp_t gfp_mask, > + int try_purge) > +{ > + struct vmap_area *va; > + struct vmap_area *cur_va = NULL; > + struct vmap_area *first_before = NULL; > + int need_purge = 0; > + int blocked = 0; > + int purged = 0; > + unsigned long addr_end; > + > + WARN_ON(!size); > + WARN_ON(offset_in_page(size)); > + > + addr_end = addr + size; > + if (addr > addr_end) > + return ERR_PTR(-EOVERFLOW); > + > + might_sleep(); > + > + va = kmalloc_node(sizeof(struct vmap_area), > + gfp_mask & GFP_RECLAIM_MASK, node); > + if (unlikely(!va)) > + return ERR_PTR(-ENOMEM); > + > + /* > + * Only scan the relevant parts containing pointers to other objects > + * to avoid false negatives. > + */ > + kmemleak_scan_area(&va->rb_node, SIZE_MAX, gfp_mask & > GFP_RECLAIM_MASK); > + > +retry: > + spin_lock(&vmap_area_lock); > + > + pvm_find_next_prev(addr, &cur_va, &first_before); > + > + if (!cur_va) > + goto found; > + > + /* > + * If there is no VA that starts before the target address, start the > + * check from the closest VA in order to cover the case where the > + * allocation overlaps at the end. > + */ > + if (first_before && addr < first_before->va_end) > + cur_va = first_before; > + > + /* Linearly search through to make sure there is a hole */ > + while (cur_va->va_start < addr_end) { > + if (cur_va->va_end > addr) { > + if (cur_va->flags & VM_LAZY_FREE) { > + need_purge = 1; > + } else { > + blocked = 1; > + break; > + } > + } > + > + if (list_is_last(&cur_va->list, &vmap_area_list)) > + break; > + > + cur_va = list_next_entry(cur_va, list); > + } > + > + /* > + * If a non-lazy free va blocks the allocation, or > + * we are not supposed to purge, but we need to, the > + * allocation fails. > + */ > + if (blocked || (need_purge && !try_purge)) > + goto fail; > + > + if (try_purge && need_purge) { > + /* if purged once before, give up */ > + if (purged) > + goto fail; > + > + /* > + * If the va blocking the allocation is set to > + * be purged then purge all vmap_areas that are > + * set to purged since this will flush the TLBs > + * anyway. > + */ > + spin_unlock(&vmap_area_lock); > + purge_vmap_area_lazy(); > + need_purge = 0; > + purged = 1; > + goto retry; > + } > + > +found: > + va->va_start = addr; > + va->va_end = addr_end; > + va->flags = 0; > + __insert_vmap_area(va); > + spin_unlock(&vmap_area_lock); > + > + return va; > +fail: > + spin_unlock(&vmap_area_lock); > + kfree(va); > + if (need_purge && !blocked) > + return ERR_PTR(-EUCLEAN); > + return ERR_PTR(-EBUSY); > +} > + > +/** > + * __vmalloc_try_addr - try to alloc at a specific address > + * @addr: address to try > + * @size: size to try > + * @gfp_mask: flags for the page level allocator > + * @prot: protection mask for the allocated pages > + * @vm_flags: additional vm area flags (e.g. %VM_NO_GUARD) > + * @node: node to use for allocation or NUMA_NO_NODE > + * @try_purge: try to purge if needed to fulfill and allocation > + * @caller: caller's return address > + * > + * Try to allocate at the specific address. If it succeeds the address > is > + * returned. If it fails an EBUSY ERR_PTR is returned. If try_purge is > + * zero, it will return an EUCLEAN ERR_PTR if it could have allocated > if it > + * was allowed to purge. It may trigger TLB flushes if a purge is > needed, > + * and try_purge is set. > + */ > +void *__vmalloc_node_try_addr(unsigned long addr, unsigned long size, > + gfp_t gfp_mask, pgprot_t prot, unsigned long > vm_flags, > + int node, int try_purge, const void *caller) > +{ > + struct vmap_area *va; > + struct vm_struct *area; > + void *alloc_addr; > + unsigned long real_size = size; > + > + size = PAGE_ALIGN(size); > + if (!size || (size >> PAGE_SHIFT) > totalram_pages) > + return NULL; > + > + WARN_ON(in_interrupt()); > + > + if (!(vm_flags & VM_NO_GUARD)) > + size += PAGE_SIZE; > + > + va = try_alloc_vmap_area(addr, size, node, gfp_mask, try_purge); > + if (IS_ERR(va)) > + goto fail; > + > + area = kzalloc_node(sizeof(*area), gfp_mask & GFP_RECLAIM_MASK, > node); > + if (unlikely(!area)) { > + warn_alloc(gfp_mask, NULL, "kmalloc: allocation failure"); > + return ERR_PTR(-ENOMEM); > + } > + > + setup_vmalloc_vm(area, va, vm_flags, caller); > + > + alloc_addr = __vmalloc_area_node(area, gfp_mask, prot, node); > + if (!alloc_addr) { > + warn_alloc(gfp_mask, NULL, > + "vmalloc: allocation failure: %lu bytes", real_size); > + return ERR_PTR(-ENOMEM); > + } > + > + clear_vm_uninitialized_flag(area); > + > + kmemleak_vmalloc(area, real_size, gfp_mask); > + > + return alloc_addr; > +fail: > + return va; > +} > + > /** > * __vmalloc_node_range - allocate virtually contiguous memory > * @size: allocation size > @@ -2355,7 +2530,6 @@ void free_vm_area(struct vm_struct *area) > } > EXPORT_SYMBOL_GPL(free_vm_area); > > -#ifdef CONFIG_SMP > static struct vmap_area *node_to_va(struct rb_node *n) > { > return rb_entry_safe(n, struct vmap_area, rb_node); > @@ -2403,6 +2577,7 @@ static bool pvm_find_next_prev(unsigned long end, > return true; > } > > +#ifdef CONFIG_SMP > /** > * pvm_determine_end - find the highest aligned address between two > vmap_areas > * @pnext: in/out arg for the next vmap_area
Powered by blists - more mailing lists
Confused about mailing lists and their use? Read about mailing lists on Wikipedia and check out these guidelines on proper formatting of your messages.