|
Message-Id: <1497915397-93805-15-git-send-email-keescook@chromium.org> Date: Mon, 19 Jun 2017 16:36:28 -0700 From: Kees Cook <keescook@...omium.org> To: kernel-hardening@...ts.openwall.com Cc: Kees Cook <keescook@...omium.org>, David Windsor <dave@...lcore.net>, linux-mm@...ck.org, linux-kernel@...r.kernel.org Subject: [PATCH 14/23] fork: define usercopy region in thread_stack, task_struct, mm_struct slab caches From: David Windsor <dave@...lcore.net> In support of usercopy hardening, this patch defines a region in the thread_stack, task_struct and mm_struct slab caches in which userspace copy operations are allowed. Since only a single whitelisted buffer region is used, this moves the usercopyable fields next to each other in task_struct so a single window can cover them. This region is known as the slab cache's usercopy region. Slab caches can now check that each copy operation involving cache-managed memory falls entirely within the slab's usercopy region. This patch is modified from Brad Spengler/PaX Team's PAX_USERCOPY whitelisting code in the last public patch of grsecurity/PaX based on my understanding of the code. Changes or omissions from the original code are mine and don't reflect the original grsecurity/PaX code. Signed-off-by: David Windsor <dave@...lcore.net> [kees: adjust commit log] Signed-off-by: Kees Cook <keescook@...omium.org> --- include/linux/sched.h | 15 ++++++++++++--- kernel/fork.c | 18 +++++++++++++----- 2 files changed, 25 insertions(+), 8 deletions(-) diff --git a/include/linux/sched.h b/include/linux/sched.h index 2b69fc650201..345db7983af1 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h @@ -745,10 +745,19 @@ struct task_struct { /* Signal handlers: */ struct signal_struct *signal; struct sighand_struct *sighand; - sigset_t blocked; sigset_t real_blocked; - /* Restored if set_restore_sigmask() was used: */ - sigset_t saved_sigmask; + + /* + * Usercopy slab whitelisting needs blocked, saved_sigmask + * to be adjacent. + */ + struct { + sigset_t blocked; + + /* Restored if set_restore_sigmask() was used */ + sigset_t saved_sigmask; + }; + struct sigpending pending; unsigned long sas_ss_sp; size_t sas_ss_size; diff --git a/kernel/fork.c b/kernel/fork.c index e53770d2bf95..172df19baeb5 100644 --- a/kernel/fork.c +++ b/kernel/fork.c @@ -282,8 +282,9 @@ static void free_thread_stack(struct task_struct *tsk) void thread_stack_cache_init(void) { - thread_stack_cache = kmem_cache_create("thread_stack", THREAD_SIZE, - THREAD_SIZE, 0, NULL); + thread_stack_cache = kmem_cache_create_usercopy("thread_stack", + THREAD_SIZE, THREAD_SIZE, 0, 0, + THREAD_SIZE, NULL); BUG_ON(thread_stack_cache == NULL); } # endif @@ -467,9 +468,14 @@ void __init fork_init(void) int align = max_t(int, L1_CACHE_BYTES, ARCH_MIN_TASKALIGN); /* create a slab on which task_structs can be allocated */ - task_struct_cachep = kmem_cache_create("task_struct", + task_struct_cachep = kmem_cache_create_usercopy("task_struct", arch_task_struct_size, align, - SLAB_PANIC|SLAB_NOTRACK|SLAB_ACCOUNT, NULL); + SLAB_PANIC|SLAB_NOTRACK|SLAB_ACCOUNT, + offsetof(struct task_struct, blocked), + offsetof(struct task_struct, saved_sigmask) - + offsetof(struct task_struct, blocked) + + sizeof(init_task.saved_sigmask), + NULL); #endif /* do the arch specific task caches init */ @@ -2208,9 +2214,11 @@ void __init proc_caches_init(void) * maximum number of CPU's we can ever have. The cpumask_allocation * is at the end of the structure, exactly for that reason. */ - mm_cachep = kmem_cache_create("mm_struct", + mm_cachep = kmem_cache_create_usercopy("mm_struct", sizeof(struct mm_struct), ARCH_MIN_MMSTRUCT_ALIGN, SLAB_HWCACHE_ALIGN|SLAB_PANIC|SLAB_NOTRACK|SLAB_ACCOUNT, + offsetof(struct mm_struct, saved_auxv), + sizeof_field(struct mm_struct, saved_auxv), NULL); vm_area_cachep = KMEM_CACHE(vm_area_struct, SLAB_PANIC|SLAB_ACCOUNT); mmap_init(); -- 2.7.4
Powered by blists - more mailing lists
Confused about mailing lists and their use? Read about mailing lists on Wikipedia and check out these guidelines on proper formatting of your messages.