diff options
author | faux123 <reioux@gmail.com> | 2011-12-24 23:48:01 -0800 |
---|---|---|
committer | Ziyan <jaraidaniel@gmail.com> | 2016-01-08 10:43:04 +0100 |
commit | 2c38a3c551c2e6d4aa62974a56819d66382e0d27 (patch) | |
tree | 96a86666609ae3b547cf07d743303145b5558e5f /include | |
parent | b0c6b2899990b48bd431464080beb8e36215b1bc (diff) | |
download | kernel_samsung_tuna-2c38a3c551c2e6d4aa62974a56819d66382e0d27.zip kernel_samsung_tuna-2c38a3c551c2e6d4aa62974a56819d66382e0d27.tar.gz kernel_samsung_tuna-2c38a3c551c2e6d4aa62974a56819d66382e0d27.tar.bz2 |
threadgroup: extend threadgroup_lock() to cover exit and exec
From: Tejun Heo <tj@kernel.org>
threadgroup_lock() protected only protected against new addition to
the threadgroup, which was inherently somewhat incomplete and
problematic for its only user cgroup. On-going migration could race
against exec and exit leading to interesting problems - the symmetry
between various attach methods, task exiting during method execution,
->exit() racing against attach methods, migrating task switching basic
properties during exec and so on.
This patch extends threadgroup_lock() such that it protects against
all three threadgroup altering operations - fork, exit and exec. For
exit, threadgroup_change_begin/end() calls are added to exit path.
For exec, threadgroup_[un]lock() are updated to also grab and release
cred_guard_mutex.
With this change, threadgroup_lock() guarantees that the target
threadgroup will remain stable - no new task will be added, no new
PF_EXITING will be set and exec won't happen.
The next patch will update cgroup so that it can take full advantage
of this change.
Signed-off-by: Tejun Heo <tj@kernel.org>
Cc: Oleg Nesterov <oleg@redhat.com>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Paul Menage <menage@google.com>
Cc: Li Zefan <lizf@cn.fujitsu.com>
backported by faux123
Conflicts:
kernel/exit.c
Change-Id: Id975b5cbacb5d8db993e43688a722a28225dc557
Diffstat (limited to 'include')
-rw-r--r-- | include/linux/sched.h | 32 |
1 files changed, 26 insertions, 6 deletions
diff --git a/include/linux/sched.h b/include/linux/sched.h index 20a73ad..13f4339 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h @@ -635,11 +635,12 @@ struct signal_struct { #endif #ifdef CONFIG_CGROUPS /* - * The group_rwsem prevents threads from forking with - * CLONE_THREAD while held for writing. Use this for fork-sensitive - * threadgroup-wide operations. It's taken for reading in fork.c in - * copy_process(). - * Currently only needed write-side by cgroups. + * group_rwsem prevents new tasks from entering the threadgroup and + * member tasks from exiting. fork and exit paths are protected + * with this rwsem using threadgroup_change_begin/end(). Users + * which require threadgroup to remain stable should use + * threadgroup_[un]lock() which also takes care of exec path. + * Currently, cgroup is the only user. */ struct rw_semaphore group_rwsem; #endif @@ -2354,7 +2355,6 @@ static inline void unlock_task_sighand(struct task_struct *tsk, spin_unlock_irqrestore(&tsk->sighand->siglock, *flags); } -/* See the declaration of group_rwsem in signal_struct. */ #ifdef CONFIG_CGROUPS static inline void threadgroup_change_begin(struct task_struct *tsk) { @@ -2364,13 +2364,33 @@ static inline void threadgroup_change_done(struct task_struct *tsk) { up_read(&tsk->signal->group_rwsem); } + +/** + * threadgroup_lock - lock threadgroup + * @tsk: member task of the threadgroup to lock + * + * Lock the threadgroup @tsk belongs to. No new task is allowed to enter + * and member tasks aren't allowed to exit (as indicated by PF_EXITING) or + * perform exec. This is useful for cases where the threadgroup needs to + * stay stable across blockable operations. + */ static inline void threadgroup_lock(struct task_struct *tsk) { + /* exec uses exit for de-threading, grab cred_guard_mutex first */ + mutex_lock(&tsk->signal->cred_guard_mutex); down_write(&tsk->signal->group_rwsem); } + +/** + * threadgroup_unlock - unlock threadgroup + * @tsk: member task of the threadgroup to unlock + * + * Reverse threadgroup_lock(). + */ static inline void threadgroup_unlock(struct task_struct *tsk) { up_write(&tsk->signal->group_rwsem); + mutex_unlock(&tsk->signal->cred_guard_mutex); } #else static inline void threadgroup_change_begin(struct task_struct *tsk) {} |