extern void exit_files(struct task_struct *);
extern void __cleanup_signal(struct signal_struct *);
+extern void cleanup_sighand(struct task_struct *);
extern void __exit_signal(struct task_struct *);
-extern void __exit_sighand(struct task_struct *);
extern void exit_itimers(struct signal_struct *);
extern NORET_TYPE void do_group_exit(int);
return 0;
}
+void cleanup_sighand(struct task_struct *tsk)
+{
+ struct sighand_struct * sighand = tsk->sighand;
+
+ /* Ok, we're done with the signal handlers */
+ tsk->sighand = NULL;
+ if (atomic_dec_and_test(&sighand->count))
+ kmem_cache_free(sighand_cachep, sighand);
+}
+
static inline int copy_signal(unsigned long clone_flags, struct task_struct * tsk)
{
struct signal_struct *sig;
bad_fork_cleanup_signal:
cleanup_signal(p);
bad_fork_cleanup_sighand:
- __exit_sighand(p);
+ cleanup_sighand(p);
bad_fork_cleanup_fs:
exit_fs(p); /* blocking */
bad_fork_cleanup_files:
/*
* Flush all pending signals for a task.
*/
-
-void
-flush_signals(struct task_struct *t)
+void flush_signals(struct task_struct *t)
{
unsigned long flags;
spin_unlock_irqrestore(&t->sighand->siglock, flags);
}
-/*
- * This function expects the tasklist_lock write-locked.
- */
-void __exit_sighand(struct task_struct *tsk)
-{
- struct sighand_struct * sighand = tsk->sighand;
-
- /* Ok, we're done with the signal handlers */
- tsk->sighand = NULL;
- if (atomic_dec_and_test(&sighand->count))
- kmem_cache_free(sighand_cachep, sighand);
-}
-
/*
* This function expects the tasklist_lock write-locked.
*/
}
tsk->signal = NULL;
- __exit_sighand(tsk);
+ cleanup_sighand(tsk);
spin_unlock(&sighand->siglock);
rcu_read_unlock();