Fix threaded process job control bug. SMP tested.

Reviewed by: julian
This commit is contained in:
davidxu 2003-03-11 00:07:53 +00:00
parent 29e1315705
commit bb4f70ad77
5 changed files with 68 additions and 84 deletions

View File

@ -732,7 +732,8 @@ loop:
mtx_unlock(&Giant);
return (0);
}
if (P_SHOULDSTOP(p) && ((p->p_flag & P_WAITED) == 0) &&
if (P_SHOULDSTOP(p) && (p->p_suspcount == p->p_numthreads) &&
((p->p_flag & P_WAITED) == 0) &&
(p->p_flag & P_TRACED || uap->options & WUNTRACED)) {
p->p_flag |= P_WAITED;
sx_xunlock(&proctree_lock);

View File

@ -452,6 +452,7 @@ kse_exit(struct thread *td, struct kse_exit_args *uap)
kse_purge_group(td);
ke->ke_flags |= KEF_EXIT;
}
thread_stopped(p);
thread_exit();
/* NOTREACHED */
}
@ -1512,6 +1513,7 @@ thread_user_enter(struct proc *p, struct thread *td)
if ((p->p_flag & P_SINGLE_EXIT) && (p->p_singlethread != td)) {
PROC_LOCK(p);
mtx_lock_spin(&sched_lock);
thread_stopped(p);
thread_exit();
/* NOTREACHED */
}
@ -1627,21 +1629,17 @@ thread_userret(struct thread *td, struct trapframe *frame)
mtx_unlock_spin(&sched_lock);
} else if (td->td_mailbox) {
error = thread_export_context(td);
if (error) {
PROC_LOCK(td->td_proc);
mtx_lock_spin(&sched_lock);
/* possibly upcall with error? */
} else {
PROC_LOCK(td->td_proc);
mtx_lock_spin(&sched_lock);
/*
* There are upcall threads waiting for
* work to do, wake one of them up.
* XXXKSE Maybe wake all of them up.
*/
if (kg->kg_upsleeps)
wakeup_one(&kg->kg_completed);
}
/* possibly upcall with error? */
PROC_LOCK(p);
/*
* There are upcall threads waiting for
* work to do, wake one of them up.
* XXXKSE Maybe wake all of them up.
*/
if (!error && kg->kg_upsleeps)
wakeup_one(&kg->kg_completed);
mtx_lock_spin(&sched_lock);
thread_stopped(p);
thread_exit();
/* NOTREACHED */
}
@ -1918,13 +1916,14 @@ thread_suspend_check(int return_instead)
if (return_instead)
return (1);
mtx_lock_spin(&sched_lock);
thread_stopped(p);
/*
* If the process is waiting for us to exit,
* this thread should just suicide.
* Assumes that P_SINGLE_EXIT implies P_STOPPED_SINGLE.
*/
if ((p->p_flag & P_SINGLE_EXIT) && (p->p_singlethread != td)) {
mtx_lock_spin(&sched_lock);
while (mtx_owned(&Giant))
mtx_unlock(&Giant);
thread_exit();
@ -1935,18 +1934,6 @@ thread_suspend_check(int return_instead)
* moves to the processes's suspend queue
* and stays there.
*/
mtx_lock_spin(&sched_lock);
if ((p->p_flag & P_STOPPED_SIG) &&
(p->p_suspcount+1 == p->p_numthreads)) {
mtx_unlock_spin(&sched_lock);
PROC_LOCK(p->p_pptr);
if ((p->p_pptr->p_procsig->ps_flag &
PS_NOCLDSTOP) == 0) {
psignal(p->p_pptr, SIGCHLD);
}
PROC_UNLOCK(p->p_pptr);
mtx_lock_spin(&sched_lock);
}
mtx_assert(&Giant, MA_NOTOWNED);
thread_suspend_one(td);
PROC_UNLOCK(p);
@ -1969,6 +1956,7 @@ thread_suspend_one(struct thread *td)
struct proc *p = td->td_proc;
mtx_assert(&sched_lock, MA_OWNED);
KASSERT(!TD_IS_SUSPENDED(td), ("already suspended"));
p->p_suspcount++;
TD_SET_SUSPENDED(td);
TAILQ_INSERT_TAIL(&p->p_suspended, td, td_runq);

View File

@ -1495,26 +1495,18 @@ psignal(p, sig)
if (prop & SA_STOP) {
if (p->p_flag & P_PPWAIT)
goto out;
p->p_flag |= P_STOPPED_SIG;
p->p_xstat = sig;
mtx_lock_spin(&sched_lock);
FOREACH_THREAD_IN_PROC(p, td) {
if (TD_IS_SLEEPING(td) &&
(td->td_flags & TDF_SINTR))
thread_suspend_one(td);
}
if (p->p_suspcount == p->p_numthreads) {
mtx_unlock_spin(&sched_lock);
stop(p);
p->p_xstat = sig;
SIGDELSET(p->p_siglist, sig);
PROC_LOCK(p->p_pptr);
if ((p->p_pptr->p_procsig->ps_flag &
PS_NOCLDSTOP) == 0) {
psignal(p->p_pptr, SIGCHLD);
}
PROC_UNLOCK(p->p_pptr);
} else {
mtx_unlock_spin(&sched_lock);
}
thread_stopped(p);
if (p->p_numthreads == p->p_suspcount)
SIGDELSET(p->p_siglist, p->p_xstat);
mtx_unlock_spin(&sched_lock);
goto out;
}
else
@ -1754,19 +1746,10 @@ issignal(td)
(p->p_pgrp->pg_jobc == 0 &&
prop & SA_TTYSTOP))
break; /* == ignore */
p->p_flag |= P_STOPPED_SIG;
p->p_xstat = sig;
mtx_lock_spin(&sched_lock);
if (p->p_suspcount+1 == p->p_numthreads) {
mtx_unlock_spin(&sched_lock);
PROC_LOCK(p->p_pptr);
if ((p->p_pptr->p_procsig->ps_flag &
PS_NOCLDSTOP) == 0) {
psignal(p->p_pptr, SIGCHLD);
}
PROC_UNLOCK(p->p_pptr);
mtx_lock_spin(&sched_lock);
}
stop(p);
thread_stopped(p);
thread_suspend_one(td);
PROC_UNLOCK(p);
DROP_GIANT();
@ -1816,8 +1799,7 @@ issignal(td)
* lock held.
*/
static void
stop(p)
register struct proc *p;
stop(struct proc *p)
{
PROC_LOCK_ASSERT(p, MA_OWNED);
@ -1826,6 +1808,30 @@ stop(p)
wakeup(p->p_pptr);
}
void
thread_stopped(struct proc *p)
{
struct proc *p1 = curthread->td_proc;
int n;
PROC_LOCK_ASSERT(p, MA_OWNED);
mtx_assert(&sched_lock, MA_OWNED);
n = p->p_suspcount;
if (p == p1)
n++;
if ((p->p_flag & P_STOPPED_SIG) && (n == p->p_numthreads)) {
mtx_unlock_spin(&sched_lock);
stop(p);
PROC_LOCK(p->p_pptr);
if ((p->p_pptr->p_procsig->ps_flag &
PS_NOCLDSTOP) == 0) {
psignal(p->p_pptr, SIGCHLD);
}
PROC_UNLOCK(p->p_pptr);
mtx_lock_spin(&sched_lock);
}
}
/*
* Take the action for the specified signal
* from the current set of pending signals.

View File

@ -452,6 +452,7 @@ kse_exit(struct thread *td, struct kse_exit_args *uap)
kse_purge_group(td);
ke->ke_flags |= KEF_EXIT;
}
thread_stopped(p);
thread_exit();
/* NOTREACHED */
}
@ -1512,6 +1513,7 @@ thread_user_enter(struct proc *p, struct thread *td)
if ((p->p_flag & P_SINGLE_EXIT) && (p->p_singlethread != td)) {
PROC_LOCK(p);
mtx_lock_spin(&sched_lock);
thread_stopped(p);
thread_exit();
/* NOTREACHED */
}
@ -1627,21 +1629,17 @@ thread_userret(struct thread *td, struct trapframe *frame)
mtx_unlock_spin(&sched_lock);
} else if (td->td_mailbox) {
error = thread_export_context(td);
if (error) {
PROC_LOCK(td->td_proc);
mtx_lock_spin(&sched_lock);
/* possibly upcall with error? */
} else {
PROC_LOCK(td->td_proc);
mtx_lock_spin(&sched_lock);
/*
* There are upcall threads waiting for
* work to do, wake one of them up.
* XXXKSE Maybe wake all of them up.
*/
if (kg->kg_upsleeps)
wakeup_one(&kg->kg_completed);
}
/* possibly upcall with error? */
PROC_LOCK(p);
/*
* There are upcall threads waiting for
* work to do, wake one of them up.
* XXXKSE Maybe wake all of them up.
*/
if (!error && kg->kg_upsleeps)
wakeup_one(&kg->kg_completed);
mtx_lock_spin(&sched_lock);
thread_stopped(p);
thread_exit();
/* NOTREACHED */
}
@ -1918,13 +1916,14 @@ thread_suspend_check(int return_instead)
if (return_instead)
return (1);
mtx_lock_spin(&sched_lock);
thread_stopped(p);
/*
* If the process is waiting for us to exit,
* this thread should just suicide.
* Assumes that P_SINGLE_EXIT implies P_STOPPED_SINGLE.
*/
if ((p->p_flag & P_SINGLE_EXIT) && (p->p_singlethread != td)) {
mtx_lock_spin(&sched_lock);
while (mtx_owned(&Giant))
mtx_unlock(&Giant);
thread_exit();
@ -1935,18 +1934,6 @@ thread_suspend_check(int return_instead)
* moves to the processes's suspend queue
* and stays there.
*/
mtx_lock_spin(&sched_lock);
if ((p->p_flag & P_STOPPED_SIG) &&
(p->p_suspcount+1 == p->p_numthreads)) {
mtx_unlock_spin(&sched_lock);
PROC_LOCK(p->p_pptr);
if ((p->p_pptr->p_procsig->ps_flag &
PS_NOCLDSTOP) == 0) {
psignal(p->p_pptr, SIGCHLD);
}
PROC_UNLOCK(p->p_pptr);
mtx_lock_spin(&sched_lock);
}
mtx_assert(&Giant, MA_NOTOWNED);
thread_suspend_one(td);
PROC_UNLOCK(p);
@ -1969,6 +1956,7 @@ thread_suspend_one(struct thread *td)
struct proc *p = td->td_proc;
mtx_assert(&sched_lock, MA_OWNED);
KASSERT(!TD_IS_SUSPENDED(td), ("already suspended"));
p->p_suspcount++;
TD_SET_SUSPENDED(td);
TAILQ_INSERT_TAIL(&p->p_suspended, td, td_runq);

View File

@ -945,6 +945,7 @@ void upcall_unlink(struct kse_upcall *ku);
void upcall_remove(struct thread *td);
void upcall_stash(struct kse_upcall *ke);
void thread_sanity_check(struct thread *td, char *);
void thread_stopped(struct proc *);
#endif /* _KERNEL */
#endif /* !_SYS_PROC_H_ */