Replace system call thr_getscheduler, thr_setscheduler, thr_setschedparam
with rtprio_thread, while rtprio system call is for process only, the new system call rtprio_thread is responsible for LWP.
This commit is contained in:
parent
7614e6f347
commit
73fa3e5b88
Notes:
svn2git
2020-12-20 02:59:44 +00:00
svn path=/head/; revision=162497
@ -272,6 +272,101 @@ donice(struct thread *td, struct proc *p, int n)
|
|||||||
return (0);
|
return (0);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Set realtime priority for LWP.
|
||||||
|
*
|
||||||
|
* MPSAFE
|
||||||
|
*/
|
||||||
|
#ifndef _SYS_SYSPROTO_H_
|
||||||
|
struct rtprio_thread_args {
|
||||||
|
int function;
|
||||||
|
lwpid_t lwpid;
|
||||||
|
struct rtprio *rtp;
|
||||||
|
};
|
||||||
|
#endif
|
||||||
|
|
||||||
|
int
|
||||||
|
rtprio_thread(struct thread *td, struct rtprio_thread_args *uap)
|
||||||
|
{
|
||||||
|
struct proc *curp;
|
||||||
|
struct proc *p;
|
||||||
|
struct rtprio rtp;
|
||||||
|
struct thread *td1;
|
||||||
|
int cierror, error;
|
||||||
|
|
||||||
|
/* Perform copyin before acquiring locks if needed. */
|
||||||
|
if (uap->function == RTP_SET)
|
||||||
|
cierror = copyin(uap->rtp, &rtp, sizeof(struct rtprio));
|
||||||
|
else
|
||||||
|
cierror = 0;
|
||||||
|
|
||||||
|
curp = td->td_proc;
|
||||||
|
/*
|
||||||
|
* Though lwpid is unique, only current process is supported
|
||||||
|
* since there is no efficient way to look up a LWP yet.
|
||||||
|
*/
|
||||||
|
p = curp;
|
||||||
|
PROC_LOCK(p);
|
||||||
|
|
||||||
|
switch (uap->function) {
|
||||||
|
case RTP_LOOKUP:
|
||||||
|
if ((error = p_cansee(td, p)))
|
||||||
|
break;
|
||||||
|
mtx_lock_spin(&sched_lock);
|
||||||
|
if (uap->lwpid == 0 || uap->lwpid == td->td_tid)
|
||||||
|
td1 = td;
|
||||||
|
else
|
||||||
|
td1 = thread_find(p, uap->lwpid);
|
||||||
|
if (td1 != NULL)
|
||||||
|
pri_to_rtp(td1->td_ksegrp, &rtp);
|
||||||
|
else
|
||||||
|
error = ESRCH;
|
||||||
|
mtx_unlock_spin(&sched_lock);
|
||||||
|
PROC_UNLOCK(p);
|
||||||
|
return (copyout(&rtp, uap->rtp, sizeof(struct rtprio)));
|
||||||
|
case RTP_SET:
|
||||||
|
if ((error = p_cansched(td, p)) || (error = cierror))
|
||||||
|
break;
|
||||||
|
|
||||||
|
/* Disallow setting rtprio in most cases if not superuser. */
|
||||||
|
if (suser(td) != 0) {
|
||||||
|
/* can't set realtime priority */
|
||||||
|
/*
|
||||||
|
* Realtime priority has to be restricted for reasons which should be
|
||||||
|
* obvious. However, for idle priority, there is a potential for
|
||||||
|
* system deadlock if an idleprio process gains a lock on a resource
|
||||||
|
* that other processes need (and the idleprio process can't run
|
||||||
|
* due to a CPU-bound normal process). Fix me! XXX
|
||||||
|
*/
|
||||||
|
#if 0
|
||||||
|
if (RTP_PRIO_IS_REALTIME(rtp.type)) {
|
||||||
|
#else
|
||||||
|
if (rtp.type != RTP_PRIO_NORMAL) {
|
||||||
|
#endif
|
||||||
|
error = EPERM;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
mtx_lock_spin(&sched_lock);
|
||||||
|
if (uap->lwpid == 0 || uap->lwpid == td->td_tid)
|
||||||
|
td1 = td;
|
||||||
|
else
|
||||||
|
td1 = thread_find(p, uap->lwpid);
|
||||||
|
if (td1 != NULL)
|
||||||
|
error = rtp_to_pri(&rtp, td1->td_ksegrp);
|
||||||
|
else
|
||||||
|
error = ESRCH;
|
||||||
|
mtx_unlock_spin(&sched_lock);
|
||||||
|
break;
|
||||||
|
default:
|
||||||
|
error = EINVAL;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
PROC_UNLOCK(p);
|
||||||
|
return (error);
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Set realtime priority.
|
* Set realtime priority.
|
||||||
*
|
*
|
||||||
|
@ -58,7 +58,7 @@ static int create_thread(struct thread *td, mcontext_t *ctx,
|
|||||||
char *stack_base, size_t stack_size,
|
char *stack_base, size_t stack_size,
|
||||||
char *tls_base,
|
char *tls_base,
|
||||||
long *child_tid, long *parent_tid,
|
long *child_tid, long *parent_tid,
|
||||||
int flags, struct thr_sched_param *sched);
|
int flags, struct rtprio *rtp);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* System call interface.
|
* System call interface.
|
||||||
@ -83,29 +83,23 @@ thr_new(struct thread *td, struct thr_new_args *uap)
|
|||||||
/* struct thr_param * */
|
/* struct thr_param * */
|
||||||
{
|
{
|
||||||
struct thr_param param;
|
struct thr_param param;
|
||||||
struct thr_sched_param sched_param, *sched;
|
struct rtprio rtp, *rtpp;
|
||||||
int error;
|
int error;
|
||||||
|
|
||||||
if (uap->param_size < sizeof(param))
|
if (uap->param_size < sizeof(param))
|
||||||
return (EINVAL);
|
return (EINVAL);
|
||||||
|
bzero(¶m, sizeof(param));
|
||||||
if ((error = copyin(uap->param, ¶m, sizeof(param))))
|
if ((error = copyin(uap->param, ¶m, sizeof(param))))
|
||||||
return (error);
|
return (error);
|
||||||
sched = NULL;
|
rtpp = NULL;
|
||||||
if (param.sched_param != NULL) {
|
if (param.rtp != 0) {
|
||||||
if (param.sched_param_size != sizeof(struct thr_sched_param))
|
error = copyin(param.rtp, &rtp, sizeof(struct rtprio));
|
||||||
return (EINVAL);
|
rtpp = &rtp;
|
||||||
|
|
||||||
error = copyin(param.sched_param, &sched_param,
|
|
||||||
sizeof(sched_param));
|
|
||||||
if (error)
|
|
||||||
return (error);
|
|
||||||
sched = &sched_param;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
error = create_thread(td, NULL, param.start_func, param.arg,
|
error = create_thread(td, NULL, param.start_func, param.arg,
|
||||||
param.stack_base, param.stack_size, param.tls_base,
|
param.stack_base, param.stack_size, param.tls_base,
|
||||||
param.child_tid, param.parent_tid, param.flags,
|
param.child_tid, param.parent_tid, param.flags,
|
||||||
sched);
|
rtpp);
|
||||||
return (error);
|
return (error);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -115,7 +109,7 @@ create_thread(struct thread *td, mcontext_t *ctx,
|
|||||||
char *stack_base, size_t stack_size,
|
char *stack_base, size_t stack_size,
|
||||||
char *tls_base,
|
char *tls_base,
|
||||||
long *child_tid, long *parent_tid,
|
long *child_tid, long *parent_tid,
|
||||||
int flags, struct thr_sched_param *sched)
|
int flags, struct rtprio *rtp)
|
||||||
{
|
{
|
||||||
stack_t stack;
|
stack_t stack;
|
||||||
struct thread *newtd;
|
struct thread *newtd;
|
||||||
@ -132,18 +126,18 @@ create_thread(struct thread *td, mcontext_t *ctx,
|
|||||||
if (p->p_numthreads >= max_threads_per_proc)
|
if (p->p_numthreads >= max_threads_per_proc)
|
||||||
return (EPROCLIM);
|
return (EPROCLIM);
|
||||||
|
|
||||||
if (sched != NULL) {
|
if (rtp != NULL) {
|
||||||
switch(sched->policy) {
|
switch(rtp->type) {
|
||||||
case SCHED_FIFO:
|
case RTP_PRIO_REALTIME:
|
||||||
case SCHED_RR:
|
case RTP_PRIO_FIFO:
|
||||||
/* Only root can set scheduler policy */
|
/* Only root can set scheduler policy */
|
||||||
if (suser(td) != 0)
|
if (suser(td) != 0)
|
||||||
return (EPERM);
|
return (EPERM);
|
||||||
if (sched->param.sched_priority < RTP_PRIO_MIN ||
|
if (rtp->prio > RTP_PRIO_MAX)
|
||||||
sched->param.sched_priority > RTP_PRIO_MAX)
|
|
||||||
return (EINVAL);
|
return (EINVAL);
|
||||||
break;
|
break;
|
||||||
case SCHED_OTHER:
|
case RTP_PRIO_NORMAL:
|
||||||
|
rtp->prio = 0;
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
return (EINVAL);
|
return (EINVAL);
|
||||||
@ -218,32 +212,12 @@ create_thread(struct thread *td, mcontext_t *ctx,
|
|||||||
/* let the scheduler know about these things. */
|
/* let the scheduler know about these things. */
|
||||||
sched_fork_ksegrp(td, newkg);
|
sched_fork_ksegrp(td, newkg);
|
||||||
sched_fork_thread(td, newtd);
|
sched_fork_thread(td, newtd);
|
||||||
if (sched != NULL) {
|
if (rtp != NULL) {
|
||||||
struct rtprio rtp;
|
if (!(kg->kg_pri_class == PRI_TIMESHARE &&
|
||||||
switch (sched->policy) {
|
rtp->type == RTP_PRIO_NORMAL)) {
|
||||||
case SCHED_FIFO:
|
rtp_to_pri(rtp, newkg);
|
||||||
rtp.type = PRI_FIFO;
|
|
||||||
rtp.prio = RTP_PRIO_MAX - sched->param.sched_priority;
|
|
||||||
rtp_to_pri(&rtp, newkg);
|
|
||||||
sched_prio(newtd, newkg->kg_user_pri);
|
sched_prio(newtd, newkg->kg_user_pri);
|
||||||
break;
|
} /* ignore timesharing class */
|
||||||
case SCHED_RR:
|
|
||||||
rtp.type = PRI_REALTIME;
|
|
||||||
rtp.prio = RTP_PRIO_MAX - sched->param.sched_priority;
|
|
||||||
rtp_to_pri(&rtp, newkg);
|
|
||||||
sched_prio(newtd, newkg->kg_user_pri);
|
|
||||||
break;
|
|
||||||
case SCHED_OTHER:
|
|
||||||
if (newkg->kg_pri_class != PRI_TIMESHARE) {
|
|
||||||
rtp.type = PRI_TIMESHARE;
|
|
||||||
rtp.prio = 0;
|
|
||||||
rtp_to_pri(&rtp, newkg);
|
|
||||||
sched_prio(newtd, newkg->kg_user_pri);
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
default:
|
|
||||||
panic("sched policy");
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
TD_SET_CAN_RUN(newtd);
|
TD_SET_CAN_RUN(newtd);
|
||||||
/* if ((flags & THR_SUSPENDED) == 0) */
|
/* if ((flags & THR_SUSPENDED) == 0) */
|
||||||
@ -436,154 +410,3 @@ thr_set_name(struct thread *td, struct thr_set_name_args *uap)
|
|||||||
PROC_UNLOCK(p);
|
PROC_UNLOCK(p);
|
||||||
return (error);
|
return (error);
|
||||||
}
|
}
|
||||||
|
|
||||||
int
|
|
||||||
thr_setscheduler(struct thread *td, struct thr_setscheduler_args *uap)
|
|
||||||
{
|
|
||||||
struct proc *p;
|
|
||||||
struct thread *ttd;
|
|
||||||
struct rtprio rtp;
|
|
||||||
struct sched_param param;
|
|
||||||
int ret;
|
|
||||||
|
|
||||||
if (uap->param_size != sizeof(struct sched_param))
|
|
||||||
return (EINVAL);
|
|
||||||
|
|
||||||
ret = copyin(uap->param, ¶m, sizeof(struct sched_param));
|
|
||||||
if (ret != 0)
|
|
||||||
return (ret);
|
|
||||||
|
|
||||||
ret = suser(td);
|
|
||||||
if (ret != 0)
|
|
||||||
return (ret);
|
|
||||||
|
|
||||||
switch(uap->policy) {
|
|
||||||
case SCHED_FIFO:
|
|
||||||
rtp.type = PRI_FIFO;
|
|
||||||
rtp.prio = RTP_PRIO_MAX - param.sched_priority;
|
|
||||||
break;
|
|
||||||
case SCHED_RR:
|
|
||||||
rtp.type = PRI_REALTIME;
|
|
||||||
rtp.prio = RTP_PRIO_MAX - param.sched_priority;
|
|
||||||
break;
|
|
||||||
case SCHED_OTHER:
|
|
||||||
rtp.type = PRI_TIMESHARE;
|
|
||||||
rtp.prio = 0;
|
|
||||||
break;
|
|
||||||
default:
|
|
||||||
return (EINVAL);
|
|
||||||
}
|
|
||||||
|
|
||||||
p = td->td_proc;
|
|
||||||
PROC_LOCK(p);
|
|
||||||
if (ret != 0) {
|
|
||||||
PROC_UNLOCK(p);
|
|
||||||
return (ret);
|
|
||||||
}
|
|
||||||
|
|
||||||
ttd = thread_find(p, uap->id);
|
|
||||||
if (ttd == NULL) {
|
|
||||||
PROC_UNLOCK(p);
|
|
||||||
return (ESRCH);
|
|
||||||
}
|
|
||||||
mtx_lock_spin(&sched_lock);
|
|
||||||
ret = rtp_to_pri(&rtp, ttd->td_ksegrp);
|
|
||||||
if (ret == 0)
|
|
||||||
ttd->td_flags |= TDF_NEEDRESCHED;
|
|
||||||
mtx_unlock_spin(&sched_lock);
|
|
||||||
PROC_UNLOCK(p);
|
|
||||||
return (ret);
|
|
||||||
}
|
|
||||||
|
|
||||||
int
|
|
||||||
thr_getscheduler(struct thread *td, struct thr_getscheduler_args *uap)
|
|
||||||
{
|
|
||||||
struct proc *p;
|
|
||||||
struct thread *ttd;
|
|
||||||
struct rtprio rtp;
|
|
||||||
struct sched_param param;
|
|
||||||
int policy;
|
|
||||||
int ret;
|
|
||||||
|
|
||||||
if (uap->param_size != sizeof(struct sched_param))
|
|
||||||
return (EINVAL);
|
|
||||||
|
|
||||||
p = td->td_proc;
|
|
||||||
PROC_LOCK(p);
|
|
||||||
ttd = thread_find(p, uap->id);
|
|
||||||
if (ttd == NULL) {
|
|
||||||
PROC_UNLOCK(p);
|
|
||||||
return (ESRCH);
|
|
||||||
}
|
|
||||||
mtx_lock_spin(&sched_lock);
|
|
||||||
pri_to_rtp(ttd->td_ksegrp, &rtp);
|
|
||||||
switch(ttd->td_ksegrp->kg_pri_class) {
|
|
||||||
case PRI_FIFO:
|
|
||||||
policy = SCHED_FIFO;
|
|
||||||
param.sched_priority = RTP_PRIO_MAX - rtp.prio;
|
|
||||||
break;
|
|
||||||
case PRI_REALTIME:
|
|
||||||
policy = SCHED_RR;
|
|
||||||
param.sched_priority = RTP_PRIO_MAX - rtp.prio;
|
|
||||||
break;
|
|
||||||
case PRI_TIMESHARE:
|
|
||||||
default: /* XXX SCHED_IDLE */
|
|
||||||
policy = SCHED_OTHER;
|
|
||||||
param.sched_priority = 0;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
mtx_unlock_spin(&sched_lock);
|
|
||||||
PROC_UNLOCK(p);
|
|
||||||
|
|
||||||
ret = copyout(&policy, uap->policy, sizeof(policy));
|
|
||||||
if (ret == 0)
|
|
||||||
ret = copyout(¶m, uap->param, sizeof(param));
|
|
||||||
return (ret);
|
|
||||||
}
|
|
||||||
|
|
||||||
int
|
|
||||||
thr_setschedparam(struct thread *td, struct thr_setschedparam_args *uap)
|
|
||||||
{
|
|
||||||
struct proc *p;
|
|
||||||
struct thread *ttd;
|
|
||||||
struct rtprio rtp;
|
|
||||||
struct sched_param param;
|
|
||||||
int ret;
|
|
||||||
|
|
||||||
if (uap->param_size != sizeof(struct sched_param))
|
|
||||||
return (EINVAL);
|
|
||||||
|
|
||||||
ret = copyin(uap->param, ¶m, sizeof(struct sched_param));
|
|
||||||
if (ret != 0)
|
|
||||||
return (ret);
|
|
||||||
ret = suser(td);
|
|
||||||
if (ret != 0)
|
|
||||||
return (ret);
|
|
||||||
p = td->td_proc;
|
|
||||||
PROC_LOCK(p);
|
|
||||||
ttd = thread_find(p, uap->id);
|
|
||||||
if (ttd == NULL) {
|
|
||||||
PROC_UNLOCK(p);
|
|
||||||
return (ESRCH);
|
|
||||||
}
|
|
||||||
mtx_lock_spin(&sched_lock);
|
|
||||||
switch(ttd->td_ksegrp->kg_pri_class) {
|
|
||||||
case PRI_FIFO:
|
|
||||||
rtp.prio = RTP_PRIO_MAX - param.sched_priority;
|
|
||||||
break;
|
|
||||||
case PRI_REALTIME:
|
|
||||||
rtp.prio = RTP_PRIO_MAX - param.sched_priority;
|
|
||||||
break;
|
|
||||||
case PRI_TIMESHARE:
|
|
||||||
rtp.prio = 0;
|
|
||||||
break;
|
|
||||||
default:
|
|
||||||
return (EINVAL);
|
|
||||||
}
|
|
||||||
ret = rtp_to_pri(&rtp, ttd->td_ksegrp);
|
|
||||||
if (ret == 0)
|
|
||||||
ttd->td_flags |= TDF_NEEDRESCHED;
|
|
||||||
mtx_unlock_spin(&sched_lock);
|
|
||||||
PROC_UNLOCK(p);
|
|
||||||
return (ret);
|
|
||||||
}
|
|
||||||
|
@ -819,15 +819,10 @@
|
|||||||
463 AUE_NULL STD { int abort2(const char *why, int nargs, void **args); }
|
463 AUE_NULL STD { int abort2(const char *why, int nargs, void **args); }
|
||||||
464 AUE_NULL STD { int thr_set_name(long id, const char *name); }
|
464 AUE_NULL STD { int thr_set_name(long id, const char *name); }
|
||||||
465 AUE_NULL NOSTD { int aio_fsync(int op, struct aiocb *aiocbp); }
|
465 AUE_NULL NOSTD { int aio_fsync(int op, struct aiocb *aiocbp); }
|
||||||
466 AUE_NULL STD { int thr_setscheduler(long id, int policy,\
|
466 AUE_RTPRIO STD { int rtprio_thread(int function, \
|
||||||
const struct sched_param *param, \
|
lwpid_t lwpid, struct rtprio *rtp); }
|
||||||
int param_size); }
|
467 AUE_NULL UNIMPL nosys
|
||||||
467 AUE_NULL STD { int thr_getscheduler(long id, int *policy,\
|
468 AUE_NULL UNIMPL nosys
|
||||||
struct sched_param *param, \
|
|
||||||
int param_size); }
|
|
||||||
468 AUE_NULL STD { int thr_setschedparam(long id, \
|
|
||||||
const struct sched_param *param, \
|
|
||||||
int param_size); }
|
|
||||||
469 AUE_NULL UNIMPL __getpath_fromfd
|
469 AUE_NULL UNIMPL __getpath_fromfd
|
||||||
470 AUE_NULL UNIMPL __getpath_fromaddr
|
470 AUE_NULL UNIMPL __getpath_fromaddr
|
||||||
; Please copy any additions and changes to the following compatability tables:
|
; Please copy any additions and changes to the following compatability tables:
|
||||||
|
@ -86,6 +86,7 @@ void pri_to_rtp(struct ksegrp *, struct rtprio *);
|
|||||||
|
|
||||||
__BEGIN_DECLS
|
__BEGIN_DECLS
|
||||||
int rtprio(int, pid_t, struct rtprio *);
|
int rtprio(int, pid_t, struct rtprio *);
|
||||||
|
int rtprio_thread(int, lwpid_t, struct rtprio *);
|
||||||
__END_DECLS
|
__END_DECLS
|
||||||
#endif /* !_KERNEL */
|
#endif /* !_KERNEL */
|
||||||
#endif /* !_SYS_RTPRIO_H_ */
|
#endif /* !_SYS_RTPRIO_H_ */
|
||||||
|
@ -32,11 +32,6 @@
|
|||||||
|
|
||||||
#include <posix4/sched.h>
|
#include <posix4/sched.h>
|
||||||
|
|
||||||
struct thr_sched_param {
|
|
||||||
int policy;
|
|
||||||
struct sched_param param;
|
|
||||||
};
|
|
||||||
|
|
||||||
/* Create the thread in the suspended state. */
|
/* Create the thread in the suspended state. */
|
||||||
#define THR_SUSPENDED 0x0001
|
#define THR_SUSPENDED 0x0001
|
||||||
/* Create the system scope thread. */
|
/* Create the system scope thread. */
|
||||||
@ -52,9 +47,8 @@ struct thr_param {
|
|||||||
long *child_tid; /* address to store new TID. */
|
long *child_tid; /* address to store new TID. */
|
||||||
long *parent_tid; /* parent accesses the new TID here. */
|
long *parent_tid; /* parent accesses the new TID here. */
|
||||||
int flags; /* thread flags. */
|
int flags; /* thread flags. */
|
||||||
struct thr_sched_param *sched_param; /* POSIX scheduler parameters .*/
|
struct rtprio *rtp; /* Real-time scheduling priority */
|
||||||
long sched_param_size; /* scheduler parameter size */
|
void *spare[3]; /* TODO: cpu affinity mask etc. */
|
||||||
void *spare[2]; /* TODO: cpu affinity mask etc. */
|
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -70,12 +64,6 @@ int thr_kill(long id, int sig);
|
|||||||
int thr_suspend(const struct timespec *timeout);
|
int thr_suspend(const struct timespec *timeout);
|
||||||
int thr_wake(long id);
|
int thr_wake(long id);
|
||||||
int thr_set_name(long id, const char *name);
|
int thr_set_name(long id, const char *name);
|
||||||
int thr_setscheduler(long id, int policy, const struct sched_param *param,
|
|
||||||
int param_size);
|
|
||||||
int thr_getscheduler(long id, int *policy, struct sched_param *param,
|
|
||||||
int param_size);
|
|
||||||
int thr_setschedparam(long id, const struct sched_param *param,
|
|
||||||
int param_size);
|
|
||||||
#endif /* !_KERNEL */
|
#endif /* !_KERNEL */
|
||||||
|
|
||||||
#endif /* ! _SYS_THR_H_ */
|
#endif /* ! _SYS_THR_H_ */
|
||||||
|
Loading…
Reference in New Issue
Block a user