temp
This commit is contained in:
parent
23f774324a
commit
bec8de6929
@ -100,9 +100,9 @@ MTX_SYSINIT(kq_global, &kq_global, "kqueue order", MTX_DEF);
|
|||||||
TASKQUEUE_DEFINE_THREAD(kqueue_ctx);
|
TASKQUEUE_DEFINE_THREAD(kqueue_ctx);
|
||||||
|
|
||||||
static struct kevq * kevqlist_find(struct kevqlist *kevq_list, struct kqueue *kq);
|
static struct kevq * kevqlist_find(struct kevqlist *kevq_list, struct kqueue *kq);
|
||||||
static void kevq_wakeup(struct kevq* kevq);
|
|
||||||
static void kevq_thred_init(struct kevq_thred *kevq_th);
|
static void kevq_thred_init(struct kevq_thred *kevq_th);
|
||||||
static void kevq_thred_destroy(struct kevq_thred *kevq_th);
|
static void kevq_thred_destroy(struct kevq_thred *kevq_th);
|
||||||
|
static void kevq_wakeup(struct kevq* kevq);
|
||||||
static void kevq_init(struct kevq *kevq);
|
static void kevq_init(struct kevq *kevq);
|
||||||
static void kevq_release(struct kevq* kevq, int locked);
|
static void kevq_release(struct kevq* kevq, int locked);
|
||||||
static int kevq_acquire_kq(struct kqueue *kq, struct thread *td, struct kevq **kevqp);
|
static int kevq_acquire_kq(struct kqueue *kq, struct thread *td, struct kevq **kevqp);
|
||||||
@ -158,7 +158,14 @@ static struct fileops kqueueops = {
|
|||||||
.fo_fill_kinfo = kqueue_fill_kinfo,
|
.fo_fill_kinfo = kqueue_fill_kinfo,
|
||||||
};
|
};
|
||||||
|
|
||||||
static void knote_activate(struct knote *kn, int haskqlock);
|
static bool knote_leave_flux_ul(struct knote *kn);
|
||||||
|
static bool knote_leave_flux(struct knote *kn);
|
||||||
|
static void knote_enter_flux(struct knote *kn);
|
||||||
|
static void knote_enter_flux_ul(struct knote *kn);
|
||||||
|
static void knote_flux_wakeup_ul(struct knote *kn);
|
||||||
|
static void knote_flux_wakeup(struct knote *kn);
|
||||||
|
static void knote_activate_ul(struct knote *kn);
|
||||||
|
static void knote_activate(struct knote *kn);
|
||||||
static int knote_attach(struct knote *kn, struct kqueue *kq);
|
static int knote_attach(struct knote *kn, struct kqueue *kq);
|
||||||
static void knote_drop(struct knote *kn, struct thread *td);
|
static void knote_drop(struct knote *kn, struct thread *td);
|
||||||
static void knote_drop_detached(struct knote *kn, struct thread *td);
|
static void knote_drop_detached(struct knote *kn, struct thread *td);
|
||||||
@ -167,7 +174,7 @@ static void knote_dequeue(struct knote *kn);
|
|||||||
static void knote_init(void);
|
static void knote_init(void);
|
||||||
static struct knote *knote_alloc(int mflag);
|
static struct knote *knote_alloc(int mflag);
|
||||||
static void knote_free(struct knote *kn);
|
static void knote_free(struct knote *kn);
|
||||||
static struct kevq* knote_sched(struct knote *kn);
|
static void knote_sched(struct knote *kn);
|
||||||
|
|
||||||
static void filt_kqdetach(struct knote *kn);
|
static void filt_kqdetach(struct knote *kn);
|
||||||
static int filt_kqueue(struct knote *kn, long hint);
|
static int filt_kqueue(struct knote *kn, long hint);
|
||||||
@ -239,24 +246,19 @@ SYSCTL_UINT(_kern, OID_AUTO, kq_calloutmax, CTLFLAG_RW,
|
|||||||
#define KEVQ_LOCK(kevq) do { \
|
#define KEVQ_LOCK(kevq) do { \
|
||||||
mtx_lock(&(kevq)->lock); \
|
mtx_lock(&(kevq)->lock); \
|
||||||
} while (0)
|
} while (0)
|
||||||
#define KN_FLUX_WAKEUP(kn, haslock) do { \
|
|
||||||
if ((haslock)) \
|
|
||||||
KN_FLUX_OWNED((kn)); \
|
|
||||||
else \
|
|
||||||
KN_FLUX_LOCK((kn)); \
|
|
||||||
if ((kn)->kn_fluxwait) { \
|
|
||||||
(kn)->kn_fluxwait = 0; \
|
|
||||||
wakeup((kn)); \
|
|
||||||
} \
|
|
||||||
if (!(haslock)) \
|
|
||||||
KN_FLUX_UNLOCK((kn)); \
|
|
||||||
} while (0)
|
|
||||||
#define KQ_UNLOCK(kq) do { \
|
#define KQ_UNLOCK(kq) do { \
|
||||||
mtx_unlock(&(kq)->kq_lock); \
|
mtx_unlock(&(kq)->kq_lock); \
|
||||||
} while (0)
|
} while (0)
|
||||||
#define KN_FLUX_UNLOCK(kn) do { \
|
#define KN_FLUX_UNLOCK(kn) do { \
|
||||||
mtx_unlock(&(kn)->kn_fluxlock); \
|
mtx_unlock(&(kn)->kn_fluxlock); \
|
||||||
} while (0)
|
} while (0)
|
||||||
|
#define KN_LEAVE_FLUX_WAKEUP(kn) do { \
|
||||||
|
KN_FLUX_NOTOWNED((kn)); \
|
||||||
|
KN_FLUX_LOCK((kn)); \
|
||||||
|
knote_leave_flux((kn)); \
|
||||||
|
knote_flux_wakeup((kn)); \
|
||||||
|
KN_FLUX_UNLOCK((kn)); \
|
||||||
|
} while(0)
|
||||||
#define KEVQ_TH_UNLOCK(kevqth) do { \
|
#define KEVQ_TH_UNLOCK(kevqth) do { \
|
||||||
mtx_unlock(&(kevqth)->lock); \
|
mtx_unlock(&(kevqth)->lock); \
|
||||||
} while (0)
|
} while (0)
|
||||||
@ -272,7 +274,7 @@ SYSCTL_UINT(_kern, OID_AUTO, kq_calloutmax, CTLFLAG_RW,
|
|||||||
#define KN_FLUX_OWNED(kn) do { \
|
#define KN_FLUX_OWNED(kn) do { \
|
||||||
mtx_assert(&(kn)->kn_fluxlock, MA_OWNED); \
|
mtx_assert(&(kn)->kn_fluxlock, MA_OWNED); \
|
||||||
} while (0)
|
} while (0)
|
||||||
#define KN_FLUX_NOTOWNED(kq) do { \
|
#define KN_FLUX_NOTOWNED(kn) do { \
|
||||||
mtx_assert(&(kn)->kn_fluxlock, MA_NOTOWNED); \
|
mtx_assert(&(kn)->kn_fluxlock, MA_NOTOWNED); \
|
||||||
} while (0)
|
} while (0)
|
||||||
#define KEVQ_OWNED(kevq) do { \
|
#define KEVQ_OWNED(kevq) do { \
|
||||||
@ -315,46 +317,49 @@ kn_in_flux(struct knote *kn)
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void
|
static void
|
||||||
kn_enter_flux(struct knote *kn, int haslock)
|
knote_enter_flux_ul(struct knote *kn)
|
||||||
{
|
{
|
||||||
#ifdef KQ_DEBUG
|
KN_FLUX_NOTOWNED(kn);
|
||||||
printf("KQUEUE: kn_enter_flux: %p\n", kn);
|
|
||||||
#endif
|
|
||||||
if(haslock) {
|
|
||||||
KN_FLUX_OWNED(kn);
|
|
||||||
} else {
|
|
||||||
KN_FLUX_LOCK(kn);
|
KN_FLUX_LOCK(kn);
|
||||||
}
|
knote_enter_flux(kn);
|
||||||
MPASS(kn->kn_influx < INT_MAX);
|
|
||||||
kn->kn_influx++;
|
|
||||||
if(!haslock) {
|
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void
|
||||||
|
knote_enter_flux(struct knote *kn)
|
||||||
|
{
|
||||||
|
#ifdef KQ_DEBUG
|
||||||
|
printf("KQUEUE: knote_enter_flux: %p\n", kn);
|
||||||
|
#endif
|
||||||
|
KN_FLUX_OWNED(kn);
|
||||||
|
MPASS(kn->kn_influx < INT_MAX);
|
||||||
|
kn->kn_influx++;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* TODO: change *_ul functions to macros? */
|
||||||
|
static bool
|
||||||
|
knote_leave_flux_ul(struct knote *kn)
|
||||||
|
{
|
||||||
|
bool ret;
|
||||||
|
KN_FLUX_NOTOWNED(kn);
|
||||||
|
KN_FLUX_LOCK(kn);
|
||||||
|
ret = knote_leave_flux(kn);
|
||||||
|
KN_FLUX_UNLOCK(kn);
|
||||||
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool
|
static bool
|
||||||
kn_leave_flux(struct knote *kn, int haslock)
|
knote_leave_flux(struct knote *kn)
|
||||||
{
|
{
|
||||||
#ifdef KQ_DEBUG
|
#ifdef KQ_DEBUG
|
||||||
printf("KQUEUE: kn_leave_flux: %p\n", kn);
|
printf("KQUEUE: knote_leave_flux: %p\n", kn);
|
||||||
#endif
|
#endif
|
||||||
if(haslock) {
|
|
||||||
KN_FLUX_OWNED(kn);
|
KN_FLUX_OWNED(kn);
|
||||||
} else {
|
|
||||||
KN_FLUX_LOCK(kn);
|
|
||||||
}
|
|
||||||
MPASS(kn->kn_influx > 0);
|
MPASS(kn->kn_influx > 0);
|
||||||
kn->kn_influx--;
|
kn->kn_influx--;
|
||||||
|
|
||||||
if(!haslock) {
|
|
||||||
KN_FLUX_UNLOCK(kn);
|
|
||||||
// RETURN false if the caller doesn't care about the result
|
|
||||||
// otherwise there might be a race here
|
|
||||||
return false;
|
|
||||||
} else {
|
|
||||||
return (kn->kn_influx == 0);
|
return (kn->kn_influx == 0);
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
#define KNL_ASSERT_LOCK(knl, islocked) do { \
|
#define KNL_ASSERT_LOCK(knl, islocked) do { \
|
||||||
if (islocked) \
|
if (islocked) \
|
||||||
@ -535,7 +540,7 @@ filt_procattach(struct knote *kn)
|
|||||||
* is registered.
|
* is registered.
|
||||||
*/
|
*/
|
||||||
if (immediate || (exiting && filt_proc(kn, NOTE_EXIT)))
|
if (immediate || (exiting && filt_proc(kn, NOTE_EXIT)))
|
||||||
knote_activate(kn, 0);
|
knote_activate_ul(kn);
|
||||||
|
|
||||||
PROC_UNLOCK(p);
|
PROC_UNLOCK(p);
|
||||||
|
|
||||||
@ -636,21 +641,20 @@ knote_fork(struct knlist *list, struct thread *td, int pid)
|
|||||||
* track the child. Drop the locks in preparation for
|
* track the child. Drop the locks in preparation for
|
||||||
* the call to kqueue_register().
|
* the call to kqueue_register().
|
||||||
*/
|
*/
|
||||||
kn_enter_flux(kn, 1);
|
knote_enter_flux(kn);
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* The same as knote(), activate the event.
|
* The same as knote(), activate the event.
|
||||||
*/
|
*/
|
||||||
if ((kn->kn_sfflags & NOTE_TRACK) == 0) {
|
if ((kn->kn_sfflags & NOTE_TRACK) == 0) {
|
||||||
if (kn->kn_fop->f_event(kn, NOTE_FORK))
|
error = kn->kn_fop->f_event(kn, NOTE_FORK);
|
||||||
knote_activate(kn, 1);
|
|
||||||
|
|
||||||
KN_FLUX_LOCK(kn);
|
|
||||||
kn_leave_flux(kn, 1);
|
|
||||||
KN_FLUX_WAKEUP(kn, 1);
|
|
||||||
KN_FLUX_UNLOCK(kn);
|
|
||||||
KQ_UNLOCK(kq);
|
KQ_UNLOCK(kq);
|
||||||
|
|
||||||
|
if(error)
|
||||||
|
knote_activate(kn);
|
||||||
|
|
||||||
|
KN_LEAVE_FLUX_WAKEUP(kn);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -698,13 +702,10 @@ knote_fork(struct knlist *list, struct thread *td, int pid)
|
|||||||
if (error)
|
if (error)
|
||||||
kn->kn_fflags |= NOTE_TRACKERR;
|
kn->kn_fflags |= NOTE_TRACKERR;
|
||||||
if (kn->kn_fop->f_event(kn, NOTE_FORK))
|
if (kn->kn_fop->f_event(kn, NOTE_FORK))
|
||||||
knote_activate(kn, 0);
|
knote_activate_ul(kn);
|
||||||
list->kl_lock(list->kl_lockarg);
|
list->kl_lock(list->kl_lockarg);
|
||||||
|
|
||||||
KN_FLUX_LOCK(kn);
|
KN_LEAVE_FLUX_WAKEUP(kn);
|
||||||
kn_leave_flux(kn, 1);
|
|
||||||
KN_FLUX_WAKEUP(kn, 1);
|
|
||||||
KN_FLUX_UNLOCK(kn);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -787,13 +788,9 @@ filt_timerexpire(void *knx)
|
|||||||
kn = knx;
|
kn = knx;
|
||||||
kn->kn_data++;
|
kn->kn_data++;
|
||||||
|
|
||||||
kn_enter_flux(kn, 0);
|
knote_enter_flux_ul(kn);
|
||||||
|
knote_activate_ul(kn);
|
||||||
knote_activate(kn, 0);
|
KN_LEAVE_FLUX_WAKEUP(kn);
|
||||||
|
|
||||||
/* TODO: lock? */
|
|
||||||
kn_leave_flux(kn, 0);
|
|
||||||
KN_FLUX_WAKEUP(kn, 0);
|
|
||||||
|
|
||||||
if ((kn->kn_flags & EV_ONESHOT) != 0)
|
if ((kn->kn_flags & EV_ONESHOT) != 0)
|
||||||
return;
|
return;
|
||||||
@ -1606,6 +1603,7 @@ kqueue_register(struct kqueue *kq, struct kevq *kevq, struct kevent *kev, struct
|
|||||||
KQ_GLOBAL_LOCK(&kq_global, haskqglobal);
|
KQ_GLOBAL_LOCK(&kq_global, haskqglobal);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* lock the kq lock for accessing kq_knhash table */
|
||||||
KQ_LOCK(kq);
|
KQ_LOCK(kq);
|
||||||
if (kev->ident < kq->kq_knlistsize) {
|
if (kev->ident < kq->kq_knlistsize) {
|
||||||
SLIST_FOREACH(kn, &kq->kq_knlist[kev->ident], kn_link)
|
SLIST_FOREACH(kn, &kq->kq_knlist[kev->ident], kn_link)
|
||||||
@ -1619,6 +1617,7 @@ kqueue_register(struct kqueue *kq, struct kevq *kevq, struct kevent *kev, struct
|
|||||||
goto done;
|
goto done;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* lock the kq lock for accessing kq_knhash table */
|
||||||
KQ_LOCK(kq);
|
KQ_LOCK(kq);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -1643,6 +1642,9 @@ kqueue_register(struct kqueue *kq, struct kevq *kevq, struct kevent *kev, struct
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* We need the kq lock because attaching to KQ requires KQ Lock */
|
||||||
|
KQ_OWNED(kq);
|
||||||
|
|
||||||
/* knote is in the process of changing, wait for it to stabilize. */
|
/* knote is in the process of changing, wait for it to stabilize. */
|
||||||
if (kn != NULL) {
|
if (kn != NULL) {
|
||||||
KN_FLUX_LOCK(kn);
|
KN_FLUX_LOCK(kn);
|
||||||
@ -1663,7 +1665,7 @@ kqueue_register(struct kqueue *kq, struct kevq *kevq, struct kevent *kev, struct
|
|||||||
goto findkn;
|
goto findkn;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
/* We now have exclusive access to the knote with flux lock */
|
/* We now have exclusive access to the knote with flux lock and kq lock */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* kn now contains the matching knote, or NULL if no match
|
* kn now contains the matching knote, or NULL if no match
|
||||||
@ -1699,7 +1701,7 @@ kqueue_register(struct kqueue *kq, struct kevq *kevq, struct kevent *kev, struct
|
|||||||
kn->kn_status = KN_DETACHED;
|
kn->kn_status = KN_DETACHED;
|
||||||
if ((kev->flags & EV_DISABLE) != 0)
|
if ((kev->flags & EV_DISABLE) != 0)
|
||||||
kn->kn_status |= KN_DISABLED;
|
kn->kn_status |= KN_DISABLED;
|
||||||
kn_enter_flux(kn, 0);
|
knote_enter_flux_ul(kn);
|
||||||
|
|
||||||
error = knote_attach(kn, kq);
|
error = knote_attach(kn, kq);
|
||||||
KQ_UNLOCK(kq);
|
KQ_UNLOCK(kq);
|
||||||
@ -1724,7 +1726,7 @@ kqueue_register(struct kqueue *kq, struct kevq *kevq, struct kevent *kev, struct
|
|||||||
|
|
||||||
if (kev->flags & EV_DELETE) {
|
if (kev->flags & EV_DELETE) {
|
||||||
/* We have the exclusive flux lock here */
|
/* We have the exclusive flux lock here */
|
||||||
kn_enter_flux(kn, 1);
|
knote_enter_flux(kn);
|
||||||
|
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
KQ_UNLOCK(kq);
|
KQ_UNLOCK(kq);
|
||||||
@ -1734,13 +1736,13 @@ kqueue_register(struct kqueue *kq, struct kevq *kevq, struct kevent *kev, struct
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* We have the exclusive lock */
|
/* We have the exclusive lock */
|
||||||
kn_enter_flux(kn, 1);
|
knote_enter_flux(kn);
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
|
|
||||||
// we have kq lock and knote influx
|
// we have kq lock and knote influx
|
||||||
if (kev->flags & EV_FORCEONESHOT) {
|
if (kev->flags & EV_FORCEONESHOT) {
|
||||||
kn->kn_flags |= EV_ONESHOT;
|
kn->kn_flags |= EV_ONESHOT;
|
||||||
knote_activate(kn, 1);
|
knote_activate(kn);
|
||||||
}
|
}
|
||||||
|
|
||||||
if ((kev->flags & EV_ENABLE) != 0)
|
if ((kev->flags & EV_ENABLE) != 0)
|
||||||
@ -1784,16 +1786,14 @@ kqueue_register(struct kqueue *kq, struct kevq *kevq, struct kevent *kev, struct
|
|||||||
|
|
||||||
KQ_LOCK(kq);
|
KQ_LOCK(kq);
|
||||||
if (event)
|
if (event)
|
||||||
knote_activate(kn, 1);
|
knote_activate(kn);
|
||||||
|
|
||||||
kn->kn_status &= ~KN_SCAN;
|
kn->kn_status &= ~KN_SCAN;
|
||||||
|
|
||||||
KN_FLUX_LOCK(kn);
|
KN_LEAVE_FLUX_WAKEUP(kn);
|
||||||
kn_leave_flux(kn, 1);
|
|
||||||
KN_FLUX_WAKEUP(kn, 1);
|
|
||||||
KN_FLUX_UNLOCK(kn);
|
|
||||||
|
|
||||||
KQ_UNLOCK(kq);
|
KQ_UNLOCK(kq);
|
||||||
|
|
||||||
kn_list_unlock(knl);
|
kn_list_unlock(knl);
|
||||||
|
|
||||||
done:
|
done:
|
||||||
@ -2271,7 +2271,7 @@ kqueue_scan(struct kevq *kevq, int maxevents, struct kevent_copyops *k_ops,
|
|||||||
kn_in_flux(kn)) {
|
kn_in_flux(kn)) {
|
||||||
if (influx) {
|
if (influx) {
|
||||||
influx = 0;
|
influx = 0;
|
||||||
KN_FLUX_WAKEUP(kn, 1);
|
knote_flux_wakeup(kn);
|
||||||
}
|
}
|
||||||
kn->kn_fluxwait = 1;
|
kn->kn_fluxwait = 1;
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
@ -2293,7 +2293,7 @@ kqueue_scan(struct kevq *kevq, int maxevents, struct kevent_copyops *k_ops,
|
|||||||
}
|
}
|
||||||
if (kn == marker) {
|
if (kn == marker) {
|
||||||
/* We are dequeuing our marker, wakeup threads waiting on it */
|
/* We are dequeuing our marker, wakeup threads waiting on it */
|
||||||
KN_FLUX_WAKEUP(kn, 1);
|
knote_flux_wakeup(kn);
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
if (count == maxevents) {
|
if (count == maxevents) {
|
||||||
goto retry;
|
goto retry;
|
||||||
@ -2305,7 +2305,7 @@ kqueue_scan(struct kevq *kevq, int maxevents, struct kevent_copyops *k_ops,
|
|||||||
|
|
||||||
if ((kn->kn_flags & EV_DROP) == EV_DROP) {
|
if ((kn->kn_flags & EV_DROP) == EV_DROP) {
|
||||||
kn->kn_status &= ~KN_QUEUED;
|
kn->kn_status &= ~KN_QUEUED;
|
||||||
kn_enter_flux(kn, 1);
|
knote_enter_flux(kn);
|
||||||
kevq->kn_count--;
|
kevq->kn_count--;
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
KEVQ_UNLOCK(kevq);
|
KEVQ_UNLOCK(kevq);
|
||||||
@ -2318,7 +2318,7 @@ kqueue_scan(struct kevq *kevq, int maxevents, struct kevent_copyops *k_ops,
|
|||||||
continue;
|
continue;
|
||||||
} else if ((kn->kn_flags & EV_ONESHOT) == EV_ONESHOT) {
|
} else if ((kn->kn_flags & EV_ONESHOT) == EV_ONESHOT) {
|
||||||
kn->kn_status &= ~KN_QUEUED;
|
kn->kn_status &= ~KN_QUEUED;
|
||||||
kn_enter_flux(kn, 1);
|
knote_enter_flux(kn);
|
||||||
kevq->kn_count--;
|
kevq->kn_count--;
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
KEVQ_UNLOCK(kevq);
|
KEVQ_UNLOCK(kevq);
|
||||||
@ -2332,7 +2332,7 @@ kqueue_scan(struct kevq *kevq, int maxevents, struct kevent_copyops *k_ops,
|
|||||||
kn = NULL;
|
kn = NULL;
|
||||||
} else {
|
} else {
|
||||||
kn->kn_status |= KN_SCAN;
|
kn->kn_status |= KN_SCAN;
|
||||||
kn_enter_flux(kn, 1);
|
knote_enter_flux(kn);
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
KEVQ_UNLOCK(kevq);
|
KEVQ_UNLOCK(kevq);
|
||||||
if ((kn->kn_status & KN_KQUEUE) == KN_KQUEUE) {
|
if ((kn->kn_status & KN_KQUEUE) == KN_KQUEUE) {
|
||||||
@ -2346,7 +2346,7 @@ kqueue_scan(struct kevq *kevq, int maxevents, struct kevent_copyops *k_ops,
|
|||||||
KQ_GLOBAL_UNLOCK(&kq_global, haskqglobal);
|
KQ_GLOBAL_UNLOCK(&kq_global, haskqglobal);
|
||||||
kn->kn_status &= ~(KN_QUEUED | KN_ACTIVE |
|
kn->kn_status &= ~(KN_QUEUED | KN_ACTIVE |
|
||||||
KN_SCAN);
|
KN_SCAN);
|
||||||
kn_leave_flux(kn, 0);
|
knote_leave_flux_ul(kn);
|
||||||
kevq->kn_count--;
|
kevq->kn_count--;
|
||||||
kn_list_unlock(knl);
|
kn_list_unlock(knl);
|
||||||
influx = 1;
|
influx = 1;
|
||||||
@ -2383,7 +2383,7 @@ kqueue_scan(struct kevq *kevq, int maxevents, struct kevent_copyops *k_ops,
|
|||||||
}
|
}
|
||||||
|
|
||||||
kn->kn_status &= ~KN_SCAN;
|
kn->kn_status &= ~KN_SCAN;
|
||||||
kn_leave_flux(kn, 0);
|
knote_leave_flux_ul(kn);
|
||||||
kn_list_unlock(knl);
|
kn_list_unlock(knl);
|
||||||
influx = 1;
|
influx = 1;
|
||||||
}
|
}
|
||||||
@ -2394,9 +2394,11 @@ kqueue_scan(struct kevq *kevq, int maxevents, struct kevent_copyops *k_ops,
|
|||||||
count--;
|
count--;
|
||||||
|
|
||||||
if (nkev == KQ_NEVENTS) {
|
if (nkev == KQ_NEVENTS) {
|
||||||
influx = 0;
|
|
||||||
KN_FLUX_WAKEUP(kn, 0);
|
|
||||||
KEVQ_UNLOCK(kevq);
|
KEVQ_UNLOCK(kevq);
|
||||||
|
|
||||||
|
influx = 0;
|
||||||
|
knote_flux_wakeup_ul(kn);
|
||||||
|
|
||||||
error = k_ops->k_copyout(k_ops->arg, keva, nkev);
|
error = k_ops->k_copyout(k_ops->arg, keva, nkev);
|
||||||
nkev = 0;
|
nkev = 0;
|
||||||
kevp = keva;
|
kevp = keva;
|
||||||
@ -2411,7 +2413,7 @@ kqueue_scan(struct kevq *kevq, int maxevents, struct kevent_copyops *k_ops,
|
|||||||
KEVQ_UNLOCK(kevq);
|
KEVQ_UNLOCK(kevq);
|
||||||
|
|
||||||
if (kn != NULL) {
|
if (kn != NULL) {
|
||||||
KN_FLUX_WAKEUP(kn, 0);
|
knote_flux_wakeup_ul(kn);
|
||||||
}
|
}
|
||||||
|
|
||||||
knote_free(marker);
|
knote_free(marker);
|
||||||
@ -2559,7 +2561,7 @@ kevq_destroy(struct kevq *kevq)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* This is called on every kevq when kqueue exits
|
/* This is called on every kevq when kqueue exits
|
||||||
This is also called when a thread exits/crashes
|
This is also called when a thread exits/crashes (currently racing, also to make it work need to reconfigure kq->ck_evq)
|
||||||
* a ref cnt must be held */
|
* a ref cnt must be held */
|
||||||
void
|
void
|
||||||
kevq_drain(struct kevq *kevq)
|
kevq_drain(struct kevq *kevq)
|
||||||
@ -2576,7 +2578,6 @@ kevq_drain(struct kevq *kevq)
|
|||||||
KEVQ_NOTOWNED(kevq);
|
KEVQ_NOTOWNED(kevq);
|
||||||
|
|
||||||
KEVQ_LOCK(kevq);
|
KEVQ_LOCK(kevq);
|
||||||
// now the refcnt is 1
|
|
||||||
if(kevq->kevq_state == KEVQ_CLOSING) {
|
if(kevq->kevq_state == KEVQ_CLOSING) {
|
||||||
// already closing, dereference
|
// already closing, dereference
|
||||||
kevq_release(kevq, 1);
|
kevq_release(kevq, 1);
|
||||||
@ -2602,38 +2603,35 @@ kevq_drain(struct kevq *kevq)
|
|||||||
if (kn_in_flux(kn)) {
|
if (kn_in_flux(kn)) {
|
||||||
kn->kn_fluxwait = 1;
|
kn->kn_fluxwait = 1;
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
msleep(kn, &kevq->lock, PSOCK, "kevqclo1", 0);
|
msleep(kn, &kevq->lock, PSOCK, "kevqclose2", 0);
|
||||||
goto retry;
|
goto retry;
|
||||||
}
|
}
|
||||||
|
|
||||||
KN_FLUX_OWNED(kn);
|
KN_FLUX_OWNED(kn);
|
||||||
KASSERT(!kn_in_flux(kn), ("knote is still influx"));
|
KASSERT(!kn_in_flux(kn), ("knote is still influx"));
|
||||||
|
|
||||||
kn_enter_flux(kn, 1);
|
knote_enter_flux(kn);
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
knote_dequeue(kn);
|
knote_dequeue(kn);
|
||||||
|
|
||||||
if ((kq->kq_state & KQ_FLAG_MULTI) == KQ_FLAG_MULTI && (kq->kq_state & KQ_CLOSING) != KQ_CLOSING) {
|
if ((kq->kq_state & KQ_FLAG_MULTI) == KQ_FLAG_MULTI && (kq->kq_state & KQ_CLOSING) != KQ_CLOSING) {
|
||||||
KEVQ_UNLOCK(kevq);
|
KEVQ_UNLOCK(kevq);
|
||||||
knote_activate(kn, 0);
|
knote_activate_ul(kn);
|
||||||
KEVQ_LOCK(kevq);
|
KEVQ_LOCK(kevq);
|
||||||
}
|
}
|
||||||
|
|
||||||
KN_FLUX_LOCK(kn);
|
KN_LEAVE_FLUX_WAKEUP(kn);
|
||||||
kn_leave_flux(kn, 1);
|
|
||||||
KN_FLUX_WAKEUP(kn, 1);
|
|
||||||
KN_FLUX_UNLOCK(kn);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
KASSERT(kevq->kn_count == 0, ("some knotes are left"));
|
KASSERT(kevq->kn_count == 0, ("some knotes are left"));
|
||||||
KEVQ_UNLOCK(kevq);
|
KEVQ_UNLOCK(kevq);
|
||||||
|
|
||||||
if ((kq->kq_state & KQ_FLAG_MULTI) == KQ_FLAG_MULTI) {
|
|
||||||
// drop from kq if in multithreaded mode
|
|
||||||
KQ_LOCK(kq);
|
KQ_LOCK(kq);
|
||||||
TAILQ_REMOVE(&kq->kq_kevqlist, kevq, kq_e);
|
TAILQ_REMOVE(&kq->kq_kevqlist, kevq, kq_e);
|
||||||
KQ_UNLOCK(kq);
|
KQ_UNLOCK(kq);
|
||||||
|
|
||||||
|
if ((kq->kq_state & KQ_FLAG_MULTI) == KQ_FLAG_MULTI) {
|
||||||
|
// drop from kq if in multithreaded mode
|
||||||
KEVQ_TH_LOCK(kevq->kevq_th);
|
KEVQ_TH_LOCK(kevq->kevq_th);
|
||||||
TAILQ_REMOVE(&kevq->kevq_th->kevq_tq, kevq, kevq_th_tqe);
|
TAILQ_REMOVE(&kevq->kevq_th->kevq_tq, kevq, kevq_th_tqe);
|
||||||
kevq_list = &kevq->kevq_th->kevq_hash[KEVQ_HASH((unsigned long long)kq, kevq->kevq_th->kevq_hashmask)];
|
kevq_list = &kevq->kevq_th->kevq_hash[KEVQ_HASH((unsigned long long)kq, kevq->kevq_th->kevq_hashmask)];
|
||||||
@ -2641,6 +2639,7 @@ kevq_drain(struct kevq *kevq)
|
|||||||
KEVQ_TH_UNLOCK(kevq->kevq_th);
|
KEVQ_TH_UNLOCK(kevq->kevq_th);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
/* delete the kevq */
|
/* delete the kevq */
|
||||||
kevq_destroy(kevq);
|
kevq_destroy(kevq);
|
||||||
}
|
}
|
||||||
@ -2699,7 +2698,7 @@ kqueue_drain(struct kqueue *kq, struct kevq *kevq, struct thread *td)
|
|||||||
msleep(kn, &kq->kq_lock, PSOCK, "kqclo1", 0);
|
msleep(kn, &kq->kq_lock, PSOCK, "kqclo1", 0);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
kn_enter_flux(kn, 1);
|
knote_enter_flux(kn);
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
KQ_UNLOCK(kq);
|
KQ_UNLOCK(kq);
|
||||||
knote_drop(kn, td);
|
knote_drop(kn, td);
|
||||||
@ -2716,7 +2715,7 @@ kqueue_drain(struct kqueue *kq, struct kevq *kevq, struct thread *td)
|
|||||||
msleep(kn, &kq->kq_lock, PSOCK, "kqclo2", 0);
|
msleep(kn, &kq->kq_lock, PSOCK, "kqclo2", 0);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
kn_enter_flux(kn, 1);
|
knote_enter_flux(kn);
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
KQ_UNLOCK(kq);
|
KQ_UNLOCK(kq);
|
||||||
knote_drop(kn, td);
|
knote_drop(kn, td);
|
||||||
@ -2888,7 +2887,7 @@ knote(struct knlist *list, long hint, int lockflags)
|
|||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
} else {
|
} else {
|
||||||
kq = kn->kn_kq;
|
kq = kn->kn_kq;
|
||||||
kn_enter_flux(kn, 1);
|
knote_enter_flux(kn);
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
|
|
||||||
require_kqlock = ((lockflags & KNF_NOKQLOCK) == 0);
|
require_kqlock = ((lockflags & KNF_NOKQLOCK) == 0);
|
||||||
@ -2896,70 +2895,73 @@ knote(struct knlist *list, long hint, int lockflags)
|
|||||||
if (require_kqlock)
|
if (require_kqlock)
|
||||||
KQ_LOCK(kq);
|
KQ_LOCK(kq);
|
||||||
|
|
||||||
if (kn->kn_fop->f_event(kn, hint))
|
if (kn->kn_fop->f_event(kn, hint)){
|
||||||
knote_activate(kn, require_kqlock);
|
// TODO: WTH is this?
|
||||||
|
require_kqlock ? knote_activate(kn) : knote_activate_ul(kn);
|
||||||
|
}
|
||||||
|
|
||||||
if (require_kqlock)
|
if (require_kqlock)
|
||||||
KQ_UNLOCK(kq);
|
KQ_UNLOCK(kq);
|
||||||
|
|
||||||
KN_FLUX_LOCK(kn);
|
KN_LEAVE_FLUX_WAKEUP(kn);
|
||||||
kn_leave_flux(kn, 1);
|
|
||||||
KN_FLUX_WAKEUP(kn, 1);
|
|
||||||
KN_FLUX_UNLOCK(kn);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if ((lockflags & KNF_LISTLOCKED) == 0)
|
if ((lockflags & KNF_LISTLOCKED) == 0)
|
||||||
list->kl_unlock(list->kl_lockarg);
|
list->kl_unlock(list->kl_lockarg);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void
|
||||||
|
knote_flux_wakeup_ul(struct knote *kn)
|
||||||
|
{
|
||||||
|
KN_FLUX_NOTOWNED(kn);
|
||||||
|
KN_FLUX_LOCK(kn);
|
||||||
|
knote_flux_wakeup(kn);
|
||||||
|
KN_FLUX_UNLOCK(kn);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void
|
||||||
|
knote_flux_wakeup(struct knote *kn)
|
||||||
|
{
|
||||||
|
KN_FLUX_OWNED(kn);
|
||||||
|
if ((kn)->kn_fluxwait) {
|
||||||
|
(kn)->kn_fluxwait = 0;
|
||||||
|
wakeup((kn));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static void
|
||||||
|
knote_activate_ul(struct knote *kn)
|
||||||
|
{
|
||||||
|
KQ_LOCK(kn->kn_kq);
|
||||||
|
knote_activate(kn);
|
||||||
|
KQ_UNLOCK(kn->kn_kq);
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* activate a knote
|
* activate a knote
|
||||||
* the knote should be marked in flux and the knote flux lock should not be owned
|
* the knote should be marked in flux and the knote flux lock should not be owned
|
||||||
|
* none of the other locks should be held
|
||||||
*/
|
*/
|
||||||
static void
|
static void
|
||||||
knote_activate(struct knote *kn, int haskqlock)
|
knote_activate(struct knote *kn)
|
||||||
{
|
{
|
||||||
struct kevq *kevq;
|
|
||||||
struct kqueue *kq;
|
struct kqueue *kq;
|
||||||
|
kq = kn->kn_kq;
|
||||||
|
|
||||||
#ifdef KQ_DEBUG
|
#ifdef KQ_DEBUG
|
||||||
printf("KQUEUE: knote_activate: kn %p\n", kn);
|
printf("KQUEUE: knote_activate: kn %p\n", kn);
|
||||||
#endif
|
#endif
|
||||||
kq = kn->kn_kq;
|
|
||||||
|
|
||||||
if (haskqlock) {
|
|
||||||
KQ_OWNED(kq);
|
KQ_OWNED(kq);
|
||||||
} else {
|
|
||||||
KQ_LOCK(kq);
|
|
||||||
}
|
|
||||||
|
|
||||||
KN_FLUX_NOTOWNED(kn);
|
KN_FLUX_NOTOWNED(kn);
|
||||||
KASSERT(kn_in_flux(kn), ("knote %p not in flux", kn));
|
KASSERT(kn_in_flux(kn), ("knote %p not in flux", kn));
|
||||||
|
|
||||||
kn->kn_status |= KN_ACTIVE;
|
kn->kn_status |= KN_ACTIVE;
|
||||||
|
|
||||||
retry:
|
|
||||||
if (((kn)->kn_status & (KN_QUEUED | KN_DISABLED)) == 0) {
|
if (((kn)->kn_status & (KN_QUEUED | KN_DISABLED)) == 0) {
|
||||||
kevq = knote_sched(kn);
|
knote_sched(kn);
|
||||||
|
|
||||||
if (kevq != NULL) {
|
|
||||||
// if we have a queue to queue the knote
|
|
||||||
KEVQ_LOCK(kevq);
|
|
||||||
|
|
||||||
if ((kevq->kevq_state & KEVQ_CLOSING) != 0) {
|
|
||||||
KEVQ_UNLOCK(kevq);
|
|
||||||
goto retry;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
knote_enqueue(kn, kevq);
|
kqueue_wakeup(kq);
|
||||||
|
|
||||||
KEVQ_UNLOCK(kevq);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!haskqlock) {
|
|
||||||
KQ_UNLOCK(kq);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -3198,7 +3200,7 @@ knlist_cleardel(struct knlist *knl, struct thread *td, int islocked, int killkn)
|
|||||||
}
|
}
|
||||||
knlist_remove_kq(knl, kn, 1, 1);
|
knlist_remove_kq(knl, kn, 1, 1);
|
||||||
if (killkn) {
|
if (killkn) {
|
||||||
kn_enter_flux(kn, 1);
|
knote_enter_flux(kn);
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
KQ_UNLOCK(kq);
|
KQ_UNLOCK(kq);
|
||||||
knote_drop_detached(kn, td);
|
knote_drop_detached(kn, td);
|
||||||
@ -3265,13 +3267,13 @@ knote_fdclose(struct thread *td, int fd)
|
|||||||
if (kn_in_flux(kn)) {
|
if (kn_in_flux(kn)) {
|
||||||
/* someone else might be waiting on our knote */
|
/* someone else might be waiting on our knote */
|
||||||
if (influx)
|
if (influx)
|
||||||
KN_FLUX_WAKEUP(kn, 1);
|
knote_flux_wakeup(kn);
|
||||||
kn->kn_fluxwait = 1;
|
kn->kn_fluxwait = 1;
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
msleep(kn, &kq->kq_lock, PSOCK, "kqflxwt", 0);
|
msleep(kn, &kq->kq_lock, PSOCK, "kqflxwt", 0);
|
||||||
goto again;
|
goto again;
|
||||||
}
|
}
|
||||||
kn_enter_flux(kn, 1);
|
knote_enter_flux(kn);
|
||||||
KN_FLUX_UNLOCK(kn);
|
KN_FLUX_UNLOCK(kn);
|
||||||
KQ_UNLOCK(kq);
|
KQ_UNLOCK(kq);
|
||||||
/* WTF ? influx should be 0? */
|
/* WTF ? influx should be 0? */
|
||||||
@ -3361,11 +3363,11 @@ knote_drop_detached(struct knote *kn, struct thread *td)
|
|||||||
|
|
||||||
|
|
||||||
// if no kevqs are available for queueing, returns NULL
|
// if no kevqs are available for queueing, returns NULL
|
||||||
static struct kevq*
|
static void
|
||||||
knote_sched(struct knote *kn)
|
knote_sched(struct knote *kn)
|
||||||
{
|
{
|
||||||
struct kqueue *kq = kn->kn_kq;
|
struct kqueue *kq = kn->kn_kq;
|
||||||
struct kevq *each_kevq;
|
struct kevq *next_kevq;
|
||||||
|
|
||||||
KQ_OWNED(kq);
|
KQ_OWNED(kq);
|
||||||
KASSERT(kn_in_flux(kn), ("kn not in flux"));
|
KASSERT(kn_in_flux(kn), ("kn not in flux"));
|
||||||
@ -3377,55 +3379,63 @@ knote_sched(struct knote *kn)
|
|||||||
printf("KQUEUE: knote_sched(M) affinity set: kn %p \n", kn);
|
printf("KQUEUE: knote_sched(M) affinity set: kn %p \n", kn);
|
||||||
#endif
|
#endif
|
||||||
if ((kn->kn_org_kevq->kevq_state & KEVQ_RDY) != 0) {
|
if ((kn->kn_org_kevq->kevq_state & KEVQ_RDY) != 0) {
|
||||||
return kn->kn_org_kevq;
|
next_kevq = kn->kn_org_kevq;
|
||||||
} else {
|
} else {
|
||||||
return NULL;
|
next_kevq = NULL;
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
each_kevq = kq->kq_ckevq;
|
next_kevq = kq->kq_ckevq;
|
||||||
while(1) {
|
while(1) {
|
||||||
if (each_kevq == NULL) {
|
if (next_kevq == NULL) {
|
||||||
each_kevq = TAILQ_FIRST(&kq->kq_kevqlist);
|
next_kevq = TAILQ_FIRST(&kq->kq_kevqlist);
|
||||||
if (each_kevq == NULL) {
|
if (next_kevq == NULL) {
|
||||||
#ifdef KQ_DEBUG
|
#ifdef KQ_DEBUG
|
||||||
printf("KQUEUE: knote_sched(M) no kevqs exist for queueing kn %p, discarding... \n", kn);
|
printf("KQUEUE: knote_sched(M) no kevqs exist for queueing kn %p, discarding... \n", kn);
|
||||||
#endif
|
#endif
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
each_kevq = TAILQ_NEXT(each_kevq, kq_e);
|
next_kevq = TAILQ_NEXT(next_kevq, kq_e);
|
||||||
if (each_kevq == NULL) {
|
if (next_kevq == NULL) {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if ((each_kevq->kevq_state & KEVQ_CLOSING) == 0 && (each_kevq->kevq_state & KEVQ_RDY) != 0) {
|
KASSERT(next_kevq != NULL, ("picked a null kevq"));
|
||||||
|
KEVQ_LOCK(next_kevq);
|
||||||
|
|
||||||
|
if ((next_kevq->kevq_state & KEVQ_CLOSING) == 0 && (next_kevq->kevq_state & KEVQ_RDY) != 0) {
|
||||||
|
kq->kq_ckevq = next_kevq;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (each_kevq == kq->kq_ckevq) {
|
if (next_kevq == kq->kq_ckevq) {
|
||||||
// if the previous "if" didn't break
|
// if the previous "if" didn't break
|
||||||
// we have traversed the list once and the current kevq is closing
|
// we have traversed the list once and the current kevq is closing
|
||||||
// we have no queue to queue the knote
|
// we have no queue to queue the knote
|
||||||
#ifdef KQ_DEBUG
|
#ifdef KQ_DEBUG
|
||||||
printf("KQUEUE: knote_sched(M) no open kevqs for queueing kn %p, discarding... \n", kn);
|
printf("KQUEUE: knote_sched(M) no open kevqs for queueing kn %p, discarding... \n", kn);
|
||||||
#endif
|
#endif
|
||||||
each_kevq = NULL;
|
next_kevq = NULL;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef KQ_DEBUG
|
|
||||||
printf("KQUEUE: knote_sched(M) next kevq %p for kn %p \n", each_kevq, kn);
|
|
||||||
#endif
|
|
||||||
kq->kq_ckevq = each_kevq;
|
|
||||||
return each_kevq;
|
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
#ifdef KQ_DEBUG
|
#ifdef KQ_DEBUG
|
||||||
printf("KQUEUE: knote_sched(S): kn %p to kevq %p\n", kn, kq->kq_kevq);
|
printf("KQUEUE: knote_sched(S): kn %p to kevq %p\n", kn, kq->kq_kevq);
|
||||||
#endif
|
#endif
|
||||||
return kq->kq_kevq;
|
next_kevq = kq->kq_kevq;
|
||||||
|
}
|
||||||
|
|
||||||
|
#ifdef KQ_DEBUG
|
||||||
|
printf("KQUEUE: knote_sched(M) next kevq %p for kn %p \n", next_kevq, kn);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
if (next_kevq != NULL) {
|
||||||
|
KEVQ_OWNED(next_kevq);
|
||||||
|
knote_enqueue(kn, next_kevq);
|
||||||
|
KEVQ_UNLOCK(next_kevq);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -3439,8 +3449,6 @@ knote_enqueue(struct knote *kn, struct kevq *kevq)
|
|||||||
printf("KQUEUE: knote_enqueue: kn %p to kevq %p\n", kn, kevq);
|
printf("KQUEUE: knote_enqueue: kn %p to kevq %p\n", kn, kevq);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/* TODO: optimize locking, we don't really need KQ_LOCK here except for kqueue_wakeup() */
|
|
||||||
KQ_OWNED(kq);
|
|
||||||
KEVQ_OWNED(kevq);
|
KEVQ_OWNED(kevq);
|
||||||
|
|
||||||
KASSERT((kn->kn_status & KN_QUEUED) == 0, ("knote already queued"));
|
KASSERT((kn->kn_status & KN_QUEUED) == 0, ("knote already queued"));
|
||||||
@ -3452,7 +3460,6 @@ knote_enqueue(struct knote *kn, struct kevq *kevq)
|
|||||||
TAILQ_INSERT_TAIL(&kevq->kn_head, kn, kn_tqe);
|
TAILQ_INSERT_TAIL(&kevq->kn_head, kn, kn_tqe);
|
||||||
kevq->kn_count++;
|
kevq->kn_count++;
|
||||||
|
|
||||||
kqueue_wakeup(kq);
|
|
||||||
kevq_wakeup(kevq);
|
kevq_wakeup(kevq);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user