ad355b0a9d
r357614 added CTLFLAG_NEEDGIANT to make it easier to find nodes that are still not MPSAFE (or already are but aren’t properly marked). Use it in preparation for a general review of all nodes. This is non-functional change that adds annotations to SYSCTL_NODE and SYSCTL_PROC nodes using one of the soon-to-be-required flags. Mark all obvious cases as MPSAFE. All entries that haven't been marked as MPSAFE before are by default marked as NEEDGIANT Approved by: kib (mentor, blanket) Commented by: kib, gallatin, melifaro Differential Revision: https://reviews.freebsd.org/D23718
803 lines
20 KiB
C
803 lines
20 KiB
C
/*-
|
|
* SPDX-License-Identifier: BSD-2-Clause-FreeBSD
|
|
*
|
|
* Copyright (c) 2007 Ariff Abdullah <ariff@FreeBSD.org>
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/conf.h>
|
|
#include <sys/kernel.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/proc.h>
|
|
|
|
#ifdef HAVE_KERNEL_OPTION_HEADERS
|
|
#include "opt_snd.h"
|
|
#endif
|
|
|
|
#if defined(SND_DIAGNOSTIC) || defined(SND_DEBUG)
|
|
#include <dev/sound/pcm/sound.h>
|
|
#endif
|
|
|
|
#include <dev/sound/clone.h>
|
|
|
|
/*
|
|
* So here we go again, another clonedevs manager. Unlike default clonedevs,
|
|
* this clone manager is designed to withstand various abusive behavior
|
|
* (such as 'while : ; do ls /dev/whatever ; done', etc.), reusable object
|
|
* after reaching certain expiration threshold, aggressive garbage collector,
|
|
* transparent device allocator and concurrency handling across multiple
|
|
* thread/proc. Due to limited information given by dev_clone EVENTHANDLER,
|
|
* we don't have much clues whether the caller wants a real open() or simply
|
|
* making fun of us with things like stat(), mtime() etc. Assuming that:
|
|
* 1) Time window between dev_clone EH <-> real open() should be small
|
|
* enough and 2) mtime()/stat() etc. always looks like a half way / stalled
|
|
* operation, we can decide whether a new cdev must be created, old
|
|
* (expired) cdev can be reused or an existing cdev can be shared.
|
|
*
|
|
* Most of the operations and logics are generic enough and can be applied
|
|
* on other places (such as if_tap, snp, etc). Perhaps this can be
|
|
* rearranged to complement clone_*(). However, due to this still being
|
|
* specific to the sound driver (and as a proof of concept on how it can be
|
|
* done), si_drv2 is used to keep the pointer of the clone list entry to
|
|
* avoid expensive lookup.
|
|
*/
|
|
|
|
/* clone entry */
|
|
struct snd_clone_entry {
|
|
TAILQ_ENTRY(snd_clone_entry) link;
|
|
struct snd_clone *parent;
|
|
struct cdev *devt;
|
|
struct timespec tsp;
|
|
uint32_t flags;
|
|
pid_t pid;
|
|
int unit;
|
|
};
|
|
|
|
/* clone manager */
|
|
struct snd_clone {
|
|
TAILQ_HEAD(link_head, snd_clone_entry) head;
|
|
struct timespec tsp;
|
|
int refcount;
|
|
int size;
|
|
int typemask;
|
|
int maxunit;
|
|
int deadline;
|
|
uint32_t flags;
|
|
};
|
|
|
|
#ifdef SND_DIAGNOSTIC
|
|
#define SND_CLONE_ASSERT(x, y) do { \
|
|
if (!(x)) \
|
|
panic y; \
|
|
} while (0)
|
|
#else
|
|
#define SND_CLONE_ASSERT(...) KASSERT(__VA_ARGS__)
|
|
#endif
|
|
|
|
/*
|
|
* Shamelessly ripped off from vfs_subr.c
|
|
* We need at least 1/HZ precision as default timestamping.
|
|
*/
|
|
enum { SND_TSP_SEC, SND_TSP_HZ, SND_TSP_USEC, SND_TSP_NSEC };
|
|
|
|
static int snd_timestamp_precision = SND_TSP_HZ;
|
|
TUNABLE_INT("hw.snd.timestamp_precision", &snd_timestamp_precision);
|
|
|
|
void
|
|
snd_timestamp(struct timespec *tsp)
|
|
{
|
|
struct timeval tv;
|
|
|
|
switch (snd_timestamp_precision) {
|
|
case SND_TSP_SEC:
|
|
tsp->tv_sec = time_second;
|
|
tsp->tv_nsec = 0;
|
|
break;
|
|
case SND_TSP_HZ:
|
|
getnanouptime(tsp);
|
|
break;
|
|
case SND_TSP_USEC:
|
|
microuptime(&tv);
|
|
TIMEVAL_TO_TIMESPEC(&tv, tsp);
|
|
break;
|
|
case SND_TSP_NSEC:
|
|
nanouptime(tsp);
|
|
break;
|
|
default:
|
|
snd_timestamp_precision = SND_TSP_HZ;
|
|
getnanouptime(tsp);
|
|
break;
|
|
}
|
|
}
|
|
|
|
#if defined(SND_DIAGNOSTIC) || defined(SND_DEBUG)
|
|
static int
|
|
sysctl_hw_snd_timestamp_precision(SYSCTL_HANDLER_ARGS)
|
|
{
|
|
int err, val;
|
|
|
|
val = snd_timestamp_precision;
|
|
err = sysctl_handle_int(oidp, &val, 0, req);
|
|
if (err == 0 && req->newptr != NULL) {
|
|
switch (val) {
|
|
case SND_TSP_SEC:
|
|
case SND_TSP_HZ:
|
|
case SND_TSP_USEC:
|
|
case SND_TSP_NSEC:
|
|
snd_timestamp_precision = val;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
return (err);
|
|
}
|
|
SYSCTL_PROC(_hw_snd, OID_AUTO, timestamp_precision,
|
|
CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_NEEDGIANT, 0, sizeof(int),
|
|
sysctl_hw_snd_timestamp_precision, "I",
|
|
"timestamp precision (0=s 1=hz 2=us 3=ns)");
|
|
#endif
|
|
|
|
/*
|
|
* snd_clone_create() : Return opaque allocated clone manager.
|
|
*/
|
|
struct snd_clone *
|
|
snd_clone_create(int typemask, int maxunit, int deadline, uint32_t flags)
|
|
{
|
|
struct snd_clone *c;
|
|
|
|
SND_CLONE_ASSERT(!(typemask & ~SND_CLONE_MAXUNIT),
|
|
("invalid typemask: 0x%08x", typemask));
|
|
SND_CLONE_ASSERT(maxunit == -1 ||
|
|
!(maxunit & ~(~typemask & SND_CLONE_MAXUNIT)),
|
|
("maxunit overflow: typemask=0x%08x maxunit=%d",
|
|
typemask, maxunit));
|
|
SND_CLONE_ASSERT(!(flags & ~SND_CLONE_MASK),
|
|
("invalid clone flags=0x%08x", flags));
|
|
|
|
c = malloc(sizeof(*c), M_DEVBUF, M_WAITOK | M_ZERO);
|
|
c->refcount = 0;
|
|
c->size = 0;
|
|
c->typemask = typemask;
|
|
c->maxunit = (maxunit == -1) ? (~typemask & SND_CLONE_MAXUNIT) :
|
|
maxunit;
|
|
c->deadline = deadline;
|
|
c->flags = flags;
|
|
snd_timestamp(&c->tsp);
|
|
TAILQ_INIT(&c->head);
|
|
|
|
return (c);
|
|
}
|
|
|
|
int
|
|
snd_clone_busy(struct snd_clone *c)
|
|
{
|
|
struct snd_clone_entry *ce;
|
|
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
|
|
if (c->size == 0)
|
|
return (0);
|
|
|
|
TAILQ_FOREACH(ce, &c->head, link) {
|
|
if ((ce->flags & SND_CLONE_BUSY) ||
|
|
(ce->devt != NULL && ce->devt->si_threadcount != 0))
|
|
return (EBUSY);
|
|
}
|
|
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* snd_clone_enable()/disable() : Suspend/resume clone allocation through
|
|
* snd_clone_alloc(). Everything else will not be affected by this.
|
|
*/
|
|
int
|
|
snd_clone_enable(struct snd_clone *c)
|
|
{
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
|
|
if (c->flags & SND_CLONE_ENABLE)
|
|
return (EINVAL);
|
|
|
|
c->flags |= SND_CLONE_ENABLE;
|
|
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
snd_clone_disable(struct snd_clone *c)
|
|
{
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
|
|
if (!(c->flags & SND_CLONE_ENABLE))
|
|
return (EINVAL);
|
|
|
|
c->flags &= ~SND_CLONE_ENABLE;
|
|
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* Getters / Setters. Not worth explaining :)
|
|
*/
|
|
int
|
|
snd_clone_getsize(struct snd_clone *c)
|
|
{
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
|
|
return (c->size);
|
|
}
|
|
|
|
int
|
|
snd_clone_getmaxunit(struct snd_clone *c)
|
|
{
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
|
|
return (c->maxunit);
|
|
}
|
|
|
|
int
|
|
snd_clone_setmaxunit(struct snd_clone *c, int maxunit)
|
|
{
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
SND_CLONE_ASSERT(maxunit == -1 ||
|
|
!(maxunit & ~(~c->typemask & SND_CLONE_MAXUNIT)),
|
|
("maxunit overflow: typemask=0x%08x maxunit=%d",
|
|
c->typemask, maxunit));
|
|
|
|
c->maxunit = (maxunit == -1) ? (~c->typemask & SND_CLONE_MAXUNIT) :
|
|
maxunit;
|
|
|
|
return (c->maxunit);
|
|
}
|
|
|
|
int
|
|
snd_clone_getdeadline(struct snd_clone *c)
|
|
{
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
|
|
return (c->deadline);
|
|
}
|
|
|
|
int
|
|
snd_clone_setdeadline(struct snd_clone *c, int deadline)
|
|
{
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
|
|
c->deadline = deadline;
|
|
|
|
return (c->deadline);
|
|
}
|
|
|
|
int
|
|
snd_clone_gettime(struct snd_clone *c, struct timespec *tsp)
|
|
{
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
SND_CLONE_ASSERT(tsp != NULL, ("NULL timespec"));
|
|
|
|
*tsp = c->tsp;
|
|
|
|
return (0);
|
|
}
|
|
|
|
uint32_t
|
|
snd_clone_getflags(struct snd_clone *c)
|
|
{
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
|
|
return (c->flags);
|
|
}
|
|
|
|
uint32_t
|
|
snd_clone_setflags(struct snd_clone *c, uint32_t flags)
|
|
{
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
SND_CLONE_ASSERT(!(flags & ~SND_CLONE_MASK),
|
|
("invalid clone flags=0x%08x", flags));
|
|
|
|
c->flags = flags;
|
|
|
|
return (c->flags);
|
|
}
|
|
|
|
int
|
|
snd_clone_getdevtime(struct cdev *dev, struct timespec *tsp)
|
|
{
|
|
struct snd_clone_entry *ce;
|
|
|
|
SND_CLONE_ASSERT(dev != NULL, ("NULL dev"));
|
|
SND_CLONE_ASSERT(tsp != NULL, ("NULL timespec"));
|
|
|
|
ce = dev->si_drv2;
|
|
if (ce == NULL)
|
|
return (ENODEV);
|
|
|
|
SND_CLONE_ASSERT(ce->parent != NULL, ("NULL parent"));
|
|
|
|
*tsp = ce->tsp;
|
|
|
|
return (0);
|
|
}
|
|
|
|
uint32_t
|
|
snd_clone_getdevflags(struct cdev *dev)
|
|
{
|
|
struct snd_clone_entry *ce;
|
|
|
|
SND_CLONE_ASSERT(dev != NULL, ("NULL dev"));
|
|
|
|
ce = dev->si_drv2;
|
|
if (ce == NULL)
|
|
return (0xffffffff);
|
|
|
|
SND_CLONE_ASSERT(ce->parent != NULL, ("NULL parent"));
|
|
|
|
return (ce->flags);
|
|
}
|
|
|
|
uint32_t
|
|
snd_clone_setdevflags(struct cdev *dev, uint32_t flags)
|
|
{
|
|
struct snd_clone_entry *ce;
|
|
|
|
SND_CLONE_ASSERT(dev != NULL, ("NULL dev"));
|
|
SND_CLONE_ASSERT(!(flags & ~SND_CLONE_DEVMASK),
|
|
("invalid clone dev flags=0x%08x", flags));
|
|
|
|
ce = dev->si_drv2;
|
|
if (ce == NULL)
|
|
return (0xffffffff);
|
|
|
|
SND_CLONE_ASSERT(ce->parent != NULL, ("NULL parent"));
|
|
|
|
ce->flags = flags;
|
|
|
|
return (ce->flags);
|
|
}
|
|
|
|
/* Elapsed time conversion to ms */
|
|
#define SND_CLONE_ELAPSED(x, y) \
|
|
((((x)->tv_sec - (y)->tv_sec) * 1000) + \
|
|
(((y)->tv_nsec > (x)->tv_nsec) ? \
|
|
(((1000000000L + (x)->tv_nsec - \
|
|
(y)->tv_nsec) / 1000000) - 1000) : \
|
|
(((x)->tv_nsec - (y)->tv_nsec) / 1000000)))
|
|
|
|
#define SND_CLONE_EXPIRED(x, y, z) \
|
|
((x)->deadline < 1 || \
|
|
((y)->tv_sec - (z)->tv_sec) > ((x)->deadline / 1000) || \
|
|
SND_CLONE_ELAPSED(y, z) > (x)->deadline)
|
|
|
|
/*
|
|
* snd_clone_gc() : Garbage collector for stalled, expired objects. Refer to
|
|
* clone.h for explanations on GC settings.
|
|
*/
|
|
int
|
|
snd_clone_gc(struct snd_clone *c)
|
|
{
|
|
struct snd_clone_entry *ce, *tce;
|
|
struct timespec now;
|
|
int pruned;
|
|
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
|
|
if (!(c->flags & SND_CLONE_GC_ENABLE) || c->size == 0)
|
|
return (0);
|
|
|
|
snd_timestamp(&now);
|
|
|
|
/*
|
|
* Bail out if the last clone handler was invoked below the deadline
|
|
* threshold.
|
|
*/
|
|
if ((c->flags & SND_CLONE_GC_EXPIRED) &&
|
|
!SND_CLONE_EXPIRED(c, &now, &c->tsp))
|
|
return (0);
|
|
|
|
pruned = 0;
|
|
|
|
/*
|
|
* Visit each object in reverse order. If the object is still being
|
|
* referenced by a valid open(), skip it. Look for expired objects
|
|
* and either revoke its clone invocation status or mercilessly
|
|
* throw it away.
|
|
*/
|
|
TAILQ_FOREACH_REVERSE_SAFE(ce, &c->head, link_head, link, tce) {
|
|
if (!(ce->flags & SND_CLONE_BUSY) &&
|
|
(!(ce->flags & SND_CLONE_INVOKE) ||
|
|
SND_CLONE_EXPIRED(c, &now, &ce->tsp))) {
|
|
if ((c->flags & SND_CLONE_GC_REVOKE) ||
|
|
ce->devt->si_threadcount != 0) {
|
|
ce->flags &= ~SND_CLONE_INVOKE;
|
|
ce->pid = -1;
|
|
} else {
|
|
TAILQ_REMOVE(&c->head, ce, link);
|
|
destroy_dev(ce->devt);
|
|
free(ce, M_DEVBUF);
|
|
c->size--;
|
|
}
|
|
pruned++;
|
|
}
|
|
}
|
|
|
|
/* return total pruned objects */
|
|
return (pruned);
|
|
}
|
|
|
|
void
|
|
snd_clone_destroy(struct snd_clone *c)
|
|
{
|
|
struct snd_clone_entry *ce, *tmp;
|
|
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
|
|
ce = TAILQ_FIRST(&c->head);
|
|
while (ce != NULL) {
|
|
tmp = TAILQ_NEXT(ce, link);
|
|
if (ce->devt != NULL)
|
|
destroy_dev(ce->devt);
|
|
free(ce, M_DEVBUF);
|
|
ce = tmp;
|
|
}
|
|
|
|
free(c, M_DEVBUF);
|
|
}
|
|
|
|
/*
|
|
* snd_clone_acquire() : The vital part of concurrency management. Must be
|
|
* called somewhere at the beginning of open() handler. ENODEV is not really
|
|
* fatal since it just tell the caller that this is not cloned stuff.
|
|
* EBUSY is *real*, don't forget that!
|
|
*/
|
|
int
|
|
snd_clone_acquire(struct cdev *dev)
|
|
{
|
|
struct snd_clone_entry *ce;
|
|
|
|
SND_CLONE_ASSERT(dev != NULL, ("NULL dev"));
|
|
|
|
ce = dev->si_drv2;
|
|
if (ce == NULL)
|
|
return (ENODEV);
|
|
|
|
SND_CLONE_ASSERT(ce->parent != NULL, ("NULL parent"));
|
|
|
|
ce->flags &= ~SND_CLONE_INVOKE;
|
|
|
|
if (ce->flags & SND_CLONE_BUSY)
|
|
return (EBUSY);
|
|
|
|
ce->flags |= SND_CLONE_BUSY;
|
|
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* snd_clone_release() : Release busy status. Must be called somewhere at
|
|
* the end of close() handler, or somewhere after fail open().
|
|
*/
|
|
int
|
|
snd_clone_release(struct cdev *dev)
|
|
{
|
|
struct snd_clone_entry *ce;
|
|
|
|
SND_CLONE_ASSERT(dev != NULL, ("NULL dev"));
|
|
|
|
ce = dev->si_drv2;
|
|
if (ce == NULL)
|
|
return (ENODEV);
|
|
|
|
SND_CLONE_ASSERT(ce->parent != NULL, ("NULL parent"));
|
|
|
|
ce->flags &= ~SND_CLONE_INVOKE;
|
|
|
|
if (!(ce->flags & SND_CLONE_BUSY))
|
|
return (EBADF);
|
|
|
|
ce->flags &= ~SND_CLONE_BUSY;
|
|
ce->pid = -1;
|
|
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* snd_clone_ref/unref() : Garbage collector reference counter. To make
|
|
* garbage collector run automatically, the sequence must be something like
|
|
* this (both in open() and close() handlers):
|
|
*
|
|
* open() - 1) snd_clone_acquire()
|
|
* 2) .... check check ... if failed, snd_clone_release()
|
|
* 3) Success. Call snd_clone_ref()
|
|
*
|
|
* close() - 1) .... check check check ....
|
|
* 2) Success. snd_clone_release()
|
|
* 3) snd_clone_unref() . Garbage collector will run at this point
|
|
* if this is the last referenced object.
|
|
*/
|
|
int
|
|
snd_clone_ref(struct cdev *dev)
|
|
{
|
|
struct snd_clone_entry *ce;
|
|
struct snd_clone *c;
|
|
|
|
SND_CLONE_ASSERT(dev != NULL, ("NULL dev"));
|
|
|
|
ce = dev->si_drv2;
|
|
if (ce == NULL)
|
|
return (0);
|
|
|
|
c = ce->parent;
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL parent"));
|
|
SND_CLONE_ASSERT(c->refcount >= 0, ("refcount < 0"));
|
|
|
|
return (++c->refcount);
|
|
}
|
|
|
|
int
|
|
snd_clone_unref(struct cdev *dev)
|
|
{
|
|
struct snd_clone_entry *ce;
|
|
struct snd_clone *c;
|
|
|
|
SND_CLONE_ASSERT(dev != NULL, ("NULL dev"));
|
|
|
|
ce = dev->si_drv2;
|
|
if (ce == NULL)
|
|
return (0);
|
|
|
|
c = ce->parent;
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL parent"));
|
|
SND_CLONE_ASSERT(c->refcount > 0, ("refcount <= 0"));
|
|
|
|
c->refcount--;
|
|
|
|
/*
|
|
* Run automatic garbage collector, if needed.
|
|
*/
|
|
if ((c->flags & SND_CLONE_GC_UNREF) &&
|
|
(!(c->flags & SND_CLONE_GC_LASTREF) ||
|
|
(c->refcount == 0 && (c->flags & SND_CLONE_GC_LASTREF))))
|
|
(void)snd_clone_gc(c);
|
|
|
|
return (c->refcount);
|
|
}
|
|
|
|
void
|
|
snd_clone_register(struct snd_clone_entry *ce, struct cdev *dev)
|
|
{
|
|
SND_CLONE_ASSERT(ce != NULL, ("NULL snd_clone_entry"));
|
|
SND_CLONE_ASSERT(dev != NULL, ("NULL dev"));
|
|
SND_CLONE_ASSERT(dev->si_drv2 == NULL, ("dev->si_drv2 not NULL"));
|
|
SND_CLONE_ASSERT((ce->flags & SND_CLONE_ALLOC) == SND_CLONE_ALLOC,
|
|
("invalid clone alloc flags=0x%08x", ce->flags));
|
|
SND_CLONE_ASSERT(ce->devt == NULL, ("ce->devt not NULL"));
|
|
SND_CLONE_ASSERT(ce->unit == dev2unit(dev),
|
|
("invalid unit ce->unit=0x%08x dev2unit=0x%08x",
|
|
ce->unit, dev2unit(dev)));
|
|
|
|
SND_CLONE_ASSERT(ce->parent != NULL, ("NULL parent"));
|
|
|
|
dev->si_drv2 = ce;
|
|
ce->devt = dev;
|
|
ce->flags &= ~SND_CLONE_ALLOC;
|
|
ce->flags |= SND_CLONE_INVOKE;
|
|
}
|
|
|
|
struct snd_clone_entry *
|
|
snd_clone_alloc(struct snd_clone *c, struct cdev **dev, int *unit, int tmask)
|
|
{
|
|
struct snd_clone_entry *ce, *after, *bce, *cce, *nce, *tce;
|
|
struct timespec now;
|
|
int cunit, allocunit;
|
|
pid_t curpid;
|
|
|
|
SND_CLONE_ASSERT(c != NULL, ("NULL snd_clone"));
|
|
SND_CLONE_ASSERT(dev != NULL, ("NULL dev pointer"));
|
|
SND_CLONE_ASSERT((c->typemask & tmask) == tmask,
|
|
("invalid tmask: typemask=0x%08x tmask=0x%08x",
|
|
c->typemask, tmask));
|
|
SND_CLONE_ASSERT(unit != NULL, ("NULL unit pointer"));
|
|
SND_CLONE_ASSERT(*unit == -1 || !(*unit & (c->typemask | tmask)),
|
|
("typemask collision: typemask=0x%08x tmask=0x%08x *unit=%d",
|
|
c->typemask, tmask, *unit));
|
|
|
|
if (!(c->flags & SND_CLONE_ENABLE) ||
|
|
(*unit != -1 && *unit > c->maxunit))
|
|
return (NULL);
|
|
|
|
ce = NULL;
|
|
after = NULL;
|
|
bce = NULL; /* "b"usy candidate */
|
|
cce = NULL; /* "c"urthread/proc candidate */
|
|
nce = NULL; /* "n"ull, totally unbusy candidate */
|
|
tce = NULL; /* Last "t"ry candidate */
|
|
cunit = 0;
|
|
allocunit = (*unit == -1) ? 0 : *unit;
|
|
curpid = curthread->td_proc->p_pid;
|
|
|
|
snd_timestamp(&now);
|
|
|
|
TAILQ_FOREACH(ce, &c->head, link) {
|
|
/*
|
|
* Sort incrementally according to device type.
|
|
*/
|
|
if (tmask > (ce->unit & c->typemask)) {
|
|
if (cunit == 0)
|
|
after = ce;
|
|
continue;
|
|
} else if (tmask < (ce->unit & c->typemask))
|
|
break;
|
|
|
|
/*
|
|
* Shoot.. this is where the grumpiness begin. Just
|
|
* return immediately.
|
|
*/
|
|
if (*unit != -1 && *unit == (ce->unit & ~tmask))
|
|
goto snd_clone_alloc_out;
|
|
|
|
cunit++;
|
|
/*
|
|
* Simmilar device type. Sort incrementally according
|
|
* to allocation unit. While here, look for free slot
|
|
* and possible collision for new / future allocation.
|
|
*/
|
|
if (*unit == -1 && (ce->unit & ~tmask) == allocunit)
|
|
allocunit++;
|
|
if ((ce->unit & ~tmask) < allocunit)
|
|
after = ce;
|
|
/*
|
|
* Clone logic:
|
|
* 1. Look for non busy, but keep track of the best
|
|
* possible busy cdev.
|
|
* 2. Look for the best (oldest referenced) entry that is
|
|
* in a same process / thread.
|
|
* 3. Look for the best (oldest referenced), absolute free
|
|
* entry.
|
|
* 4. Lastly, look for the best (oldest referenced)
|
|
* any entries that doesn't fit with anything above.
|
|
*/
|
|
if (ce->flags & SND_CLONE_BUSY) {
|
|
if (ce->devt != NULL && (bce == NULL ||
|
|
timespeccmp(&ce->tsp, &bce->tsp, <)))
|
|
bce = ce;
|
|
continue;
|
|
}
|
|
if (ce->pid == curpid &&
|
|
(cce == NULL || timespeccmp(&ce->tsp, &cce->tsp, <)))
|
|
cce = ce;
|
|
else if (!(ce->flags & SND_CLONE_INVOKE) &&
|
|
(nce == NULL || timespeccmp(&ce->tsp, &nce->tsp, <)))
|
|
nce = ce;
|
|
else if (tce == NULL || timespeccmp(&ce->tsp, &tce->tsp, <))
|
|
tce = ce;
|
|
}
|
|
if (*unit != -1)
|
|
goto snd_clone_alloc_new;
|
|
else if (cce != NULL) {
|
|
/* Same proc entry found, go for it */
|
|
ce = cce;
|
|
goto snd_clone_alloc_out;
|
|
} else if (nce != NULL) {
|
|
/*
|
|
* Next, try absolute free entry. If the calculated
|
|
* allocunit is smaller, create new entry instead.
|
|
*/
|
|
if (allocunit < (nce->unit & ~tmask))
|
|
goto snd_clone_alloc_new;
|
|
ce = nce;
|
|
goto snd_clone_alloc_out;
|
|
} else if (allocunit > c->maxunit) {
|
|
/*
|
|
* Maximum allowable unit reached. Try returning any
|
|
* available cdev and hope for the best. If the lookup is
|
|
* done for things like stat(), mtime() etc. , things should
|
|
* be ok. Otherwise, open() handler should do further checks
|
|
* and decide whether to return correct error code or not.
|
|
*/
|
|
if (tce != NULL) {
|
|
ce = tce;
|
|
goto snd_clone_alloc_out;
|
|
} else if (bce != NULL) {
|
|
ce = bce;
|
|
goto snd_clone_alloc_out;
|
|
}
|
|
return (NULL);
|
|
}
|
|
|
|
snd_clone_alloc_new:
|
|
/*
|
|
* No free entries found, and we still haven't reached maximum
|
|
* allowable units. Allocate, setup a minimal unique entry with busy
|
|
* status so nobody will monkey on this new entry. Unit magic is set
|
|
* right here to avoid collision with other contesting handler.
|
|
* The caller must be carefull here to maintain its own
|
|
* synchronization, as long as it will not conflict with malloc(9)
|
|
* operations.
|
|
*
|
|
* That said, go figure.
|
|
*/
|
|
ce = malloc(sizeof(*ce), M_DEVBUF,
|
|
((c->flags & SND_CLONE_WAITOK) ? M_WAITOK : M_NOWAIT) | M_ZERO);
|
|
if (ce == NULL) {
|
|
if (*unit != -1)
|
|
return (NULL);
|
|
/*
|
|
* We're being dense, ignorance is bliss,
|
|
* Super Regulatory Measure (TM).. TRY AGAIN!
|
|
*/
|
|
if (nce != NULL) {
|
|
ce = nce;
|
|
goto snd_clone_alloc_out;
|
|
} else if (tce != NULL) {
|
|
ce = tce;
|
|
goto snd_clone_alloc_out;
|
|
} else if (bce != NULL) {
|
|
ce = bce;
|
|
goto snd_clone_alloc_out;
|
|
}
|
|
return (NULL);
|
|
}
|
|
/* Setup new entry */
|
|
ce->parent = c;
|
|
ce->unit = tmask | allocunit;
|
|
ce->pid = curpid;
|
|
ce->tsp = now;
|
|
ce->flags |= SND_CLONE_ALLOC;
|
|
if (after != NULL) {
|
|
TAILQ_INSERT_AFTER(&c->head, after, ce, link);
|
|
} else {
|
|
TAILQ_INSERT_HEAD(&c->head, ce, link);
|
|
}
|
|
c->size++;
|
|
c->tsp = now;
|
|
/*
|
|
* Save new allocation unit for caller which will be used
|
|
* by make_dev().
|
|
*/
|
|
*unit = allocunit;
|
|
|
|
return (ce);
|
|
|
|
snd_clone_alloc_out:
|
|
/*
|
|
* Set, mark, timestamp the entry if this is a truly free entry.
|
|
* Leave busy entry alone.
|
|
*/
|
|
if (!(ce->flags & SND_CLONE_BUSY)) {
|
|
ce->pid = curpid;
|
|
ce->tsp = now;
|
|
ce->flags |= SND_CLONE_INVOKE;
|
|
}
|
|
c->tsp = now;
|
|
*dev = ce->devt;
|
|
|
|
return (NULL);
|
|
}
|