6a408cbd71
Give DEVFS a proper inode called struct cdev_priv. It is important to keep in mind that this "inode" is shared between all DEVFS mountpoints, therefore it is protected by the global device mutex. Link the cdev_priv's into a list, protected by the global device mutex. Keep track of each cdev_priv's state with a flag bit and of references from mountpoints with a dedicated usecount. Reap the benefits of much improved kernel memory allocator and the generally better defined device driver APIs to get rid of the tables of pointers + serial numbers, their overflow tables, the atomics to muck about in them and all the trouble that resulted in. This makes RAM the only limit on how many devices we can have. The cdev_priv is actually a super struct containing the normal cdev as the "public" part, and therefore allocation and freeing has moved to devfs_devs.c from kern_conf.c. The overall responsibility is (to be) split such that kern/kern_conf.c is the stuff that deals with drivers and struct cdev and fs/devfs handles filesystems and struct cdev_priv and their private liason exposed only in devfs_int.h. Move the inode number from cdev to cdev_priv and allocate inode numbers properly with unr. Local dirents in the mountpoints (directories, symlinks) allocate inodes from the same pool to guarantee against overlaps. Various other fields are going to migrate from cdev to cdev_priv in the future in order to hide them. A few fields may migrate from devfs_dirent to cdev_priv as well. Protect the DEVFS mountpoint with an sx lock instead of lockmgr, this lock also protects the directory tree of the mountpoint. Give each mountpoint a unique integer index, allocated with unr. Use it into an array of devfs_dirent pointers in each cdev_priv. Initially the array points to a single element also inside cdev_priv, but as more devfs instances are mounted, the array is extended with malloc(9) as necessary when the filesystem populates its directory tree. Retire the cdev alias lists, the cdev_priv now know about all the relevant devfs_dirents (and their vnodes) and devfs_revoke() will pick them up from there. We still spelunk into other mountpoints and fondle their data without 100% good locking. It may make better sense to vector the revoke event into the tty code and there do a destroy_dev/make_dev on the tty's devices, but that's for further study. Lots of shuffling of stuff and churn of bits for no good reason[2]. XXX: There is still nothing preventing the dev_clone EVENTHANDLER from being invoked at the same time in two devfs mountpoints. It is not obvious what the best course of action is here. XXX: comment out an if statement that lost its body, until I can find out what should go there so it doesn't do damage in the meantime. XXX: Leave in a few extra malloc types and KASSERTS to help track down any remaining issues. Much testing provided by: Kris Much confusion caused by (races in): md(4) [1] You are not supposed to understand anything past this point. [2] This line should simplify life for the peanut gallery.
883 lines
19 KiB
C
883 lines
19 KiB
C
/*-
|
|
* Copyright (c) 1999-2002 Poul-Henning Kamp
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/kernel.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/bio.h>
|
|
#include <sys/lock.h>
|
|
#include <sys/mutex.h>
|
|
#include <sys/module.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/conf.h>
|
|
#include <sys/vnode.h>
|
|
#include <sys/queue.h>
|
|
#include <sys/poll.h>
|
|
#include <sys/ctype.h>
|
|
#include <sys/tty.h>
|
|
#include <sys/ucred.h>
|
|
#include <machine/stdarg.h>
|
|
|
|
#include <fs/devfs/devfs_int.h>
|
|
|
|
static MALLOC_DEFINE(M_DEVT, "cdev", "cdev storage");
|
|
|
|
struct mtx devmtx;
|
|
static void destroy_devl(struct cdev *dev);
|
|
static struct cdev *make_dev_credv(struct cdevsw *devsw, int minornr,
|
|
struct ucred *cr, uid_t uid, gid_t gid, int mode, const char *fmt,
|
|
va_list ap);
|
|
|
|
void
|
|
dev_lock(void)
|
|
{
|
|
if (!mtx_initialized(&devmtx))
|
|
mtx_init(&devmtx, "cdev", NULL, MTX_DEF);
|
|
mtx_lock(&devmtx);
|
|
}
|
|
|
|
void
|
|
dev_unlock(void)
|
|
{
|
|
|
|
mtx_unlock(&devmtx);
|
|
}
|
|
|
|
void
|
|
dev_ref(struct cdev *dev)
|
|
{
|
|
|
|
mtx_assert(&devmtx, MA_NOTOWNED);
|
|
mtx_lock(&devmtx);
|
|
dev->si_refcount++;
|
|
mtx_unlock(&devmtx);
|
|
}
|
|
|
|
void
|
|
dev_refl(struct cdev *dev)
|
|
{
|
|
|
|
mtx_assert(&devmtx, MA_OWNED);
|
|
dev->si_refcount++;
|
|
}
|
|
|
|
void
|
|
dev_rel(struct cdev *dev)
|
|
{
|
|
int flag = 0;
|
|
|
|
mtx_assert(&devmtx, MA_NOTOWNED);
|
|
dev_lock();
|
|
dev->si_refcount--;
|
|
KASSERT(dev->si_refcount >= 0,
|
|
("dev_rel(%s) gave negative count", devtoname(dev)));
|
|
#if 0
|
|
if (dev->si_usecount == 0 &&
|
|
(dev->si_flags & SI_CHEAPCLONE) && (dev->si_flags & SI_NAMED))
|
|
;
|
|
else
|
|
#endif
|
|
if (dev->si_devsw == NULL && dev->si_refcount == 0) {
|
|
LIST_REMOVE(dev, si_list);
|
|
flag = 1;
|
|
}
|
|
dev_unlock();
|
|
if (flag)
|
|
devfs_free(dev);
|
|
}
|
|
|
|
struct cdevsw *
|
|
dev_refthread(struct cdev *dev)
|
|
{
|
|
struct cdevsw *csw;
|
|
|
|
mtx_assert(&devmtx, MA_NOTOWNED);
|
|
dev_lock();
|
|
csw = dev->si_devsw;
|
|
if (csw != NULL)
|
|
dev->si_threadcount++;
|
|
dev_unlock();
|
|
return (csw);
|
|
}
|
|
|
|
void
|
|
dev_relthread(struct cdev *dev)
|
|
{
|
|
|
|
mtx_assert(&devmtx, MA_NOTOWNED);
|
|
dev_lock();
|
|
dev->si_threadcount--;
|
|
dev_unlock();
|
|
}
|
|
|
|
int
|
|
nullop(void)
|
|
{
|
|
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
eopnotsupp(void)
|
|
{
|
|
|
|
return (EOPNOTSUPP);
|
|
}
|
|
|
|
static int
|
|
enxio(void)
|
|
{
|
|
return (ENXIO);
|
|
}
|
|
|
|
static int
|
|
enodev(void)
|
|
{
|
|
return (ENODEV);
|
|
}
|
|
|
|
/* Define a dead_cdevsw for use when devices leave unexpectedly. */
|
|
|
|
#define dead_open (d_open_t *)enxio
|
|
#define dead_close (d_close_t *)enxio
|
|
#define dead_read (d_read_t *)enxio
|
|
#define dead_write (d_write_t *)enxio
|
|
#define dead_ioctl (d_ioctl_t *)enxio
|
|
#define dead_poll (d_poll_t *)enodev
|
|
#define dead_mmap (d_mmap_t *)enodev
|
|
|
|
static void
|
|
dead_strategy(struct bio *bp)
|
|
{
|
|
|
|
biofinish(bp, NULL, ENXIO);
|
|
}
|
|
|
|
#define dead_dump (dumper_t *)enxio
|
|
#define dead_kqfilter (d_kqfilter_t *)enxio
|
|
|
|
static struct cdevsw dead_cdevsw = {
|
|
.d_version = D_VERSION,
|
|
.d_flags = D_NEEDGIANT, /* XXX: does dead_strategy need this ? */
|
|
.d_open = dead_open,
|
|
.d_close = dead_close,
|
|
.d_read = dead_read,
|
|
.d_write = dead_write,
|
|
.d_ioctl = dead_ioctl,
|
|
.d_poll = dead_poll,
|
|
.d_mmap = dead_mmap,
|
|
.d_strategy = dead_strategy,
|
|
.d_name = "dead",
|
|
.d_dump = dead_dump,
|
|
.d_kqfilter = dead_kqfilter
|
|
};
|
|
|
|
/* Default methods if driver does not specify method */
|
|
|
|
#define null_open (d_open_t *)nullop
|
|
#define null_close (d_close_t *)nullop
|
|
#define no_read (d_read_t *)enodev
|
|
#define no_write (d_write_t *)enodev
|
|
#define no_ioctl (d_ioctl_t *)enodev
|
|
#define no_mmap (d_mmap_t *)enodev
|
|
#define no_kqfilter (d_kqfilter_t *)enodev
|
|
|
|
static void
|
|
no_strategy(struct bio *bp)
|
|
{
|
|
|
|
biofinish(bp, NULL, ENODEV);
|
|
}
|
|
|
|
static int
|
|
no_poll(struct cdev *dev __unused, int events, struct thread *td __unused)
|
|
{
|
|
/*
|
|
* Return true for read/write. If the user asked for something
|
|
* special, return POLLNVAL, so that clients have a way of
|
|
* determining reliably whether or not the extended
|
|
* functionality is present without hard-coding knowledge
|
|
* of specific filesystem implementations.
|
|
* Stay in sync with vop_nopoll().
|
|
*/
|
|
if (events & ~POLLSTANDARD)
|
|
return (POLLNVAL);
|
|
|
|
return (events & (POLLIN | POLLOUT | POLLRDNORM | POLLWRNORM));
|
|
}
|
|
|
|
#define no_dump (dumper_t *)enodev
|
|
|
|
static int
|
|
giant_open(struct cdev *dev, int oflags, int devtype, struct thread *td)
|
|
{
|
|
int retval;
|
|
|
|
mtx_lock(&Giant);
|
|
retval = dev->si_devsw->d_gianttrick->
|
|
d_open(dev, oflags, devtype, td);
|
|
mtx_unlock(&Giant);
|
|
return (retval);
|
|
}
|
|
|
|
static int
|
|
giant_fdopen(struct cdev *dev, int oflags, struct thread *td, int fdidx)
|
|
{
|
|
int retval;
|
|
|
|
mtx_lock(&Giant);
|
|
retval = dev->si_devsw->d_gianttrick->
|
|
d_fdopen(dev, oflags, td, fdidx);
|
|
mtx_unlock(&Giant);
|
|
return (retval);
|
|
}
|
|
|
|
static int
|
|
giant_close(struct cdev *dev, int fflag, int devtype, struct thread *td)
|
|
{
|
|
int retval;
|
|
|
|
mtx_lock(&Giant);
|
|
retval = dev->si_devsw->d_gianttrick->
|
|
d_close(dev, fflag, devtype, td);
|
|
mtx_unlock(&Giant);
|
|
return (retval);
|
|
}
|
|
|
|
static void
|
|
giant_strategy(struct bio *bp)
|
|
{
|
|
|
|
mtx_lock(&Giant);
|
|
bp->bio_dev->si_devsw->d_gianttrick->
|
|
d_strategy(bp);
|
|
mtx_unlock(&Giant);
|
|
}
|
|
|
|
static int
|
|
giant_ioctl(struct cdev *dev, u_long cmd, caddr_t data, int fflag, struct thread *td)
|
|
{
|
|
int retval;
|
|
|
|
mtx_lock(&Giant);
|
|
retval = dev->si_devsw->d_gianttrick->
|
|
d_ioctl(dev, cmd, data, fflag, td);
|
|
mtx_unlock(&Giant);
|
|
return (retval);
|
|
}
|
|
|
|
static int
|
|
giant_read(struct cdev *dev, struct uio *uio, int ioflag)
|
|
{
|
|
int retval;
|
|
|
|
mtx_lock(&Giant);
|
|
retval = dev->si_devsw->d_gianttrick->
|
|
d_read(dev, uio, ioflag);
|
|
mtx_unlock(&Giant);
|
|
return (retval);
|
|
}
|
|
|
|
static int
|
|
giant_write(struct cdev *dev, struct uio *uio, int ioflag)
|
|
{
|
|
int retval;
|
|
|
|
mtx_lock(&Giant);
|
|
retval = dev->si_devsw->d_gianttrick->
|
|
d_write(dev, uio, ioflag);
|
|
mtx_unlock(&Giant);
|
|
return (retval);
|
|
}
|
|
|
|
static int
|
|
giant_poll(struct cdev *dev, int events, struct thread *td)
|
|
{
|
|
int retval;
|
|
|
|
mtx_lock(&Giant);
|
|
retval = dev->si_devsw->d_gianttrick->
|
|
d_poll(dev, events, td);
|
|
mtx_unlock(&Giant);
|
|
return (retval);
|
|
}
|
|
|
|
static int
|
|
giant_kqfilter(struct cdev *dev, struct knote *kn)
|
|
{
|
|
int retval;
|
|
|
|
mtx_lock(&Giant);
|
|
retval = dev->si_devsw->d_gianttrick->
|
|
d_kqfilter(dev, kn);
|
|
mtx_unlock(&Giant);
|
|
return (retval);
|
|
}
|
|
|
|
static int
|
|
giant_mmap(struct cdev *dev, vm_offset_t offset, vm_paddr_t *paddr, int nprot)
|
|
{
|
|
int retval;
|
|
|
|
mtx_lock(&Giant);
|
|
retval = dev->si_devsw->d_gianttrick->
|
|
d_mmap(dev, offset, paddr, nprot);
|
|
mtx_unlock(&Giant);
|
|
return (retval);
|
|
}
|
|
|
|
|
|
/*
|
|
* struct cdev * and u_dev_t primitives
|
|
*/
|
|
|
|
int
|
|
minor(struct cdev *x)
|
|
{
|
|
if (x == NULL)
|
|
return NODEV;
|
|
return(x->si_drv0 & MAXMINOR);
|
|
}
|
|
|
|
int
|
|
dev2unit(struct cdev *x)
|
|
{
|
|
|
|
if (x == NULL)
|
|
return NODEV;
|
|
return (minor2unit(minor(x)));
|
|
}
|
|
|
|
u_int
|
|
minor2unit(u_int _minor)
|
|
{
|
|
|
|
KASSERT((_minor & ~MAXMINOR) == 0, ("Illegal minor %x", _minor));
|
|
return ((_minor & 0xff) | ((_minor >> 8) & 0xffff00));
|
|
}
|
|
|
|
int
|
|
unit2minor(int unit)
|
|
{
|
|
|
|
KASSERT(unit <= 0xffffff, ("Invalid unit (%d) in unit2minor", unit));
|
|
return ((unit & 0xff) | ((unit << 8) & ~0xffff));
|
|
}
|
|
|
|
static struct cdev *
|
|
newdev(struct cdevsw *csw, int y, struct cdev *si)
|
|
{
|
|
struct cdev *si2;
|
|
dev_t udev;
|
|
|
|
mtx_assert(&devmtx, MA_OWNED);
|
|
udev = y;
|
|
LIST_FOREACH(si2, &csw->d_devs, si_list) {
|
|
if (si2->si_drv0 == udev) {
|
|
devfs_free(si);
|
|
return (si2);
|
|
}
|
|
}
|
|
si->si_drv0 = udev;
|
|
si->si_devsw = csw;
|
|
LIST_INSERT_HEAD(&csw->d_devs, si, si_list);
|
|
return (si);
|
|
}
|
|
|
|
int
|
|
uminor(dev_t dev)
|
|
{
|
|
return (dev & MAXMINOR);
|
|
}
|
|
|
|
int
|
|
umajor(dev_t dev)
|
|
{
|
|
return ((dev & ~MAXMINOR) >> 8);
|
|
}
|
|
|
|
static void
|
|
fini_cdevsw(struct cdevsw *devsw)
|
|
{
|
|
struct cdevsw *gt;
|
|
|
|
if (devsw->d_gianttrick != NULL) {
|
|
gt = devsw->d_gianttrick;
|
|
memcpy(devsw, gt, sizeof *devsw);
|
|
free(gt, M_DEVT);
|
|
devsw->d_gianttrick = NULL;
|
|
}
|
|
devsw->d_flags &= ~D_INIT;
|
|
}
|
|
|
|
static void
|
|
prep_cdevsw(struct cdevsw *devsw)
|
|
{
|
|
struct cdevsw *dsw2;
|
|
|
|
if (devsw->d_flags & D_NEEDGIANT)
|
|
dsw2 = malloc(sizeof *dsw2, M_DEVT, M_WAITOK);
|
|
else
|
|
dsw2 = NULL;
|
|
dev_lock();
|
|
|
|
if (devsw->d_version != D_VERSION_01) {
|
|
printf(
|
|
"WARNING: Device driver \"%s\" has wrong version %s\n",
|
|
devsw->d_name, "and is disabled. Recompile KLD module.");
|
|
devsw->d_open = dead_open;
|
|
devsw->d_close = dead_close;
|
|
devsw->d_read = dead_read;
|
|
devsw->d_write = dead_write;
|
|
devsw->d_ioctl = dead_ioctl;
|
|
devsw->d_poll = dead_poll;
|
|
devsw->d_mmap = dead_mmap;
|
|
devsw->d_strategy = dead_strategy;
|
|
devsw->d_dump = dead_dump;
|
|
devsw->d_kqfilter = dead_kqfilter;
|
|
}
|
|
|
|
if (devsw->d_flags & D_TTY) {
|
|
if (devsw->d_ioctl == NULL) devsw->d_ioctl = ttyioctl;
|
|
if (devsw->d_read == NULL) devsw->d_read = ttyread;
|
|
if (devsw->d_write == NULL) devsw->d_write = ttywrite;
|
|
if (devsw->d_kqfilter == NULL) devsw->d_kqfilter = ttykqfilter;
|
|
if (devsw->d_poll == NULL) devsw->d_poll = ttypoll;
|
|
}
|
|
|
|
if (devsw->d_flags & D_NEEDGIANT) {
|
|
if (devsw->d_gianttrick == NULL) {
|
|
memcpy(dsw2, devsw, sizeof *dsw2);
|
|
devsw->d_gianttrick = dsw2;
|
|
} else
|
|
free(dsw2, M_DEVT);
|
|
}
|
|
|
|
#define FIXUP(member, noop, giant) \
|
|
do { \
|
|
if (devsw->member == NULL) { \
|
|
devsw->member = noop; \
|
|
} else if (devsw->d_flags & D_NEEDGIANT) \
|
|
devsw->member = giant; \
|
|
} \
|
|
while (0)
|
|
|
|
FIXUP(d_open, null_open, giant_open);
|
|
FIXUP(d_fdopen, NULL, giant_fdopen);
|
|
FIXUP(d_close, null_close, giant_close);
|
|
FIXUP(d_read, no_read, giant_read);
|
|
FIXUP(d_write, no_write, giant_write);
|
|
FIXUP(d_ioctl, no_ioctl, giant_ioctl);
|
|
FIXUP(d_poll, no_poll, giant_poll);
|
|
FIXUP(d_mmap, no_mmap, giant_mmap);
|
|
FIXUP(d_strategy, no_strategy, giant_strategy);
|
|
FIXUP(d_kqfilter, no_kqfilter, giant_kqfilter);
|
|
|
|
if (devsw->d_dump == NULL) devsw->d_dump = no_dump;
|
|
|
|
LIST_INIT(&devsw->d_devs);
|
|
|
|
devsw->d_flags |= D_INIT;
|
|
|
|
dev_unlock();
|
|
}
|
|
|
|
static struct cdev *
|
|
make_dev_credv(struct cdevsw *devsw, int minornr, struct ucred *cr, uid_t uid,
|
|
gid_t gid, int mode, const char *fmt, va_list ap)
|
|
{
|
|
struct cdev *dev;
|
|
int i;
|
|
|
|
KASSERT((minornr & ~MAXMINOR) == 0,
|
|
("Invalid minor (0x%x) in make_dev", minornr));
|
|
|
|
if (!(devsw->d_flags & D_INIT))
|
|
prep_cdevsw(devsw);
|
|
dev = devfs_alloc();
|
|
dev_lock();
|
|
dev = newdev(devsw, minornr, dev);
|
|
if (dev->si_flags & SI_CHEAPCLONE &&
|
|
dev->si_flags & SI_NAMED) {
|
|
/*
|
|
* This is allowed as it removes races and generally
|
|
* simplifies cloning devices.
|
|
* XXX: still ??
|
|
*/
|
|
dev_unlock();
|
|
return (dev);
|
|
}
|
|
KASSERT(!(dev->si_flags & SI_NAMED),
|
|
("make_dev() by driver %s on pre-existing device (min=%x, name=%s)",
|
|
devsw->d_name, minor(dev), devtoname(dev)));
|
|
|
|
i = vsnrprintf(dev->__si_namebuf, sizeof dev->__si_namebuf, 32, fmt, ap);
|
|
if (i > (sizeof dev->__si_namebuf - 1)) {
|
|
printf("WARNING: Device name truncated! (%s)\n",
|
|
dev->__si_namebuf);
|
|
}
|
|
|
|
dev->si_flags |= SI_NAMED;
|
|
if (cr != NULL)
|
|
dev->si_cred = crhold(cr);
|
|
else
|
|
dev->si_cred = NULL;
|
|
dev->si_uid = uid;
|
|
dev->si_gid = gid;
|
|
dev->si_mode = mode;
|
|
|
|
devfs_create(dev);
|
|
dev_unlock();
|
|
return (dev);
|
|
}
|
|
|
|
struct cdev *
|
|
make_dev(struct cdevsw *devsw, int minornr, uid_t uid, gid_t gid, int mode,
|
|
const char *fmt, ...)
|
|
{
|
|
struct cdev *dev;
|
|
va_list ap;
|
|
|
|
va_start(ap, fmt);
|
|
dev = make_dev_credv(devsw, minornr, NULL, uid, gid, mode, fmt, ap);
|
|
va_end(ap);
|
|
return (dev);
|
|
}
|
|
|
|
struct cdev *
|
|
make_dev_cred(struct cdevsw *devsw, int minornr, struct ucred *cr, uid_t uid,
|
|
gid_t gid, int mode, const char *fmt, ...)
|
|
{
|
|
struct cdev *dev;
|
|
va_list ap;
|
|
|
|
va_start(ap, fmt);
|
|
dev = make_dev_credv(devsw, minornr, cr, uid, gid, mode, fmt, ap);
|
|
va_end(ap);
|
|
|
|
return (dev);
|
|
}
|
|
|
|
static void
|
|
dev_dependsl(struct cdev *pdev, struct cdev *cdev)
|
|
{
|
|
|
|
cdev->si_parent = pdev;
|
|
cdev->si_flags |= SI_CHILD;
|
|
LIST_INSERT_HEAD(&pdev->si_children, cdev, si_siblings);
|
|
}
|
|
|
|
|
|
void
|
|
dev_depends(struct cdev *pdev, struct cdev *cdev)
|
|
{
|
|
|
|
dev_lock();
|
|
dev_dependsl(pdev, cdev);
|
|
dev_unlock();
|
|
}
|
|
|
|
struct cdev *
|
|
make_dev_alias(struct cdev *pdev, const char *fmt, ...)
|
|
{
|
|
struct cdev *dev;
|
|
va_list ap;
|
|
int i;
|
|
|
|
dev = devfs_alloc();
|
|
dev_lock();
|
|
dev->si_flags |= SI_ALIAS;
|
|
dev->si_flags |= SI_NAMED;
|
|
va_start(ap, fmt);
|
|
i = vsnrprintf(dev->__si_namebuf, sizeof dev->__si_namebuf, 32, fmt, ap);
|
|
if (i > (sizeof dev->__si_namebuf - 1)) {
|
|
printf("WARNING: Device name truncated! (%s)\n",
|
|
dev->__si_namebuf);
|
|
}
|
|
va_end(ap);
|
|
|
|
devfs_create(dev);
|
|
dev_unlock();
|
|
dev_depends(pdev, dev);
|
|
return (dev);
|
|
}
|
|
|
|
static void
|
|
destroy_devl(struct cdev *dev)
|
|
{
|
|
struct cdevsw *csw;
|
|
|
|
mtx_assert(&devmtx, MA_OWNED);
|
|
KASSERT(dev->si_flags & SI_NAMED,
|
|
("WARNING: Driver mistake: destroy_dev on %d\n", minor(dev)));
|
|
|
|
devfs_destroy(dev);
|
|
|
|
/* Remove name marking */
|
|
dev->si_flags &= ~SI_NAMED;
|
|
|
|
/* If we are a child, remove us from the parents list */
|
|
if (dev->si_flags & SI_CHILD) {
|
|
LIST_REMOVE(dev, si_siblings);
|
|
dev->si_flags &= ~SI_CHILD;
|
|
}
|
|
|
|
/* Kill our children */
|
|
while (!LIST_EMPTY(&dev->si_children))
|
|
destroy_devl(LIST_FIRST(&dev->si_children));
|
|
|
|
/* Remove from clone list */
|
|
if (dev->si_flags & SI_CLONELIST) {
|
|
LIST_REMOVE(dev, si_clone);
|
|
dev->si_flags &= ~SI_CLONELIST;
|
|
}
|
|
|
|
csw = dev->si_devsw;
|
|
dev->si_devsw = NULL; /* already NULL for SI_ALIAS */
|
|
while (csw != NULL && csw->d_purge != NULL && dev->si_threadcount) {
|
|
printf("Purging %lu threads from %s\n",
|
|
dev->si_threadcount, devtoname(dev));
|
|
csw->d_purge(dev);
|
|
msleep(csw, &devmtx, PRIBIO, "devprg", hz/10);
|
|
}
|
|
if (csw != NULL && csw->d_purge != NULL)
|
|
printf("All threads purged from %s\n", devtoname(dev));
|
|
|
|
dev->si_drv1 = 0;
|
|
dev->si_drv2 = 0;
|
|
bzero(&dev->__si_u, sizeof(dev->__si_u));
|
|
|
|
if (!(dev->si_flags & SI_ALIAS)) {
|
|
/* Remove from cdevsw list */
|
|
LIST_REMOVE(dev, si_list);
|
|
|
|
/* If cdevsw has no more struct cdev *'s, clean it */
|
|
if (LIST_EMPTY(&csw->d_devs))
|
|
fini_cdevsw(csw);
|
|
}
|
|
dev->si_flags &= ~SI_ALIAS;
|
|
|
|
if (dev->si_refcount > 0) {
|
|
LIST_INSERT_HEAD(&dead_cdevsw.d_devs, dev, si_list);
|
|
} else {
|
|
devfs_free(dev);
|
|
}
|
|
}
|
|
|
|
void
|
|
destroy_dev(struct cdev *dev)
|
|
{
|
|
|
|
dev_lock();
|
|
destroy_devl(dev);
|
|
dev_unlock();
|
|
}
|
|
|
|
const char *
|
|
devtoname(struct cdev *dev)
|
|
{
|
|
char *p;
|
|
struct cdevsw *csw;
|
|
int mynor;
|
|
|
|
if (dev->si_name[0] == '#' || dev->si_name[0] == '\0') {
|
|
p = dev->si_name;
|
|
csw = dev_refthread(dev);
|
|
if (csw != NULL) {
|
|
sprintf(p, "(%s)", csw->d_name);
|
|
dev_relthread(dev);
|
|
}
|
|
p += strlen(p);
|
|
mynor = minor(dev);
|
|
if (mynor < 0 || mynor > 255)
|
|
sprintf(p, "/%#x", (u_int)mynor);
|
|
else
|
|
sprintf(p, "/%d", mynor);
|
|
}
|
|
return (dev->si_name);
|
|
}
|
|
|
|
int
|
|
dev_stdclone(char *name, char **namep, const char *stem, int *unit)
|
|
{
|
|
int u, i;
|
|
|
|
i = strlen(stem);
|
|
if (bcmp(stem, name, i) != 0)
|
|
return (0);
|
|
if (!isdigit(name[i]))
|
|
return (0);
|
|
u = 0;
|
|
if (name[i] == '0' && isdigit(name[i+1]))
|
|
return (0);
|
|
while (isdigit(name[i])) {
|
|
u *= 10;
|
|
u += name[i++] - '0';
|
|
}
|
|
if (u > 0xffffff)
|
|
return (0);
|
|
*unit = u;
|
|
if (namep)
|
|
*namep = &name[i];
|
|
if (name[i])
|
|
return (2);
|
|
return (1);
|
|
}
|
|
|
|
/*
|
|
* Helper functions for cloning device drivers.
|
|
*
|
|
* The objective here is to make it unnecessary for the device drivers to
|
|
* use rman or similar to manage their unit number space. Due to the way
|
|
* we do "on-demand" devices, using rman or other "private" methods
|
|
* will be very tricky to lock down properly once we lock down this file.
|
|
*
|
|
* Instead we give the drivers these routines which puts the struct cdev *'s
|
|
* that are to be managed on their own list, and gives the driver the ability
|
|
* to ask for the first free unit number or a given specified unit number.
|
|
*
|
|
* In addition these routines support paired devices (pty, nmdm and similar)
|
|
* by respecting a number of "flag" bits in the minor number.
|
|
*
|
|
*/
|
|
|
|
struct clonedevs {
|
|
LIST_HEAD(,cdev) head;
|
|
};
|
|
|
|
void
|
|
clone_setup(struct clonedevs **cdp)
|
|
{
|
|
|
|
*cdp = malloc(sizeof **cdp, M_DEVBUF, M_WAITOK | M_ZERO);
|
|
LIST_INIT(&(*cdp)->head);
|
|
}
|
|
|
|
int
|
|
clone_create(struct clonedevs **cdp, struct cdevsw *csw, int *up, struct cdev **dp, u_int extra)
|
|
{
|
|
struct clonedevs *cd;
|
|
struct cdev *dev, *ndev, *dl, *de;
|
|
int unit, low, u;
|
|
|
|
KASSERT(*cdp != NULL,
|
|
("clone_setup() not called in driver \"%s\"", csw->d_name));
|
|
KASSERT(!(extra & CLONE_UNITMASK),
|
|
("Illegal extra bits (0x%x) in clone_create", extra));
|
|
KASSERT(*up <= CLONE_UNITMASK,
|
|
("Too high unit (0x%x) in clone_create", *up));
|
|
|
|
if (!(csw->d_flags & D_INIT))
|
|
prep_cdevsw(csw);
|
|
|
|
/*
|
|
* Search the list for a lot of things in one go:
|
|
* A preexisting match is returned immediately.
|
|
* The lowest free unit number if we are passed -1, and the place
|
|
* in the list where we should insert that new element.
|
|
* The place to insert a specified unit number, if applicable
|
|
* the end of the list.
|
|
*/
|
|
unit = *up;
|
|
ndev = devfs_alloc();
|
|
dev_lock();
|
|
low = extra;
|
|
de = dl = NULL;
|
|
cd = *cdp;
|
|
LIST_FOREACH(dev, &cd->head, si_clone) {
|
|
KASSERT(dev->si_flags & SI_CLONELIST,
|
|
("Dev %p(%s) should be on clonelist", dev, dev->si_name));
|
|
u = dev2unit(dev);
|
|
if (u == (unit | extra)) {
|
|
*dp = dev;
|
|
devfs_free(ndev);
|
|
dev_unlock();
|
|
return (0);
|
|
}
|
|
if (unit == -1 && u == low) {
|
|
low++;
|
|
de = dev;
|
|
continue;
|
|
}
|
|
if (u > (unit | extra)) {
|
|
dl = dev;
|
|
break;
|
|
}
|
|
}
|
|
if (unit == -1)
|
|
unit = low & CLONE_UNITMASK;
|
|
dev = newdev(csw, unit2minor(unit | extra), ndev);
|
|
if (dev->si_flags & SI_CLONELIST) {
|
|
printf("dev %p (%s) is on clonelist\n", dev, dev->si_name);
|
|
printf("unit=%d\n", unit);
|
|
LIST_FOREACH(dev, &cd->head, si_clone) {
|
|
printf("\t%p %s\n", dev, dev->si_name);
|
|
}
|
|
panic("foo");
|
|
}
|
|
KASSERT(!(dev->si_flags & SI_CLONELIST),
|
|
("Dev %p(%s) should not be on clonelist", dev, dev->si_name));
|
|
if (dl != NULL)
|
|
LIST_INSERT_BEFORE(dl, dev, si_clone);
|
|
else if (de != NULL)
|
|
LIST_INSERT_AFTER(de, dev, si_clone);
|
|
else
|
|
LIST_INSERT_HEAD(&cd->head, dev, si_clone);
|
|
dev->si_flags |= SI_CLONELIST;
|
|
*up = unit;
|
|
dev_unlock();
|
|
return (1);
|
|
}
|
|
|
|
/*
|
|
* Kill everything still on the list. The driver should already have
|
|
* disposed of any softc hung of the struct cdev *'s at this time.
|
|
*/
|
|
void
|
|
clone_cleanup(struct clonedevs **cdp)
|
|
{
|
|
struct cdev *dev, *tdev;
|
|
struct clonedevs *cd;
|
|
|
|
cd = *cdp;
|
|
if (cd == NULL)
|
|
return;
|
|
dev_lock();
|
|
LIST_FOREACH_SAFE(dev, &cd->head, si_clone, tdev) {
|
|
KASSERT(dev->si_flags & SI_CLONELIST,
|
|
("Dev %p(%s) should be on clonelist", dev, dev->si_name));
|
|
KASSERT(dev->si_flags & SI_NAMED,
|
|
("Driver has goofed in cloning underways udev %x", dev->si_drv0));
|
|
destroy_devl(dev);
|
|
}
|
|
dev_unlock();
|
|
free(cd, M_DEVBUF);
|
|
*cdp = NULL;
|
|
}
|