freebsd-dev/sys/dev/bhnd/bhndb/bhndb_subr.c
2020-09-01 21:43:05 +00:00

1666 lines
44 KiB
C

/*-
* Copyright (c) 2015-2016 Landon Fuller <landon@landonf.org>
* Copyright (c) 2017 The FreeBSD Foundation
* All rights reserved.
*
* Portions of this software were developed by Landon Fuller
* under sponsorship from the FreeBSD Foundation.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer,
* without modification.
* 2. Redistributions in binary form must reproduce at minimum a disclaimer
* similar to the "NO WARRANTY" disclaimer below ("Disclaimer") and any
* redistribution must be conditioned upon including a substantially
* similar Disclaimer requirement for further binary redistribution.
*
* NO WARRANTY
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTIBILITY
* AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL
* THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY,
* OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
* IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
* THE POSSIBILITY OF SUCH DAMAGES.
*/
#include <sys/cdefs.h>
__FBSDID("$FreeBSD$");
#include <sys/param.h>
#include <sys/kernel.h>
#include <sys/limits.h>
#include "bhndb_private.h"
#include "bhndbvar.h"
static int bhndb_dma_tag_create(device_t dev, bus_dma_tag_t parent_dmat,
const struct bhnd_dma_translation *translation,
bus_dma_tag_t *dmat);
/**
* Attach a BHND bridge device to @p parent.
*
* @param parent A parent PCI device.
* @param[out] bhndb On success, the probed and attached bhndb bridge device.
* @param unit The device unit number, or -1 to select the next available unit
* number.
*
* @retval 0 success
* @retval non-zero Failed to attach the bhndb device.
*/
int
bhndb_attach_bridge(device_t parent, device_t *bhndb, int unit)
{
int error;
*bhndb = device_add_child(parent, "bhndb", unit);
if (*bhndb == NULL)
return (ENXIO);
if (!(error = device_probe_and_attach(*bhndb)))
return (0);
if ((device_delete_child(parent, *bhndb)))
device_printf(parent, "failed to detach bhndb child\n");
return (error);
}
/*
* Call BHNDB_SUSPEND_RESOURCE() for all resources in @p rl.
*/
static void
bhndb_do_suspend_resources(device_t dev, struct resource_list *rl)
{
struct resource_list_entry *rle;
/* Suspend all child resources. */
STAILQ_FOREACH(rle, rl, link) {
/* Skip non-allocated resources */
if (rle->res == NULL)
continue;
BHNDB_SUSPEND_RESOURCE(device_get_parent(dev), dev, rle->type,
rle->res);
}
}
/**
* Helper function for implementing BUS_RESUME_CHILD() on bridged
* bhnd(4) buses.
*
* This implementation of BUS_RESUME_CHILD() uses BUS_GET_RESOURCE_LIST()
* to find the child's resources and call BHNDB_SUSPEND_RESOURCE() for all
* child resources, ensuring that the device's allocated bridge resources
* will be available to other devices during bus resumption.
*
* Before suspending any resources, @p child is suspended by
* calling bhnd_generic_suspend_child().
*
* If @p child is not a direct child of @p dev, suspension is delegated to
* the @p dev parent.
*/
int
bhnd_generic_br_suspend_child(device_t dev, device_t child)
{
struct resource_list *rl;
int error;
if (device_get_parent(child) != dev)
BUS_SUSPEND_CHILD(device_get_parent(dev), child);
if (device_is_suspended(child))
return (EBUSY);
/* Suspend the child device */
if ((error = bhnd_generic_suspend_child(dev, child)))
return (error);
/* Fetch the resource list. If none, there's nothing else to do */
rl = BUS_GET_RESOURCE_LIST(device_get_parent(child), child);
if (rl == NULL)
return (0);
/* Suspend all child resources. */
bhndb_do_suspend_resources(dev, rl);
return (0);
}
/**
* Helper function for implementing BUS_RESUME_CHILD() on bridged
* bhnd(4) bus devices.
*
* This implementation of BUS_RESUME_CHILD() uses BUS_GET_RESOURCE_LIST()
* to find the child's resources and call BHNDB_RESUME_RESOURCE() for all
* child resources, before delegating to bhnd_generic_resume_child().
*
* If resource resumption fails, @p child will not be resumed.
*
* If @p child is not a direct child of @p dev, suspension is delegated to
* the @p dev parent.
*/
int
bhnd_generic_br_resume_child(device_t dev, device_t child)
{
struct resource_list *rl;
struct resource_list_entry *rle;
int error;
if (device_get_parent(child) != dev)
BUS_RESUME_CHILD(device_get_parent(dev), child);
if (!device_is_suspended(child))
return (EBUSY);
/* Fetch the resource list. If none, there's nothing else to do */
rl = BUS_GET_RESOURCE_LIST(device_get_parent(child), child);
if (rl == NULL)
return (bhnd_generic_resume_child(dev, child));
/* Resume all resources */
STAILQ_FOREACH(rle, rl, link) {
/* Skip non-allocated resources */
if (rle->res == NULL)
continue;
error = BHNDB_RESUME_RESOURCE(device_get_parent(dev), dev,
rle->type, rle->res);
if (error) {
/* Put all resources back into a suspend state */
bhndb_do_suspend_resources(dev, rl);
return (error);
}
}
/* Now that all resources are resumed, resume child */
if ((error = bhnd_generic_resume_child(dev, child))) {
/* Put all resources back into a suspend state */
bhndb_do_suspend_resources(dev, rl);
}
return (error);
}
/**
* Find a host resource of @p type that maps the given range.
*
* @param hr The resource state to search.
* @param type The resource type to search for (see SYS_RES_*).
* @param start The start address of the range to search for.
* @param count The size of the range to search for.
*
* @retval resource the host resource containing the requested range.
* @retval NULL if no resource containing the requested range can be found.
*/
struct resource *
bhndb_host_resource_for_range(struct bhndb_host_resources *hr, int type,
rman_res_t start, rman_res_t count)
{
for (u_int i = 0; hr->resource_specs[i].type != -1; i++) {
struct resource *r = hr->resources[i];
if (hr->resource_specs[i].type != type)
continue;
/* Verify range */
if (rman_get_start(r) > start)
continue;
if (rman_get_end(r) < (start + count - 1))
continue;
return (r);
}
return (NULL);
}
/**
* Find a host resource of that matches the given register window definition.
*
* @param hr The resource state to search.
* @param win A register window definition.
*
* @retval resource the host resource corresponding to @p win.
* @retval NULL if no resource corresponding to @p win can be found.
*/
struct resource *
bhndb_host_resource_for_regwin(struct bhndb_host_resources *hr,
const struct bhndb_regwin *win)
{
const struct resource_spec *rspecs;
rspecs = hr->resource_specs;
for (u_int i = 0; rspecs[i].type != -1; i++) {
if (win->res.type != rspecs[i].type)
continue;
if (win->res.rid != rspecs[i].rid)
continue;
/* Found declared resource */
return (hr->resources[i]);
}
device_printf(hr->owner, "missing regwin resource spec "
"(type=%d, rid=%d)\n", win->res.type, win->res.rid);
return (NULL);
}
/**
* Allocate and initialize a new resource state structure.
*
* @param dev The bridge device.
* @param parent_dev The parent device from which host resources should be
* allocated.
* @param cfg The hardware configuration to be used.
*/
struct bhndb_resources *
bhndb_alloc_resources(device_t dev, device_t parent_dev,
const struct bhndb_hwcfg *cfg)
{
struct bhndb_resources *r;
const struct bhndb_regwin *win;
bus_size_t last_window_size;
int rnid;
int error;
bool free_ht_mem, free_br_mem, free_br_irq;
free_ht_mem = false;
free_br_mem = false;
free_br_irq = false;
r = malloc(sizeof(*r), M_BHND, M_NOWAIT|M_ZERO);
if (r == NULL)
return (NULL);
/* Basic initialization */
r->dev = dev;
r->cfg = cfg;
r->res = NULL;
r->min_prio = BHNDB_PRIORITY_NONE;
STAILQ_INIT(&r->bus_regions);
STAILQ_INIT(&r->bus_intrs);
mtx_init(&r->dw_steal_mtx, device_get_nameunit(dev),
"bhndb dwa_steal lock", MTX_SPIN);
/* Initialize host address space resource manager. */
r->ht_mem_rman.rm_start = 0;
r->ht_mem_rman.rm_end = ~0;
r->ht_mem_rman.rm_type = RMAN_ARRAY;
r->ht_mem_rman.rm_descr = "BHNDB host memory";
if ((error = rman_init(&r->ht_mem_rman))) {
device_printf(r->dev, "could not initialize ht_mem_rman\n");
goto failed;
}
free_ht_mem = true;
/* Initialize resource manager for the bridged address space. */
r->br_mem_rman.rm_start = 0;
r->br_mem_rman.rm_end = BUS_SPACE_MAXADDR_32BIT;
r->br_mem_rman.rm_type = RMAN_ARRAY;
r->br_mem_rman.rm_descr = "BHNDB bridged memory";
if ((error = rman_init(&r->br_mem_rman))) {
device_printf(r->dev, "could not initialize br_mem_rman\n");
goto failed;
}
free_br_mem = true;
error = rman_manage_region(&r->br_mem_rman, 0, BUS_SPACE_MAXADDR_32BIT);
if (error) {
device_printf(r->dev, "could not configure br_mem_rman\n");
goto failed;
}
/* Initialize resource manager for the bridged interrupt controller. */
r->br_irq_rman.rm_start = 0;
r->br_irq_rman.rm_end = RM_MAX_END;
r->br_irq_rman.rm_type = RMAN_ARRAY;
r->br_irq_rman.rm_descr = "BHNDB bridged interrupts";
if ((error = rman_init(&r->br_irq_rman))) {
device_printf(r->dev, "could not initialize br_irq_rman\n");
goto failed;
}
free_br_irq = true;
error = rman_manage_region(&r->br_irq_rman, 0, RM_MAX_END);
if (error) {
device_printf(r->dev, "could not configure br_irq_rman\n");
goto failed;
}
/* Fetch the dynamic regwin count and verify that it does not exceed
* what is representable via our freelist bitstring. */
r->dwa_count = bhndb_regwin_count(cfg->register_windows,
BHNDB_REGWIN_T_DYN);
if (r->dwa_count >= INT_MAX) {
device_printf(r->dev, "max dynamic regwin count exceeded\n");
goto failed;
}
/* Allocate the dynamic window allocation table. */
r->dw_alloc = malloc(sizeof(r->dw_alloc[0]) * r->dwa_count, M_BHND,
M_NOWAIT);
if (r->dw_alloc == NULL)
goto failed;
/* Allocate the dynamic window allocation freelist */
r->dwa_freelist = bit_alloc(r->dwa_count, M_BHND, M_NOWAIT);
if (r->dwa_freelist == NULL)
goto failed;
/* Initialize the dynamic window table */
rnid = 0;
last_window_size = 0;
for (win = cfg->register_windows;
win->win_type != BHNDB_REGWIN_T_INVALID; win++)
{
struct bhndb_dw_alloc *dwa;
/* Skip non-DYN windows */
if (win->win_type != BHNDB_REGWIN_T_DYN)
continue;
/* Validate the window size */
if (win->win_size == 0) {
device_printf(r->dev, "ignoring zero-length dynamic "
"register window\n");
continue;
} else if (last_window_size == 0) {
last_window_size = win->win_size;
} else if (last_window_size != win->win_size) {
/*
* No existing hardware should trigger this.
*
* If you run into this in the future, the dynamic
* window allocator and the resource priority system
* will need to be extended to support multiple register
* window allocation pools.
*/
device_printf(r->dev, "devices that vend multiple "
"dynamic register window sizes are not currently "
"supported\n");
goto failed;
}
dwa = &r->dw_alloc[rnid];
dwa->win = win;
dwa->parent_res = NULL;
dwa->rnid = rnid;
dwa->target = 0x0;
LIST_INIT(&dwa->refs);
rnid++;
}
/* Allocate host resources */
error = bhndb_alloc_host_resources(&r->res, dev, parent_dev, r->cfg);
if (error) {
device_printf(r->dev,
"could not allocate host resources on %s: %d\n",
device_get_nameunit(parent_dev), error);
goto failed;
}
/* Populate (and validate) parent resource references for all
* dynamic windows */
for (size_t i = 0; i < r->dwa_count; i++) {
struct bhndb_dw_alloc *dwa;
const struct bhndb_regwin *win;
dwa = &r->dw_alloc[i];
win = dwa->win;
/* Find and validate corresponding resource. */
dwa->parent_res = bhndb_host_resource_for_regwin(r->res, win);
if (dwa->parent_res == NULL) {
device_printf(r->dev, "no host resource found for %u "
"register window with offset %#jx and "
"size %#jx\n",
win->win_type,
(uintmax_t)win->win_offset,
(uintmax_t)win->win_size);
error = ENXIO;
goto failed;
}
if (rman_get_size(dwa->parent_res) < win->win_offset +
win->win_size)
{
device_printf(r->dev, "resource %d too small for "
"register window with offset %llx and size %llx\n",
rman_get_rid(dwa->parent_res),
(unsigned long long) win->win_offset,
(unsigned long long) win->win_size);
error = EINVAL;
goto failed;
}
}
/* Add allocated memory resources to our host memory resource manager */
for (u_int i = 0; r->res->resource_specs[i].type != -1; i++) {
struct resource *res;
/* skip non-memory resources */
if (r->res->resource_specs[i].type != SYS_RES_MEMORY)
continue;
/* add host resource to set of managed regions */
res = r->res->resources[i];
error = rman_manage_region(&r->ht_mem_rman,
rman_get_start(res), rman_get_end(res));
if (error) {
device_printf(r->dev,
"could not register host memory region with "
"ht_mem_rman: %d\n", error);
goto failed;
}
}
return (r);
failed:
if (free_ht_mem)
rman_fini(&r->ht_mem_rman);
if (free_br_mem)
rman_fini(&r->br_mem_rman);
if (free_br_irq)
rman_fini(&r->br_irq_rman);
if (r->dw_alloc != NULL)
free(r->dw_alloc, M_BHND);
if (r->dwa_freelist != NULL)
free(r->dwa_freelist, M_BHND);
if (r->res != NULL)
bhndb_release_host_resources(r->res);
mtx_destroy(&r->dw_steal_mtx);
free(r, M_BHND);
return (NULL);
}
/**
* Create a new DMA tag for the given @p translation.
*
* @param dev The bridge device.
* @param parent_dmat The parent DMA tag, or NULL if none.
* @param translation The DMA translation for which a DMA tag will
* be created.
* @param[out] dmat On success, the newly created DMA tag.
*
* @retval 0 success
* @retval non-zero if creating the new DMA tag otherwise fails, a regular
* unix error code will be returned.
*/
static int
bhndb_dma_tag_create(device_t dev, bus_dma_tag_t parent_dmat,
const struct bhnd_dma_translation *translation, bus_dma_tag_t *dmat)
{
bus_dma_tag_t translation_tag;
bhnd_addr_t dt_mask;
bus_addr_t lowaddr, highaddr;
bus_size_t maxsegsz;
int error;
highaddr = BUS_SPACE_MAXADDR;
maxsegsz = BUS_SPACE_MAXSIZE;
/* Determine full addressable mask */
dt_mask = (translation->addr_mask | translation->addrext_mask);
KASSERT(dt_mask != 0, ("DMA addr_mask invalid: %#jx",
(uintmax_t)dt_mask));
/* (addr_mask|addrext_mask) is our maximum supported address */
lowaddr = MIN(dt_mask, BUS_SPACE_MAXADDR);
/* Constrain to translation window size */
if (translation->addr_mask < maxsegsz)
maxsegsz = translation->addr_mask;
/* Create our DMA tag */
error = bus_dma_tag_create(parent_dmat,
1, 0, /* alignment, boundary */
lowaddr, highaddr,
NULL, NULL, /* filter, filterarg */
BUS_SPACE_MAXSIZE, 0, /* maxsize, nsegments */
maxsegsz, 0, /* maxsegsize, flags */
NULL, NULL, /* lockfunc, lockarg */
&translation_tag);
if (error) {
device_printf(dev, "failed to create bridge DMA tag: %d\n",
error);
return (error);
}
*dmat = translation_tag;
return (0);
}
/**
* Deallocate the given bridge resource structure and any associated resources.
*
* @param br Resource state to be deallocated.
*/
void
bhndb_free_resources(struct bhndb_resources *br)
{
struct bhndb_region *region, *r_next;
struct bhndb_dw_alloc *dwa;
struct bhndb_dw_rentry *dwr, *dwr_next;
struct bhndb_intr_handler *ih;
bool leaked_regions, leaked_intrs;
leaked_regions = false;
leaked_intrs = false;
/* No window regions may still be held */
if (!bhndb_dw_all_free(br)) {
for (int i = 0; i < br->dwa_count; i++) {
dwa = &br->dw_alloc[i];
/* Skip free dynamic windows */
if (bhndb_dw_is_free(br, dwa))
continue;
device_printf(br->dev,
"leaked dynamic register window %d\n", dwa->rnid);
leaked_regions = true;
}
}
/* There should be no interrupt handlers still registered */
STAILQ_FOREACH(ih, &br->bus_intrs, ih_link) {
device_printf(br->dev, "interrupt handler leaked %p\n",
ih->ih_cookiep);
}
if (leaked_intrs || leaked_regions) {
panic("leaked%s%s", leaked_intrs ? " active interrupts" : "",
leaked_regions ? " active register windows" : "");
}
/* Release host resources allocated through our parent. */
if (br->res != NULL)
bhndb_release_host_resources(br->res);
/* Clean up resource reservations */
for (size_t i = 0; i < br->dwa_count; i++) {
dwa = &br->dw_alloc[i];
LIST_FOREACH_SAFE(dwr, &dwa->refs, dw_link, dwr_next) {
LIST_REMOVE(dwr, dw_link);
free(dwr, M_BHND);
}
}
/* Release bus regions */
STAILQ_FOREACH_SAFE(region, &br->bus_regions, link, r_next) {
STAILQ_REMOVE(&br->bus_regions, region, bhndb_region, link);
free(region, M_BHND);
}
/* Release our resource managers */
rman_fini(&br->ht_mem_rman);
rman_fini(&br->br_mem_rman);
rman_fini(&br->br_irq_rman);
free(br->dw_alloc, M_BHND);
free(br->dwa_freelist, M_BHND);
mtx_destroy(&br->dw_steal_mtx);
free(br, M_BHND);
}
/**
* Allocate host bus resources defined by @p hwcfg.
*
* On success, the caller assumes ownership of the allocated host resources,
* which must be freed via bhndb_release_host_resources().
*
* @param[out] resources On success, the allocated host resources.
* @param dev The bridge device.
* @param parent_dev The parent device from which host resources
* should be allocated (e.g. via
* bus_alloc_resources()).
* @param hwcfg The hardware configuration defining the host
* resources to be allocated
*/
int
bhndb_alloc_host_resources(struct bhndb_host_resources **resources,
device_t dev, device_t parent_dev, const struct bhndb_hwcfg *hwcfg)
{
struct bhndb_host_resources *hr;
const struct bhnd_dma_translation *dt;
bus_dma_tag_t parent_dmat;
size_t nres, ndt;
int error;
parent_dmat = bus_get_dma_tag(parent_dev);
hr = malloc(sizeof(*hr), M_BHND, M_WAITOK);
hr->owner = parent_dev;
hr->cfg = hwcfg;
hr->resource_specs = NULL;
hr->resources = NULL;
hr->dma_tags = NULL;
hr->num_dma_tags = 0;
/* Determine our bridge resource count from the hardware config. */
nres = 0;
for (size_t i = 0; hwcfg->resource_specs[i].type != -1; i++)
nres++;
/* Determine the total count and validate our DMA translation table. */
ndt = 0;
for (dt = hwcfg->dma_translations; dt != NULL &&
!BHND_DMA_IS_TRANSLATION_TABLE_END(dt); dt++)
{
/* Validate the defined translation */
if ((dt->base_addr & dt->addr_mask) != 0) {
device_printf(dev, "invalid DMA translation; base "
"address %#jx overlaps address mask %#jx",
(uintmax_t)dt->base_addr, (uintmax_t)dt->addr_mask);
error = EINVAL;
goto failed;
}
if ((dt->addrext_mask & dt->addr_mask) != 0) {
device_printf(dev, "invalid DMA translation; addrext "
"mask %#jx overlaps address mask %#jx",
(uintmax_t)dt->addrext_mask,
(uintmax_t)dt->addr_mask);
error = EINVAL;
goto failed;
}
/* Increment our entry count */
ndt++;
}
/* Allocate our DMA tags */
hr->dma_tags = malloc(sizeof(*hr->dma_tags) * ndt, M_BHND,
M_WAITOK|M_ZERO);
for (size_t i = 0; i < ndt; i++) {
error = bhndb_dma_tag_create(dev, parent_dmat,
&hwcfg->dma_translations[i], &hr->dma_tags[i]);
if (error)
goto failed;
hr->num_dma_tags++;
}
/* Allocate space for a non-const copy of our resource_spec
* table; this will be updated with the RIDs assigned by
* bus_alloc_resources. */
hr->resource_specs = malloc(sizeof(hr->resource_specs[0]) * (nres + 1),
M_BHND, M_WAITOK);
/* Initialize and terminate the table */
for (size_t i = 0; i < nres; i++)
hr->resource_specs[i] = hwcfg->resource_specs[i];
hr->resource_specs[nres].type = -1;
/* Allocate space for our resource references */
hr->resources = malloc(sizeof(hr->resources[0]) * nres, M_BHND,
M_WAITOK);
/* Allocate host resources */
error = bus_alloc_resources(hr->owner, hr->resource_specs,
hr->resources);
if (error) {
device_printf(dev, "could not allocate bridge resources via "
"%s: %d\n", device_get_nameunit(parent_dev), error);
goto failed;
}
*resources = hr;
return (0);
failed:
if (hr->resource_specs != NULL)
free(hr->resource_specs, M_BHND);
if (hr->resources != NULL)
free(hr->resources, M_BHND);
for (size_t i = 0; i < hr->num_dma_tags; i++)
bus_dma_tag_destroy(hr->dma_tags[i]);
if (hr->dma_tags != NULL)
free(hr->dma_tags, M_BHND);
free(hr, M_BHND);
return (error);
}
/**
* Deallocate a set of bridge host resources.
*
* @param hr The resources to be freed.
*/
void
bhndb_release_host_resources(struct bhndb_host_resources *hr)
{
bus_release_resources(hr->owner, hr->resource_specs, hr->resources);
for (size_t i = 0; i < hr->num_dma_tags; i++)
bus_dma_tag_destroy(hr->dma_tags[i]);
free(hr->resources, M_BHND);
free(hr->resource_specs, M_BHND);
free(hr->dma_tags, M_BHND);
free(hr, M_BHND);
}
/**
* Search @p cores for the core serving as the bhnd host bridge.
*
* This function uses a heuristic valid on all known PCI/PCIe/PCMCIA-bridged
* bhnd(4) devices to determine the hostb core:
*
* - The core must have a Broadcom vendor ID.
* - The core devclass must match the bridge type.
* - The core must be the first device on the bus with the bridged device
* class.
*
* @param cores The core table to search.
* @param ncores The number of cores in @p cores.
* @param bridge_devclass The expected device class of the bridge core.
* @param[out] core If found, the matching host bridge core info.
*
* @retval 0 success
* @retval ENOENT not found
*/
int
bhndb_find_hostb_core(struct bhnd_core_info *cores, u_int ncores,
bhnd_devclass_t bridge_devclass, struct bhnd_core_info *core)
{
struct bhnd_core_match md;
struct bhnd_core_info *match;
u_int match_core_idx;
/* Set up a match descriptor for the required device class. */
md = (struct bhnd_core_match) {
BHND_MATCH_CORE_CLASS(bridge_devclass),
BHND_MATCH_CORE_UNIT(0)
};
/* Find the matching core with the lowest core index */
match = NULL;
match_core_idx = UINT_MAX;
for (u_int i = 0; i < ncores; i++) {
if (!bhnd_core_matches(&cores[i], &md))
continue;
/* Lower core indices take precedence */
if (match != NULL && match_core_idx < match->core_idx)
continue;
match = &cores[i];
match_core_idx = match->core_idx;
}
if (match == NULL)
return (ENOENT);
*core = *match;
return (0);
}
/**
* Allocate a host interrupt source and its backing SYS_RES_IRQ host resource.
*
* @param owner The device to be used to allocate a SYS_RES_IRQ
* resource with @p rid.
* @param rid The resource ID of the IRQ to be allocated.
* @param start The start value to be passed to bus_alloc_resource().
* @param end The end value to be passed to bus_alloc_resource().
* @param count The count to be passed to bus_alloc_resource().
* @param flags The flags to be passed to bus_alloc_resource().
*
* @retval non-NULL success
* @retval NULL if allocation fails.
*/
struct bhndb_intr_isrc *
bhndb_alloc_intr_isrc(device_t owner, int rid, rman_res_t start, rman_res_t end,
rman_res_t count, u_int flags)
{
struct bhndb_intr_isrc *isrc;
isrc = malloc(sizeof(*isrc), M_BHND, M_NOWAIT);
if (isrc == NULL)
return (NULL);
isrc->is_owner = owner;
isrc->is_rid = rid;
isrc->is_res = bus_alloc_resource(owner, SYS_RES_IRQ, &isrc->is_rid,
start, end, count, flags);
if (isrc->is_res == NULL) {
free(isrc, M_BHND);
return (NULL);
}
return (isrc);
}
/**
* Free a host interrupt source and its backing host resource.
*
* @param isrc The interrupt source to be freed.
*/
void
bhndb_free_intr_isrc(struct bhndb_intr_isrc *isrc)
{
bus_release_resource(isrc->is_owner, SYS_RES_IRQ, isrc->is_rid,
isrc->is_res);
free(isrc, M_BHND);
}
/**
* Allocate and initialize a new interrupt handler entry.
*
* @param owner The child device that owns this entry.
* @param r The child's interrupt resource.
* @param isrc The isrc mapped for this entry.
*
* @retval non-NULL success
* @retval NULL if allocation fails.
*/
struct bhndb_intr_handler *
bhndb_alloc_intr_handler(device_t owner, struct resource *r,
struct bhndb_intr_isrc *isrc)
{
struct bhndb_intr_handler *ih;
ih = malloc(sizeof(*ih), M_BHND, M_NOWAIT | M_ZERO);
ih->ih_owner = owner;
ih->ih_res = r;
ih->ih_isrc = isrc;
ih->ih_cookiep = NULL;
ih->ih_active = false;
return (ih);
}
/**
* Free an interrupt handler entry.
*
* @param br The resource state owning @p ih.
* @param ih The interrupt handler entry to be removed.
*/
void
bhndb_free_intr_handler(struct bhndb_intr_handler *ih)
{
KASSERT(!ih->ih_active, ("free of active interrupt handler %p",
ih->ih_cookiep));
free(ih, M_BHND);
}
/**
* Add an active interrupt handler to the given resource state.
*
* @param br The resource state to be modified.
* @param ih The interrupt handler entry to be added.
*/
void
bhndb_register_intr_handler(struct bhndb_resources *br,
struct bhndb_intr_handler *ih)
{
KASSERT(!ih->ih_active, ("duplicate registration of interrupt "
"handler %p", ih->ih_cookiep));
KASSERT(ih->ih_cookiep != NULL, ("missing cookiep"));
ih->ih_active = true;
STAILQ_INSERT_HEAD(&br->bus_intrs, ih, ih_link);
}
/**
* Remove an interrupt handler from the given resource state.
*
* @param br The resource state containing @p ih.
* @param ih The interrupt handler entry to be removed.
*/
void
bhndb_deregister_intr_handler(struct bhndb_resources *br,
struct bhndb_intr_handler *ih)
{
KASSERT(ih->ih_active, ("duplicate deregistration of interrupt "
"handler %p", ih->ih_cookiep));
KASSERT(bhndb_find_intr_handler(br, ih) == ih,
("unknown interrupt handler %p", ih));
STAILQ_REMOVE(&br->bus_intrs, ih, bhndb_intr_handler, ih_link);
ih->ih_active = false;
}
/**
* Return the interrupt handler entry corresponding to @p cookiep, or NULL
* if no entry is found.
*
* @param br The resource state to search for the given @p cookiep.
* @param cookiep The interrupt handler's bus-assigned cookiep value.
*/
struct bhndb_intr_handler *
bhndb_find_intr_handler(struct bhndb_resources *br, void *cookiep)
{
struct bhndb_intr_handler *ih;
STAILQ_FOREACH(ih, &br->bus_intrs, ih_link) {
if (ih == cookiep)
return (ih);
}
/* Not found */
return (NULL);
}
/**
* Find the maximum start and end limits of the bridged resource @p r.
*
* If the resource is not currently mapped by the bridge, ENOENT will be
* returned.
*
* @param br The resource state to search.
* @param type The resource type (see SYS_RES_*).
* @param r The resource to search for in @p br.
* @param[out] start On success, the minimum supported start address.
* @param[out] end On success, the maximum supported end address.
*
* @retval 0 success
* @retval ENOENT no active mapping found for @p r of @p type
*/
int
bhndb_find_resource_limits(struct bhndb_resources *br, int type,
struct resource *r, rman_res_t *start, rman_res_t *end)
{
struct bhndb_dw_alloc *dynamic;
struct bhndb_region *sregion;
struct bhndb_intr_handler *ih;
switch (type) {
case SYS_RES_IRQ:
/* Is this one of ours? */
STAILQ_FOREACH(ih, &br->bus_intrs, ih_link) {
if (ih->ih_res == r)
continue;
/* We don't support adjusting IRQ resource limits */
*start = rman_get_start(r);
*end = rman_get_end(r);
return (0);
}
/* Not found */
return (ENOENT);
case SYS_RES_MEMORY: {
/* Check for an enclosing dynamic register window */
if ((dynamic = bhndb_dw_find_resource(br, r))) {
*start = dynamic->target;
*end = dynamic->target + dynamic->win->win_size - 1;
return (0);
}
/* Check for a static region */
sregion = bhndb_find_resource_region(br, rman_get_start(r),
rman_get_size(r));
if (sregion != NULL && sregion->static_regwin != NULL) {
*start = sregion->addr;
*end = sregion->addr + sregion->size - 1;
return (0);
}
/* Not found */
return (ENOENT);
}
default:
device_printf(br->dev, "unknown resource type: %d\n", type);
return (ENOENT);
}
}
/**
* Add a bus region entry to @p r for the given base @p addr and @p size.
*
* @param br The resource state to which the bus region entry will be added.
* @param addr The base address of this region.
* @param size The size of this region.
* @param priority The resource priority to be assigned to allocations
* made within this bus region.
* @param alloc_flags resource allocation flags (@see bhndb_alloc_flags)
* @param static_regwin If available, a static register window mapping this
* bus region entry. If not available, NULL.
*
* @retval 0 success
* @retval non-zero if adding the bus region fails.
*/
int
bhndb_add_resource_region(struct bhndb_resources *br, bhnd_addr_t addr,
bhnd_size_t size, bhndb_priority_t priority, uint32_t alloc_flags,
const struct bhndb_regwin *static_regwin)
{
struct bhndb_region *reg;
/* Insert in the bus resource list */
reg = malloc(sizeof(*reg), M_BHND, M_NOWAIT);
if (reg == NULL)
return (ENOMEM);
*reg = (struct bhndb_region) {
.addr = addr,
.size = size,
.priority = priority,
.alloc_flags = alloc_flags,
.static_regwin = static_regwin
};
STAILQ_INSERT_HEAD(&br->bus_regions, reg, link);
return (0);
}
/**
* Return true if a mapping of @p size bytes at @p addr is provided by either
* one contiguous bus region, or by multiple discontiguous regions.
*
* @param br The resource state to query.
* @param addr The requested starting address.
* @param size The requested size.
*/
bool
bhndb_has_static_region_mapping(struct bhndb_resources *br,
bhnd_addr_t addr, bhnd_size_t size)
{
struct bhndb_region *region;
bhnd_addr_t r_addr;
r_addr = addr;
while ((region = bhndb_find_resource_region(br, r_addr, 1)) != NULL) {
/* Must be backed by a static register window */
if (region->static_regwin == NULL)
return (false);
/* Adjust the search offset */
r_addr += region->size;
/* Have we traversed a complete (if discontiguous) mapping? */
if (r_addr == addr + size)
return (true);
}
/* No complete mapping found */
return (false);
}
/**
* Find the bus region that maps @p size bytes at @p addr.
*
* @param br The resource state to search.
* @param addr The requested starting address.
* @param size The requested size.
*
* @retval bhndb_region A region that fully contains the requested range.
* @retval NULL If no mapping region can be found.
*/
struct bhndb_region *
bhndb_find_resource_region(struct bhndb_resources *br, bhnd_addr_t addr,
bhnd_size_t size)
{
struct bhndb_region *region;
STAILQ_FOREACH(region, &br->bus_regions, link) {
/* Request must fit within the region's mapping */
if (addr < region->addr)
continue;
if (addr + size > region->addr + region->size)
continue;
return (region);
}
/* Not found */
return (NULL);
}
/**
* Find the entry matching @p r in @p dwa's references, if any.
*
* @param dwa The dynamic window allocation to search
* @param r The resource to search for in @p dwa.
*/
static struct bhndb_dw_rentry *
bhndb_dw_find_resource_entry(struct bhndb_dw_alloc *dwa, struct resource *r)
{
struct bhndb_dw_rentry *rentry;
LIST_FOREACH(rentry, &dwa->refs, dw_link) {
struct resource *dw_res = rentry->dw_res;
/* Match dev/rid/addr/size */
if (rman_get_device(dw_res) != rman_get_device(r) ||
rman_get_rid(dw_res) != rman_get_rid(r) ||
rman_get_start(dw_res) != rman_get_start(r) ||
rman_get_size(dw_res) != rman_get_size(r))
{
continue;
}
/* Matching allocation found */
return (rentry);
}
return (NULL);
}
/**
* Find the dynamic region allocated for @p r, if any.
*
* @param br The resource state to search.
* @param r The resource to search for.
*
* @retval bhndb_dw_alloc The allocation record for @p r.
* @retval NULL if no dynamic window is allocated for @p r.
*/
struct bhndb_dw_alloc *
bhndb_dw_find_resource(struct bhndb_resources *br, struct resource *r)
{
struct bhndb_dw_alloc *dwa;
for (size_t i = 0; i < br->dwa_count; i++) {
dwa = &br->dw_alloc[i];
/* Skip free dynamic windows */
if (bhndb_dw_is_free(br, dwa))
continue;
/* Matching allocation found? */
if (bhndb_dw_find_resource_entry(dwa, r) != NULL)
return (dwa);
}
return (NULL);
}
/**
* Find an existing dynamic window mapping @p size bytes
* at @p addr. The window may or may not be free.
*
* @param br The resource state to search.
* @param addr The requested starting address.
* @param size The requested size.
*
* @retval bhndb_dw_alloc A window allocation that fully contains the requested
* range.
* @retval NULL If no mapping region can be found.
*/
struct bhndb_dw_alloc *
bhndb_dw_find_mapping(struct bhndb_resources *br, bhnd_addr_t addr,
bhnd_size_t size)
{
struct bhndb_dw_alloc *dwr;
const struct bhndb_regwin *win;
/* Search for an existing dynamic mapping of this address range. */
for (size_t i = 0; i < br->dwa_count; i++) {
dwr = &br->dw_alloc[i];
win = dwr->win;
/* Verify the range */
if (addr < dwr->target)
continue;
if (addr + size > dwr->target + win->win_size)
continue;
/* Found a usable mapping */
return (dwr);
}
/* not found */
return (NULL);
}
/**
* Retain a reference to @p dwa for use by @p res.
*
* @param br The resource state owning @p dwa.
* @param dwa The allocation record to be retained.
* @param res The resource that will own a reference to @p dwa.
*
* @retval 0 success
* @retval ENOMEM Failed to allocate a new reference structure.
*/
int
bhndb_dw_retain(struct bhndb_resources *br, struct bhndb_dw_alloc *dwa,
struct resource *res)
{
struct bhndb_dw_rentry *rentry;
KASSERT(bhndb_dw_find_resource_entry(dwa, res) == NULL,
("double-retain of dynamic window for same resource"));
/* Insert a reference entry; we use M_NOWAIT to allow use from
* within a non-sleepable lock */
rentry = malloc(sizeof(*rentry), M_BHND, M_NOWAIT);
if (rentry == NULL)
return (ENOMEM);
rentry->dw_res = res;
LIST_INSERT_HEAD(&dwa->refs, rentry, dw_link);
/* Update the free list */
bit_set(br->dwa_freelist, dwa->rnid);
return (0);
}
/**
* Release a reference to @p dwa previously retained by @p res. If the
* reference count of @p dwa reaches zero, it will be added to the
* free list.
*
* @param br The resource state owning @p dwa.
* @param dwa The allocation record to be released.
* @param res The resource that currently owns a reference to @p dwa.
*/
void
bhndb_dw_release(struct bhndb_resources *br, struct bhndb_dw_alloc *dwa,
struct resource *r)
{
struct bhndb_dw_rentry *rentry;
/* Find the rentry */
rentry = bhndb_dw_find_resource_entry(dwa, r);
KASSERT(rentry != NULL, ("over release of resource entry"));
LIST_REMOVE(rentry, dw_link);
free(rentry, M_BHND);
/* If this was the last reference, update the free list */
if (LIST_EMPTY(&dwa->refs))
bit_clear(br->dwa_freelist, dwa->rnid);
}
/**
* Attempt to set (or reset) the target address of @p dwa to map @p size bytes
* at @p addr.
*
* This will apply any necessary window alignment and verify that
* the window is capable of mapping the requested range prior to modifying
* therecord.
*
* @param dev The device on which to issue the BHNDB_SET_WINDOW_ADDR() request.
* @param br The resource state owning @p dwa.
* @param dwa The allocation record to be configured.
* @param addr The address to be mapped via @p dwa.
* @param size The number of bytes to be mapped at @p addr.
*
* @retval 0 success
* @retval non-zero no usable register window available.
*/
int
bhndb_dw_set_addr(device_t dev, struct bhndb_resources *br,
struct bhndb_dw_alloc *dwa, bus_addr_t addr, bus_size_t size)
{
const struct bhndb_regwin *rw;
bus_addr_t offset;
int error;
rw = dwa->win;
KASSERT(bhndb_dw_is_free(br, dwa) || mtx_owned(&br->dw_steal_mtx),
("attempting to set the target address on an in-use window"));
/* Page-align the target address */
offset = addr % rw->win_size;
dwa->target = addr - offset;
/* Verify that the window is large enough for the full target */
if (rw->win_size - offset < size)
return (ENOMEM);
/* Update the window target */
error = BHNDB_SET_WINDOW_ADDR(dev, dwa->win, dwa->target);
if (error) {
dwa->target = 0x0;
return (error);
}
return (0);
}
/**
* Steal an in-use allocation record from @p br, returning the record's current
* target in @p saved on success.
*
* This function acquires a mutex and disables interrupts; callers should
* avoid holding a stolen window longer than required to issue an I/O
* request.
*
* A successful call to bhndb_dw_steal() must be balanced with a call to
* bhndb_dw_return_stolen().
*
* @param br The resource state from which a window should be stolen.
* @param saved The stolen window's saved target address.
*
* @retval non-NULL success
* @retval NULL no dynamic window regions are defined.
*/
struct bhndb_dw_alloc *
bhndb_dw_steal(struct bhndb_resources *br, bus_addr_t *saved)
{
struct bhndb_dw_alloc *dw_stolen;
KASSERT(bhndb_dw_next_free(br) == NULL,
("attempting to steal an in-use window while free windows remain"));
/* Nothing to steal from? */
if (br->dwa_count == 0)
return (NULL);
/*
* Acquire our steal spinlock; this will be released in
* bhndb_dw_return_stolen().
*
* Acquiring also disables interrupts, which is required when one is
* stealing an in-use existing register window.
*/
mtx_lock_spin(&br->dw_steal_mtx);
dw_stolen = &br->dw_alloc[0];
*saved = dw_stolen->target;
return (dw_stolen);
}
/**
* Return an allocation record previously stolen using bhndb_dw_steal().
*
* @param dev The device on which to issue a BHNDB_SET_WINDOW_ADDR() request.
* @param br The resource state owning @p dwa.
* @param dwa The allocation record to be returned.
* @param saved The original target address provided by bhndb_dw_steal().
*/
void
bhndb_dw_return_stolen(device_t dev, struct bhndb_resources *br,
struct bhndb_dw_alloc *dwa, bus_addr_t saved)
{
int error;
mtx_assert(&br->dw_steal_mtx, MA_OWNED);
error = bhndb_dw_set_addr(dev, br, dwa, saved, 0);
if (error) {
panic("failed to restore register window target %#jx: %d\n",
(uintmax_t)saved, error);
}
mtx_unlock_spin(&br->dw_steal_mtx);
}
/**
* Return the count of @p type register windows in @p table.
*
* @param table The table to search.
* @param type The required window type, or BHNDB_REGWIN_T_INVALID to
* count all register window types.
*/
size_t
bhndb_regwin_count(const struct bhndb_regwin *table,
bhndb_regwin_type_t type)
{
const struct bhndb_regwin *rw;
size_t count;
count = 0;
for (rw = table; rw->win_type != BHNDB_REGWIN_T_INVALID; rw++) {
if (type == BHNDB_REGWIN_T_INVALID || rw->win_type == type)
count++;
}
return (count);
}
/**
* Search @p table for the first window with the given @p type.
*
* @param table The table to search.
* @param type The required window type.
* @param min_size The minimum window size.
*
* @retval bhndb_regwin The first matching window.
* @retval NULL If no window of the requested type could be found.
*/
const struct bhndb_regwin *
bhndb_regwin_find_type(const struct bhndb_regwin *table,
bhndb_regwin_type_t type, bus_size_t min_size)
{
const struct bhndb_regwin *rw;
for (rw = table; rw->win_type != BHNDB_REGWIN_T_INVALID; rw++)
{
if (rw->win_type == type && rw->win_size >= min_size)
return (rw);
}
return (NULL);
}
/**
* Search @p windows for the first matching core window.
*
* @param table The table to search.
* @param class The required core class.
* @param unit The required core unit, or -1.
* @param port_type The required port type.
* @param port The required port.
* @param region The required region.
* @param offset The required readable core register block offset.
* @param min_size The required minimum readable size at @p offset.
*
* @retval bhndb_regwin The first matching window.
* @retval NULL If no matching window was found.
*/
const struct bhndb_regwin *
bhndb_regwin_find_core(const struct bhndb_regwin *table, bhnd_devclass_t class,
int unit, bhnd_port_type port_type, u_int port, u_int region,
bus_size_t offset, bus_size_t min_size)
{
const struct bhndb_regwin *rw;
for (rw = table; rw->win_type != BHNDB_REGWIN_T_INVALID; rw++)
{
bus_size_t rw_offset;
/* Match on core, port, and region attributes */
if (rw->win_type != BHNDB_REGWIN_T_CORE)
continue;
if (rw->d.core.class != class)
continue;
if (unit != -1 && rw->d.core.unit != unit)
continue;
if (rw->d.core.port_type != port_type)
continue;
if (rw->d.core.port != port)
continue;
if (rw->d.core.region != region)
continue;
/* Verify that the requested range is mapped within
* this register window */
if (rw->d.core.offset > offset)
continue;
rw_offset = offset - rw->d.core.offset;
if (rw->win_size < rw_offset)
continue;
if (rw->win_size - rw_offset < min_size)
continue;
return (rw);
}
return (NULL);
}
/**
* Search @p windows for the best available window of at least @p min_size.
*
* Search order:
* - BHND_REGWIN_T_CORE
* - BHND_REGWIN_T_DYN
*
* @param table The table to search.
* @param class The required core class.
* @param unit The required core unit, or -1.
* @param port_type The required port type.
* @param port The required port.
* @param region The required region.
* @param offset The required readable core register block offset.
* @param min_size The required minimum readable size at @p offset.
*
* @retval bhndb_regwin The first matching window.
* @retval NULL If no matching window was found.
*/
const struct bhndb_regwin *
bhndb_regwin_find_best(const struct bhndb_regwin *table,
bhnd_devclass_t class, int unit, bhnd_port_type port_type, u_int port,
u_int region, bus_size_t offset, bus_size_t min_size)
{
const struct bhndb_regwin *rw;
/* Prefer a fixed core mapping */
rw = bhndb_regwin_find_core(table, class, unit, port_type,
port, region, offset, min_size);
if (rw != NULL)
return (rw);
/* Fall back on a generic dynamic window */
return (bhndb_regwin_find_type(table, BHNDB_REGWIN_T_DYN, min_size));
}
/**
* Return true if @p regw defines a BHNDB_REGWIN_T_CORE register window
* that matches against @p core.
*
* @param regw A register window to match against.
* @param core The bhnd(4) core info to match against @p regw.
*/
bool
bhndb_regwin_match_core(const struct bhndb_regwin *regw,
struct bhnd_core_info *core)
{
/* Only core windows are supported */
if (regw->win_type != BHNDB_REGWIN_T_CORE)
return (false);
/* Device class must match */
if (bhnd_core_class(core) != regw->d.core.class)
return (false);
/* Device unit must match */
if (core->unit != regw->d.core.unit)
return (false);
/* Matches */
return (true);
}
/**
* Search for a core resource priority descriptor in @p table that matches
* @p core.
*
* @param table The table to search.
* @param core The core to match against @p table.
*/
const struct bhndb_hw_priority *
bhndb_hw_priority_find_core(const struct bhndb_hw_priority *table,
struct bhnd_core_info *core)
{
const struct bhndb_hw_priority *hp;
for (hp = table; hp->ports != NULL; hp++) {
if (bhnd_core_matches(core, &hp->match))
return (hp);
}
/* not found */
return (NULL);
}
/**
* Search for a port resource priority descriptor in @p table.
*
* @param table The table to search.
* @param core The core to match against @p table.
* @param port_type The required port type.
* @param port The required port.
* @param region The required region.
*/
const struct bhndb_port_priority *
bhndb_hw_priorty_find_port(const struct bhndb_hw_priority *table,
struct bhnd_core_info *core, bhnd_port_type port_type, u_int port,
u_int region)
{
const struct bhndb_hw_priority *hp;
if ((hp = bhndb_hw_priority_find_core(table, core)) == NULL)
return (NULL);
for (u_int i = 0; i < hp->num_ports; i++) {
const struct bhndb_port_priority *pp = &hp->ports[i];
if (pp->type != port_type)
continue;
if (pp->port != port)
continue;
if (pp->region != region)
continue;
return (pp);
}
/* not found */
return (NULL);
}