/* * Copyright (c) 1999 Matthew R. Green * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. The name of the author may not be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * * from: NetBSD: iommuvar.h,v 1.9 2001/07/20 00:07:13 eeh Exp * * $FreeBSD$ */ #ifndef _MACHINE_IOMMUVAR_H_ #define _MACHINE_IOMMUVAR_H_ #define IO_PAGE_SIZE PAGE_SIZE_8K #define IO_PAGE_MASK PAGE_MASK_8K #define IO_PAGE_SHIFT PAGE_SHIFT_8K #define round_io_page(x) round_page(x) #define trunc_io_page(x) trunc_page(x) /* * per-IOMMU state */ struct iommu_state { vm_offset_t is_ptsb; /* TSB physical address */ u_int64_t *is_tsb; /* TSB virtual address */ int is_tsbsize; /* 0 = 8K, ... */ u_int64_t is_dvmabase; int64_t is_cr; /* IOMMU control register value */ struct rman is_dvma_rman; /* DVMA map for this instance */ vm_offset_t is_flushpa[2]; volatile int64_t *is_flushva[2]; /* * When a flush is completed, 64 bytes will be stored at the given * location, the first double word being 1, to indicate completion. * The lower 6 address bits are ignored, so the addresses need to be * suitably aligned; over-allocate a large enough margin to be able * to adjust it. * Two such buffers are needed. * Needs to be volatile or egcs optimizes away loads. */ volatile char is_flush[STRBUF_FLUSHSYNC_NBYTES * 3 - 1]; /* copies of our parents state, to allow us to be self contained */ bus_space_tag_t is_bustag; /* our bus tag */ bus_space_handle_t is_bushandle; bus_addr_t is_iommu; /* IOMMU registers */ bus_addr_t is_sb[2]; /* streaming buffer */ bus_addr_t is_dtag; /* tag diagnostics access */ bus_addr_t is_ddram; /* data ram diag. access */ bus_addr_t is_dqueue; /* LRU queue diag. access */ bus_addr_t is_dva; /* VA diag. register */ bus_addr_t is_dtcmp; /* tag compare diag. access */ }; /* interfaces for PCI/SBUS code */ void iommu_init __P((char *, struct iommu_state *, int, u_int32_t)); void iommu_reset __P((struct iommu_state *)); void iommu_enter __P((struct iommu_state *, vm_offset_t, vm_offset_t, int)); void iommu_remove __P((struct iommu_state *, vm_offset_t, size_t)); int iommu_dvmamap_create __P((bus_dma_tag_t, struct iommu_state *, int, bus_dmamap_t *)); int iommu_dvmamap_destroy __P((bus_dma_tag_t, struct iommu_state *, bus_dmamap_t)); int iommu_dvmamap_load __P((bus_dma_tag_t, struct iommu_state *, bus_dmamap_t, void *, bus_size_t, bus_dmamap_callback_t *, void *, int)); void iommu_dvmamap_unload __P((bus_dma_tag_t, struct iommu_state *, bus_dmamap_t)); void iommu_dvmamap_sync __P((bus_dma_tag_t, struct iommu_state *, bus_dmamap_t, bus_dmasync_op_t)); int iommu_dvmamem_alloc __P((bus_dma_tag_t, struct iommu_state *, void **, int, bus_dmamap_t *)); void iommu_dvmamem_free __P((bus_dma_tag_t, struct iommu_state *, void *, bus_dmamap_t)); #endif /* !_MACHINE_IOMMUVAR_H_ */