693c9516fa
Sanitizer instrumentation of course cannot automatically update shadow state when devices write to host memory. KMSAN thus hooks into busdma, both to update shadow state after a device write, and to verify that the kernel does not publish uninitalized bytes to devices. To implement this, when KMSAN is configured, each dmamap embeds a memory descriptor describing the region currently loaded into the map. bus_dmamap_sync() uses the operation flags to determine whether to validate the loaded region or to mark it as initialized in the shadow map. Note that in cases where the amount of data written is less than the buffer size, the entire buffer is marked initialized even when it is not. For example, if a NIC writes a 128B packet into a 2KB buffer, the entire buffer will be marked initialized, but subsequent accesses past the first 128 bytes are likely caused by bugs. Reviewed by: kib Sponsored by: The FreeBSD Foundation Differential Revision: https://reviews.freebsd.org/D31338
207 lines
5.7 KiB
C
207 lines
5.7 KiB
C
/*-
|
|
* Copyright (c) 2017 Jason A. Harmening.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
#ifndef _X86_BUS_DMA_H_
|
|
#define _X86_BUS_DMA_H_
|
|
|
|
#define WANT_INLINE_DMAMAP
|
|
#include <sys/bus_dma.h>
|
|
#include <sys/_null.h>
|
|
|
|
#include <x86/busdma_impl.h>
|
|
|
|
/*
|
|
* Is DMA address 1:1 mapping of physical address
|
|
*/
|
|
static inline bool
|
|
bus_dma_id_mapped(bus_dma_tag_t dmat, vm_paddr_t buf, bus_size_t buflen)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
return (tc->impl->id_mapped(dmat, buf, buflen));
|
|
}
|
|
|
|
/*
|
|
* Allocate a handle for mapping from kva/uva/physical
|
|
* address space into bus device space.
|
|
*/
|
|
static inline int
|
|
bus_dmamap_create(bus_dma_tag_t dmat, int flags, bus_dmamap_t *mapp)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
return (tc->impl->map_create(dmat, flags, mapp));
|
|
}
|
|
|
|
/*
|
|
* Destroy a handle for mapping from kva/uva/physical
|
|
* address space into bus device space.
|
|
*/
|
|
static inline int
|
|
bus_dmamap_destroy(bus_dma_tag_t dmat, bus_dmamap_t map)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
return (tc->impl->map_destroy(dmat, map));
|
|
}
|
|
|
|
/*
|
|
* Allocate a piece of memory that can be efficiently mapped into
|
|
* bus device space based on the constraints lited in the dma tag.
|
|
* A dmamap to for use with dmamap_load is also allocated.
|
|
*/
|
|
static inline int
|
|
bus_dmamem_alloc(bus_dma_tag_t dmat, void** vaddr, int flags,
|
|
bus_dmamap_t *mapp)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
return (tc->impl->mem_alloc(dmat, vaddr, flags, mapp));
|
|
}
|
|
|
|
/*
|
|
* Free a piece of memory and it's allociated dmamap, that was allocated
|
|
* via bus_dmamem_alloc. Make the same choice for free/contigfree.
|
|
*/
|
|
static inline void
|
|
bus_dmamem_free(bus_dma_tag_t dmat, void *vaddr, bus_dmamap_t map)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
tc->impl->mem_free(dmat, vaddr, map);
|
|
}
|
|
|
|
/*
|
|
* Release the mapping held by map.
|
|
*/
|
|
static inline void
|
|
bus_dmamap_unload(bus_dma_tag_t dmat, bus_dmamap_t map)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
if (map != NULL) {
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
tc->impl->map_unload(dmat, map);
|
|
}
|
|
}
|
|
|
|
static inline void
|
|
bus_dmamap_sync(bus_dma_tag_t dmat, bus_dmamap_t map, bus_dmasync_op_t op)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
if (map != NULL) {
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
tc->impl->map_sync(dmat, map, op);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Utility function to load a physical buffer. segp contains
|
|
* the starting segment on entrace, and the ending segment on exit.
|
|
*/
|
|
static inline int
|
|
_bus_dmamap_load_phys(bus_dma_tag_t dmat, bus_dmamap_t map, vm_paddr_t buf,
|
|
bus_size_t buflen, int flags, bus_dma_segment_t *segs, int *segp)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
return (tc->impl->load_phys(dmat, map, buf, buflen, flags, segs,
|
|
segp));
|
|
}
|
|
|
|
static inline int
|
|
_bus_dmamap_load_ma(bus_dma_tag_t dmat, bus_dmamap_t map, struct vm_page **ma,
|
|
bus_size_t tlen, int ma_offs, int flags, bus_dma_segment_t *segs,
|
|
int *segp)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
return (tc->impl->load_ma(dmat, map, ma, tlen, ma_offs, flags,
|
|
segs, segp));
|
|
}
|
|
|
|
/*
|
|
* Utility function to load a linear buffer. segp contains
|
|
* the starting segment on entrace, and the ending segment on exit.
|
|
*/
|
|
static inline int
|
|
_bus_dmamap_load_buffer(bus_dma_tag_t dmat, bus_dmamap_t map, void *buf,
|
|
bus_size_t buflen, struct pmap *pmap, int flags, bus_dma_segment_t *segs,
|
|
int *segp)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
return (tc->impl->load_buffer(dmat, map, buf, buflen, pmap, flags, segs,
|
|
segp));
|
|
}
|
|
|
|
static inline void
|
|
_bus_dmamap_waitok(bus_dma_tag_t dmat, bus_dmamap_t map,
|
|
struct memdesc *mem, bus_dmamap_callback_t *callback, void *callback_arg)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
if (map != NULL) {
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
tc->impl->map_waitok(dmat, map, mem, callback, callback_arg);
|
|
}
|
|
}
|
|
|
|
static inline bus_dma_segment_t *
|
|
_bus_dmamap_complete(bus_dma_tag_t dmat, bus_dmamap_t map,
|
|
bus_dma_segment_t *segs, int nsegs, int error)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
return (tc->impl->map_complete(dmat, map, segs, nsegs, error));
|
|
}
|
|
|
|
#ifdef KMSAN
|
|
static inline void
|
|
_bus_dmamap_load_kmsan(bus_dma_tag_t dmat, bus_dmamap_t map,
|
|
struct memdesc *mem)
|
|
{
|
|
struct bus_dma_tag_common *tc;
|
|
|
|
tc = (struct bus_dma_tag_common *)dmat;
|
|
return (tc->impl->load_kmsan(map, mem));
|
|
}
|
|
#endif
|
|
|
|
#endif /* !_X86_BUS_DMA_H_ */
|