freebsd-nq/sys/dev/ioat/ioat.h
Conrad Meyer 236b57fe36 ioat(4): Set __result_use_check on ioat_acquire_reserve
Even M_WAITOK callers must check for failure.  For example, if the device is
quiescing, either due to automatic error-recovery induced reset, or due to
administrative detach, the routine will return ENXIO and the acquire
reference will not be held.  So, there is no mode in which it is safe to
assume the routine succeeds without checking.

Sponsored by:	Dell EMC Isilon
2019-01-17 23:21:02 +00:00

253 lines
9.2 KiB
C

/*-
* Copyright (C) 2012 Intel Corporation
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*/
__FBSDID("$FreeBSD$");
#ifndef __IOAT_H__
#define __IOAT_H__
#include <sys/param.h>
#include <machine/bus.h>
/*
* This file defines the public interface to the IOAT driver.
*/
/*
* Enables an interrupt for this operation. Typically, you would only enable
* this on the last operation in a group
*/
#define DMA_INT_EN 0x1
/*
* Like M_NOWAIT. Operations will return NULL if they cannot allocate a
* descriptor without blocking.
*/
#define DMA_NO_WAIT 0x2
/*
* Disallow prefetching the source of the following operation. Ordinarily, DMA
* operations can be pipelined on some hardware. E.g., operation 2's source
* may be prefetched before operation 1 completes.
*/
#define DMA_FENCE 0x4
#define _DMA_GENERIC_FLAGS (DMA_INT_EN | DMA_NO_WAIT | DMA_FENCE)
/*
* Emit a CRC32C as the result of a ioat_copy_crc() or ioat_crc().
*/
#define DMA_CRC_STORE 0x8
/*
* Compare the CRC32C of a ioat_copy_crc() or ioat_crc() against an expeceted
* value. It is invalid to specify both TEST and STORE.
*/
#define DMA_CRC_TEST 0x10
#define _DMA_CRC_TESTSTORE (DMA_CRC_STORE | DMA_CRC_TEST)
/*
* Use an inline comparison CRC32C or emit an inline CRC32C result. Invalid
* without one of STORE or TEST.
*/
#define DMA_CRC_INLINE 0x20
#define _DMA_CRC_FLAGS (DMA_CRC_STORE | DMA_CRC_TEST | DMA_CRC_INLINE)
/*
* Hardware revision number. Different hardware revisions support different
* features. For example, 3.2 cannot read from MMIO space, while 3.3 can.
*/
#define IOAT_VER_3_0 0x30
#define IOAT_VER_3_2 0x32
#define IOAT_VER_3_3 0x33
/*
* Hardware capabilities. Different hardware revisions support different
* features. It is often useful to detect specific features than try to infer
* them from hardware version.
*
* Different channels may support different features too; for example, 'PQ' may
* only be supported on the first two channels of some hardware.
*/
#define IOAT_DMACAP_PB (1 << 0)
#define IOAT_DMACAP_CRC (1 << 1)
#define IOAT_DMACAP_MARKER_SKIP (1 << 2)
#define IOAT_DMACAP_OLD_XOR (1 << 3)
#define IOAT_DMACAP_DCA (1 << 4)
#define IOAT_DMACAP_MOVECRC (1 << 5)
#define IOAT_DMACAP_BFILL (1 << 6)
#define IOAT_DMACAP_EXT_APIC (1 << 7)
#define IOAT_DMACAP_XOR (1 << 8)
#define IOAT_DMACAP_PQ (1 << 9)
#define IOAT_DMACAP_DMA_DIF (1 << 10)
#define IOAT_DMACAP_DWBES (1 << 13)
#define IOAT_DMACAP_RAID16SS (1 << 17)
#define IOAT_DMACAP_DMAMC (1 << 18)
#define IOAT_DMACAP_CTOS (1 << 19)
#define IOAT_DMACAP_STR \
"\20\24Completion_Timeout_Support\23DMA_with_Multicasting_Support" \
"\22RAID_Super_descriptors\16Descriptor_Write_Back_Error_Support" \
"\13DMA_with_DIF\12PQ\11XOR\10Extended_APIC_ID\07Block_Fill\06Move_CRC" \
"\05DCA\04Old_XOR\03Marker_Skipping\02CRC\01Page_Break"
typedef void *bus_dmaengine_t;
struct bus_dmadesc;
typedef void (*bus_dmaengine_callback_t)(void *arg, int error);
unsigned ioat_get_nchannels(void);
/*
* Called first to acquire a reference to the DMA channel
*
* Flags may be M_WAITOK or M_NOWAIT.
*/
bus_dmaengine_t ioat_get_dmaengine(uint32_t channel_index, int flags);
/* Release the DMA channel */
void ioat_put_dmaengine(bus_dmaengine_t dmaengine);
/* Check the DMA engine's HW version */
int ioat_get_hwversion(bus_dmaengine_t dmaengine);
size_t ioat_get_max_io_size(bus_dmaengine_t dmaengine);
uint32_t ioat_get_capabilities(bus_dmaengine_t dmaengine);
/*
* Set interrupt coalescing on a DMA channel.
*
* The argument is in microseconds. A zero value disables coalescing. Any
* other value delays interrupt generation for N microseconds to provide
* opportunity to coalesce multiple operations into a single interrupt.
*
* Returns an error status, or zero on success.
*
* - ERANGE if the given value exceeds the delay supported by the hardware.
* (All current hardware supports a maximum of 0x3fff microseconds delay.)
* - ENODEV if the hardware does not support interrupt coalescing.
*/
int ioat_set_interrupt_coalesce(bus_dmaengine_t dmaengine, uint16_t delay);
/*
* Return the maximum supported coalescing period, for use in
* ioat_set_interrupt_coalesce(). If the hardware does not support coalescing,
* returns zero.
*/
uint16_t ioat_get_max_coalesce_period(bus_dmaengine_t dmaengine);
/*
* Acquire must be called before issuing an operation to perform. Release is
* called after. Multiple operations can be issued within the context of one
* acquire and release
*/
void ioat_acquire(bus_dmaengine_t dmaengine);
void ioat_release(bus_dmaengine_t dmaengine);
/*
* Acquire_reserve can be called to ensure there is room for N descriptors. If
* it succeeds, the next N valid operations will successfully enqueue.
*
* It may fail with:
* - ENXIO if the channel is in an errored state, or the driver is being
* unloaded
* - EAGAIN if mflags included M_NOWAIT
*
* On failure, the caller does not hold the dmaengine.
*/
int ioat_acquire_reserve(bus_dmaengine_t dmaengine, unsigned n, int mflags)
__result_use_check;
/*
* Issue a blockfill operation. The 64-bit pattern 'fillpattern' is written to
* 'len' physically contiguous bytes at 'dst'.
*
* Only supported on devices with the BFILL capability.
*/
struct bus_dmadesc *ioat_blockfill(bus_dmaengine_t dmaengine, bus_addr_t dst,
uint64_t fillpattern, bus_size_t len, bus_dmaengine_callback_t callback_fn,
void *callback_arg, uint32_t flags);
/* Issues the copy data operation */
struct bus_dmadesc *ioat_copy(bus_dmaengine_t dmaengine, bus_addr_t dst,
bus_addr_t src, bus_size_t len, bus_dmaengine_callback_t callback_fn,
void *callback_arg, uint32_t flags);
/*
* Issue a copy data operation, with constraints:
* - src1, src2, dst1, dst2 are all page-aligned addresses
* - The quantity to copy is exactly 2 pages;
* - src1 -> dst1, src2 -> dst2
*
* Why use this instead of normal _copy()? You can copy two non-contiguous
* pages (src, dst, or both) with one descriptor.
*/
struct bus_dmadesc *ioat_copy_8k_aligned(bus_dmaengine_t dmaengine,
bus_addr_t dst1, bus_addr_t dst2, bus_addr_t src1, bus_addr_t src2,
bus_dmaengine_callback_t callback_fn, void *callback_arg, uint32_t flags);
/*
* Copy len bytes from dst to src, like ioat_copy().
*
* Additionally, accumulate a CRC32C of the data.
*
* If initialseed is not NULL, the value it points to is used to seed the
* initial value of the CRC32C.
*
* If flags include DMA_CRC_STORE and not DMA_CRC_INLINE, crcptr is written
* with the 32-bit CRC32C result (in wire format).
*
* If flags include DMA_CRC_TEST and not DMA_CRC_INLINE, the computed CRC32C is
* compared with the 32-bit CRC32C pointed to by crcptr. If they do not match,
* a channel error is raised.
*
* If the DMA_CRC_INLINE flag is set, crcptr is ignored and the DMA engine uses
* the 4 bytes trailing the source data (TEST) or the destination data (STORE).
*/
struct bus_dmadesc *ioat_copy_crc(bus_dmaengine_t dmaengine, bus_addr_t dst,
bus_addr_t src, bus_size_t len, uint32_t *initialseed, bus_addr_t crcptr,
bus_dmaengine_callback_t callback_fn, void *callback_arg, uint32_t flags);
/*
* ioat_crc() is nearly identical to ioat_copy_crc(), but does not actually
* move data around.
*
* Like ioat_copy_crc, ioat_crc computes a CRC32C over len bytes pointed to by
* src. The flags affect its operation in the same way, with one exception:
*
* If flags includes both DMA_CRC_STORE and DMA_CRC_INLINE, the computed CRC32C
* is written to the 4 bytes trailing the *source* data.
*/
struct bus_dmadesc *ioat_crc(bus_dmaengine_t dmaengine, bus_addr_t src,
bus_size_t len, uint32_t *initialseed, bus_addr_t crcptr,
bus_dmaengine_callback_t callback_fn, void *callback_arg, uint32_t flags);
/*
* Issues a null operation. This issues the operation to the hardware, but the
* hardware doesn't do anything with it.
*/
struct bus_dmadesc *ioat_null(bus_dmaengine_t dmaengine,
bus_dmaengine_callback_t callback_fn, void *callback_arg, uint32_t flags);
#endif /* __IOAT_H__ */