2015-08-24 19:32:03 +00:00
|
|
|
/*-
|
|
|
|
* Copyright (C) 2012 Intel Corporation
|
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
|
|
* SUCH DAMAGE.
|
|
|
|
*/
|
|
|
|
|
|
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
|
|
|
|
#ifndef __IOAT_H__
|
|
|
|
#define __IOAT_H__
|
|
|
|
|
|
|
|
#include <sys/param.h>
|
|
|
|
#include <machine/bus.h>
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This file defines the public interface to the IOAT driver.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Enables an interrupt for this operation. Typically, you would only enable
|
|
|
|
* this on the last operation in a group
|
|
|
|
*/
|
|
|
|
#define DMA_INT_EN 0x1
|
2015-10-26 03:30:38 +00:00
|
|
|
/*
|
|
|
|
* Like M_NOWAIT. Operations will return NULL if they cannot allocate a
|
|
|
|
* descriptor without blocking.
|
|
|
|
*/
|
|
|
|
#define DMA_NO_WAIT 0x2
|
2016-01-15 01:34:43 +00:00
|
|
|
/*
|
|
|
|
* Disallow prefetching the source of the following operation. Ordinarily, DMA
|
|
|
|
* operations can be pipelined on some hardware. E.g., operation 2's source
|
|
|
|
* may be prefetched before operation 1 completes.
|
|
|
|
*/
|
|
|
|
#define DMA_FENCE 0x4
|
2016-05-03 17:07:18 +00:00
|
|
|
#define _DMA_GENERIC_FLAGS (DMA_INT_EN | DMA_NO_WAIT | DMA_FENCE)
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Emit a CRC32C as the result of a ioat_copy_crc() or ioat_crc().
|
|
|
|
*/
|
|
|
|
#define DMA_CRC_STORE 0x8
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Compare the CRC32C of a ioat_copy_crc() or ioat_crc() against an expeceted
|
|
|
|
* value. It is invalid to specify both TEST and STORE.
|
|
|
|
*/
|
|
|
|
#define DMA_CRC_TEST 0x10
|
|
|
|
#define _DMA_CRC_TESTSTORE (DMA_CRC_STORE | DMA_CRC_TEST)
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Use an inline comparison CRC32C or emit an inline CRC32C result. Invalid
|
|
|
|
* without one of STORE or TEST.
|
|
|
|
*/
|
|
|
|
#define DMA_CRC_INLINE 0x20
|
|
|
|
#define _DMA_CRC_FLAGS (DMA_CRC_STORE | DMA_CRC_TEST | DMA_CRC_INLINE)
|
2015-08-24 19:32:03 +00:00
|
|
|
|
2015-12-17 23:21:37 +00:00
|
|
|
/*
|
|
|
|
* Hardware revision number. Different hardware revisions support different
|
|
|
|
* features. For example, 3.2 cannot read from MMIO space, while 3.3 can.
|
|
|
|
*/
|
|
|
|
#define IOAT_VER_3_0 0x30
|
|
|
|
#define IOAT_VER_3_2 0x32
|
|
|
|
#define IOAT_VER_3_3 0x33
|
|
|
|
|
2016-07-12 21:56:49 +00:00
|
|
|
/*
|
|
|
|
* Hardware capabilities. Different hardware revisions support different
|
|
|
|
* features. It is often useful to detect specific features than try to infer
|
|
|
|
* them from hardware version.
|
|
|
|
*
|
|
|
|
* Different channels may support different features too; for example, 'PQ' may
|
|
|
|
* only be supported on the first two channels of some hardware.
|
|
|
|
*/
|
|
|
|
#define IOAT_DMACAP_PB (1 << 0)
|
|
|
|
#define IOAT_DMACAP_CRC (1 << 1)
|
|
|
|
#define IOAT_DMACAP_MARKER_SKIP (1 << 2)
|
|
|
|
#define IOAT_DMACAP_OLD_XOR (1 << 3)
|
|
|
|
#define IOAT_DMACAP_DCA (1 << 4)
|
|
|
|
#define IOAT_DMACAP_MOVECRC (1 << 5)
|
|
|
|
#define IOAT_DMACAP_BFILL (1 << 6)
|
|
|
|
#define IOAT_DMACAP_EXT_APIC (1 << 7)
|
|
|
|
#define IOAT_DMACAP_XOR (1 << 8)
|
|
|
|
#define IOAT_DMACAP_PQ (1 << 9)
|
|
|
|
#define IOAT_DMACAP_DMA_DIF (1 << 10)
|
|
|
|
#define IOAT_DMACAP_DWBES (1 << 13)
|
|
|
|
#define IOAT_DMACAP_RAID16SS (1 << 17)
|
|
|
|
#define IOAT_DMACAP_DMAMC (1 << 18)
|
|
|
|
#define IOAT_DMACAP_CTOS (1 << 19)
|
|
|
|
|
|
|
|
#define IOAT_DMACAP_STR \
|
|
|
|
"\20\24Completion_Timeout_Support\23DMA_with_Multicasting_Support" \
|
|
|
|
"\22RAID_Super_descriptors\16Descriptor_Write_Back_Error_Support" \
|
|
|
|
"\13DMA_with_DIF\12PQ\11XOR\10Extended_APIC_ID\07Block_Fill\06Move_CRC" \
|
|
|
|
"\05DCA\04Old_XOR\03Marker_Skipping\02CRC\01Page_Break"
|
|
|
|
|
2015-08-24 19:32:03 +00:00
|
|
|
typedef void *bus_dmaengine_t;
|
|
|
|
struct bus_dmadesc;
|
2015-10-31 20:38:06 +00:00
|
|
|
typedef void (*bus_dmaengine_callback_t)(void *arg, int error);
|
2015-08-24 19:32:03 +00:00
|
|
|
|
2016-06-04 03:54:30 +00:00
|
|
|
unsigned ioat_get_nchannels(void);
|
|
|
|
|
2015-08-24 19:32:03 +00:00
|
|
|
/*
|
|
|
|
* Called first to acquire a reference to the DMA channel
|
2016-04-09 13:15:34 +00:00
|
|
|
*
|
|
|
|
* Flags may be M_WAITOK or M_NOWAIT.
|
2015-08-24 19:32:03 +00:00
|
|
|
*/
|
2016-04-09 13:15:34 +00:00
|
|
|
bus_dmaengine_t ioat_get_dmaengine(uint32_t channel_index, int flags);
|
2015-08-24 19:32:03 +00:00
|
|
|
|
2015-10-24 23:45:33 +00:00
|
|
|
/* Release the DMA channel */
|
|
|
|
void ioat_put_dmaengine(bus_dmaengine_t dmaengine);
|
|
|
|
|
2015-12-17 23:21:37 +00:00
|
|
|
/* Check the DMA engine's HW version */
|
|
|
|
int ioat_get_hwversion(bus_dmaengine_t dmaengine);
|
2016-01-05 20:42:19 +00:00
|
|
|
size_t ioat_get_max_io_size(bus_dmaengine_t dmaengine);
|
2016-07-12 21:56:49 +00:00
|
|
|
uint32_t ioat_get_capabilities(bus_dmaengine_t dmaengine);
|
2015-12-17 23:21:37 +00:00
|
|
|
|
2015-12-14 22:01:52 +00:00
|
|
|
/*
|
|
|
|
* Set interrupt coalescing on a DMA channel.
|
|
|
|
*
|
|
|
|
* The argument is in microseconds. A zero value disables coalescing. Any
|
|
|
|
* other value delays interrupt generation for N microseconds to provide
|
|
|
|
* opportunity to coalesce multiple operations into a single interrupt.
|
|
|
|
*
|
|
|
|
* Returns an error status, or zero on success.
|
|
|
|
*
|
|
|
|
* - ERANGE if the given value exceeds the delay supported by the hardware.
|
|
|
|
* (All current hardware supports a maximum of 0x3fff microseconds delay.)
|
|
|
|
* - ENODEV if the hardware does not support interrupt coalescing.
|
|
|
|
*/
|
|
|
|
int ioat_set_interrupt_coalesce(bus_dmaengine_t dmaengine, uint16_t delay);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Return the maximum supported coalescing period, for use in
|
|
|
|
* ioat_set_interrupt_coalesce(). If the hardware does not support coalescing,
|
|
|
|
* returns zero.
|
|
|
|
*/
|
|
|
|
uint16_t ioat_get_max_coalesce_period(bus_dmaengine_t dmaengine);
|
|
|
|
|
2015-08-24 19:32:03 +00:00
|
|
|
/*
|
|
|
|
* Acquire must be called before issuing an operation to perform. Release is
|
2016-01-07 23:02:15 +00:00
|
|
|
* called after. Multiple operations can be issued within the context of one
|
2015-08-24 19:32:03 +00:00
|
|
|
* acquire and release
|
|
|
|
*/
|
|
|
|
void ioat_acquire(bus_dmaengine_t dmaengine);
|
|
|
|
void ioat_release(bus_dmaengine_t dmaengine);
|
|
|
|
|
2016-01-07 23:02:15 +00:00
|
|
|
/*
|
|
|
|
* Acquire_reserve can be called to ensure there is room for N descriptors. If
|
|
|
|
* it succeeds, the next N valid operations will successfully enqueue.
|
|
|
|
*
|
|
|
|
* It may fail with:
|
|
|
|
* - ENXIO if the channel is in an errored state, or the driver is being
|
|
|
|
* unloaded
|
|
|
|
* - EAGAIN if mflags included M_NOWAIT
|
|
|
|
*
|
|
|
|
* On failure, the caller does not hold the dmaengine.
|
|
|
|
*/
|
|
|
|
int ioat_acquire_reserve(bus_dmaengine_t dmaengine, unsigned n, int mflags);
|
|
|
|
|
2015-10-26 19:34:12 +00:00
|
|
|
/*
|
|
|
|
* Issue a blockfill operation. The 64-bit pattern 'fillpattern' is written to
|
|
|
|
* 'len' physically contiguous bytes at 'dst'.
|
2015-10-28 02:37:24 +00:00
|
|
|
*
|
|
|
|
* Only supported on devices with the BFILL capability.
|
2015-10-26 19:34:12 +00:00
|
|
|
*/
|
|
|
|
struct bus_dmadesc *ioat_blockfill(bus_dmaengine_t dmaengine, bus_addr_t dst,
|
|
|
|
uint64_t fillpattern, bus_size_t len, bus_dmaengine_callback_t callback_fn,
|
|
|
|
void *callback_arg, uint32_t flags);
|
|
|
|
|
2015-08-24 19:32:03 +00:00
|
|
|
/* Issues the copy data operation */
|
|
|
|
struct bus_dmadesc *ioat_copy(bus_dmaengine_t dmaengine, bus_addr_t dst,
|
|
|
|
bus_addr_t src, bus_size_t len, bus_dmaengine_callback_t callback_fn,
|
|
|
|
void *callback_arg, uint32_t flags);
|
|
|
|
|
2015-12-09 22:45:51 +00:00
|
|
|
/*
|
|
|
|
* Issue a copy data operation, with constraints:
|
|
|
|
* - src1, src2, dst1, dst2 are all page-aligned addresses
|
|
|
|
* - The quantity to copy is exactly 2 pages;
|
|
|
|
* - src1 -> dst1, src2 -> dst2
|
|
|
|
*
|
|
|
|
* Why use this instead of normal _copy()? You can copy two non-contiguous
|
|
|
|
* pages (src, dst, or both) with one descriptor.
|
|
|
|
*/
|
|
|
|
struct bus_dmadesc *ioat_copy_8k_aligned(bus_dmaengine_t dmaengine,
|
|
|
|
bus_addr_t dst1, bus_addr_t dst2, bus_addr_t src1, bus_addr_t src2,
|
|
|
|
bus_dmaengine_callback_t callback_fn, void *callback_arg, uint32_t flags);
|
|
|
|
|
2016-05-03 17:07:18 +00:00
|
|
|
/*
|
|
|
|
* Copy len bytes from dst to src, like ioat_copy().
|
|
|
|
*
|
|
|
|
* Additionally, accumulate a CRC32C of the data.
|
|
|
|
*
|
|
|
|
* If initialseed is not NULL, the value it points to is used to seed the
|
|
|
|
* initial value of the CRC32C.
|
|
|
|
*
|
|
|
|
* If flags include DMA_CRC_STORE and not DMA_CRC_INLINE, crcptr is written
|
|
|
|
* with the 32-bit CRC32C result (in wire format).
|
|
|
|
*
|
|
|
|
* If flags include DMA_CRC_TEST and not DMA_CRC_INLINE, the computed CRC32C is
|
|
|
|
* compared with the 32-bit CRC32C pointed to by crcptr. If they do not match,
|
|
|
|
* a channel error is raised.
|
|
|
|
*
|
|
|
|
* If the DMA_CRC_INLINE flag is set, crcptr is ignored and the DMA engine uses
|
|
|
|
* the 4 bytes trailing the source data (TEST) or the destination data (STORE).
|
|
|
|
*/
|
|
|
|
struct bus_dmadesc *ioat_copy_crc(bus_dmaengine_t dmaengine, bus_addr_t dst,
|
|
|
|
bus_addr_t src, bus_size_t len, uint32_t *initialseed, bus_addr_t crcptr,
|
|
|
|
bus_dmaengine_callback_t callback_fn, void *callback_arg, uint32_t flags);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* ioat_crc() is nearly identical to ioat_copy_crc(), but does not actually
|
|
|
|
* move data around.
|
|
|
|
*
|
|
|
|
* Like ioat_copy_crc, ioat_crc computes a CRC32C over len bytes pointed to by
|
|
|
|
* src. The flags affect its operation in the same way, with one exception:
|
|
|
|
*
|
|
|
|
* If flags includes both DMA_CRC_STORE and DMA_CRC_INLINE, the computed CRC32C
|
|
|
|
* is written to the 4 bytes trailing the *source* data.
|
|
|
|
*/
|
|
|
|
struct bus_dmadesc *ioat_crc(bus_dmaengine_t dmaengine, bus_addr_t src,
|
|
|
|
bus_size_t len, uint32_t *initialseed, bus_addr_t crcptr,
|
|
|
|
bus_dmaengine_callback_t callback_fn, void *callback_arg, uint32_t flags);
|
|
|
|
|
2015-08-24 19:32:03 +00:00
|
|
|
/*
|
|
|
|
* Issues a null operation. This issues the operation to the hardware, but the
|
|
|
|
* hardware doesn't do anything with it.
|
|
|
|
*/
|
|
|
|
struct bus_dmadesc *ioat_null(bus_dmaengine_t dmaengine,
|
|
|
|
bus_dmaengine_callback_t callback_fn, void *callback_arg, uint32_t flags);
|
|
|
|
|
|
|
|
|
|
|
|
#endif /* __IOAT_H__ */
|
|
|
|
|