freebsd-skq/sys/dev/vxge/vxgehal/vxgehal-ring.c
eadler 9d7884364e - change "is is" to "is" or "it is"
- change "the the" to "the"

Approved by:	lstewart
Approved by:	sahil (mentor)
MFC after:	3 days
2011-10-16 14:30:28 +00:00

1814 lines
49 KiB
C

/*-
* Copyright(c) 2002-2011 Exar Corp.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification are permitted provided the following conditions are met:
*
* 1. Redistributions of source code must retain the above copyright notice,
* this list of conditions and the following disclaimer.
*
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* 3. Neither the name of the Exar Corporation nor the names of its
* contributors may be used to endorse or promote products derived from
* this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/
/*$FreeBSD$*/
#include <dev/vxge/vxgehal/vxgehal.h>
/*
* __hal_ring_block_memblock_idx - Return the memblock index
* @block: Virtual address of memory block
*
* This function returns the index of memory block
*/
static inline u32
__hal_ring_block_memblock_idx(
vxge_hal_ring_block_t block)
{
return (u32)*((u64 *) ((void *)((u8 *) block +
VXGE_HAL_RING_MEMBLOCK_IDX_OFFSET)));
}
/*
* __hal_ring_block_memblock_idx_set - Sets the memblock index
* @block: Virtual address of memory block
* @memblock_idx: Index of memory block
*
* This function sets index to a memory block
*/
static inline void
__hal_ring_block_memblock_idx_set(
vxge_hal_ring_block_t block,
u32 memblock_idx)
{
*((u64 *) ((void *)((u8 *) block +
VXGE_HAL_RING_MEMBLOCK_IDX_OFFSET))) = memblock_idx;
}
/*
* __hal_ring_block_next_pointer - Returns the dma address of next block
* @block: RxD block
*
* Returns the dma address of next block stored in the RxD block
*/
static inline dma_addr_t
/* LINTED */
__hal_ring_block_next_pointer(
vxge_hal_ring_block_t *block)
{
return (dma_addr_t)*((u64 *) ((void *)((u8 *) block +
VXGE_HAL_RING_NEXT_BLOCK_POINTER_OFFSET)));
}
/*
* __hal_ring_block_next_pointer_set - Sets the next block pointer in RxD block
* @block: RxD block
* @dma_next: dma address of next block
*
* Sets the next block pointer in RxD block
*/
static inline void
__hal_ring_block_next_pointer_set(
vxge_hal_ring_block_t *block,
dma_addr_t dma_next)
{
*((u64 *) ((void *)((u8 *) block +
VXGE_HAL_RING_NEXT_BLOCK_POINTER_OFFSET))) = dma_next;
}
/*
* __hal_ring_first_block_address_get - Returns the dma address of the
* first block
* @ringh: Handle to the ring
*
* Returns the dma address of the first RxD block
*/
u64
__hal_ring_first_block_address_get(
vxge_hal_ring_h ringh)
{
__hal_ring_t *ring = (__hal_ring_t *) ringh;
vxge_hal_mempool_dma_t *dma_object;
dma_object = __hal_mempool_memblock_dma(ring->mempool, 0);
vxge_assert(dma_object != NULL);
return (dma_object->addr);
}
#if defined(VXGE_OS_DMA_REQUIRES_SYNC) && defined(VXGE_HAL_DMA_RXD_STREAMING)
/*
* __hal_ring_item_dma_offset - Return the dma offset of an item
* @mempoolh: Handle to the memory pool of the ring
* @item: Item for which to get the dma offset
*
* This function returns the dma offset of a given item
*/
static ptrdiff_t
__hal_ring_item_dma_offset(
vxge_hal_mempool_h mempoolh,
void *item)
{
u32 memblock_idx;
void *memblock;
vxge_hal_mempool_t *mempool = (vxge_hal_mempool_t *) mempoolh;
__hal_device_t *hldev;
vxge_assert((mempoolh != NULL) && (item != NULL) &&
(dma_handle != NULL));
hldev = (__hal_device_t *) mempool->devh;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"mempoolh = 0x"VXGE_OS_STXFMT", item = 0x"VXGE_OS_STXFMT,
(ptr_t) mempoolh, (ptr_t) item);
/* get owner memblock index */
memblock_idx = __hal_ring_block_memblock_idx(item);
/* get owner memblock by memblock index */
memblock = __hal_mempool_memblock(mempoolh, memblock_idx);
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
return ((u8 *) item - (u8 *) memblock);
}
#endif
/*
* __hal_ring_item_dma_addr - Return the dma address of an item
* @mempoolh: Handle to the memory pool of the ring
* @item: Item for which to get the dma offset
* @dma_handle: dma handle
*
* This function returns the dma address of a given item
*/
static dma_addr_t
__hal_ring_item_dma_addr(
vxge_hal_mempool_h mempoolh,
void *item,
pci_dma_h *dma_handle)
{
u32 memblock_idx;
void *memblock;
vxge_hal_mempool_dma_t *memblock_dma_object;
vxge_hal_mempool_t *mempool = (vxge_hal_mempool_t *) mempoolh;
__hal_device_t *hldev;
ptrdiff_t dma_item_offset;
vxge_assert((mempoolh != NULL) && (item != NULL) &&
(dma_handle != NULL));
hldev = (__hal_device_t *) mempool->devh;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"mempoolh = 0x"VXGE_OS_STXFMT", item = 0x"VXGE_OS_STXFMT", "
"dma_handle = 0x"VXGE_OS_STXFMT, (ptr_t) mempoolh,
(ptr_t) item, (ptr_t) dma_handle);
/* get owner memblock index */
memblock_idx = __hal_ring_block_memblock_idx((u8 *) item);
/* get owner memblock by memblock index */
memblock = __hal_mempool_memblock(
(vxge_hal_mempool_t *) mempoolh, memblock_idx);
/* get memblock DMA object by memblock index */
memblock_dma_object = __hal_mempool_memblock_dma(
(vxge_hal_mempool_t *) mempoolh, memblock_idx);
/* calculate offset in the memblock of this item */
/* LINTED */
dma_item_offset = (u8 *) item - (u8 *) memblock;
*dma_handle = memblock_dma_object->handle;
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
return (memblock_dma_object->addr + dma_item_offset);
}
/*
* __hal_ring_rxdblock_link - Link the RxD blocks
* @mempoolh: Handle to the memory pool of the ring
* @ring: ring
* @from: RxD block from which to link
* @to: RxD block to which to link to
*
* This function returns the dma address of a given item
*/
static void
__hal_ring_rxdblock_link(
vxge_hal_mempool_h mempoolh,
__hal_ring_t *ring,
u32 from,
u32 to)
{
vxge_hal_ring_block_t *to_item, *from_item;
dma_addr_t to_dma, from_dma;
pci_dma_h to_dma_handle, from_dma_handle;
__hal_device_t *hldev;
vxge_assert((mempoolh != NULL) && (ring != NULL));
hldev = (__hal_device_t *) ring->channel.devh;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"mempoolh = 0x"VXGE_OS_STXFMT", ring = 0x"VXGE_OS_STXFMT", "
"from = %d, to = %d", (ptr_t) mempoolh, (ptr_t) ring, from, to);
/* get "from" RxD block */
from_item = (vxge_hal_ring_block_t *) __hal_mempool_item(
(vxge_hal_mempool_t *) mempoolh, from);
vxge_assert(from_item);
/* get "to" RxD block */
to_item = (vxge_hal_ring_block_t *) __hal_mempool_item(
(vxge_hal_mempool_t *) mempoolh, to);
vxge_assert(to_item);
/* return address of the beginning of previous RxD block */
to_dma = __hal_ring_item_dma_addr(mempoolh, to_item, &to_dma_handle);
/*
* set next pointer for this RxD block to point on
* previous item's DMA start address
*/
__hal_ring_block_next_pointer_set(from_item, to_dma);
/* return "from" RxD block's DMA start address */
from_dma = __hal_ring_item_dma_addr(
mempoolh, from_item, &from_dma_handle);
#if defined(VXGE_OS_DMA_REQUIRES_SYNC) && defined(VXGE_HAL_DMA_RXD_STREAMING)
/* we must sync "from" RxD block, so hardware will see it */
vxge_os_dma_sync(ring->channel.pdev,
from_dma_handle,
from_dma + VXGE_HAL_RING_NEXT_BLOCK_POINTER_OFFSET,
__hal_ring_item_dma_offset(mempoolh, from_item) +
VXGE_HAL_RING_NEXT_BLOCK_POINTER_OFFSET,
sizeof(u64),
VXGE_OS_DMA_DIR_TODEVICE);
#endif
vxge_hal_info_log_ring(
"block%d:0x"VXGE_OS_STXFMT" => block%d:0x"VXGE_OS_STXFMT,
from, (ptr_t) from_dma, to, (ptr_t) to_dma);
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
}
/*
* __hal_ring_mempool_item_alloc - Allocate List blocks for RxD block callback
* @mempoolh: Handle to memory pool
* @memblock: Address of this memory block
* @memblock_index: Index of this memory block
* @dma_object: dma object for this block
* @item: Pointer to this item
* @index: Index of this item in memory block
* @is_last: If this is last item in the block
* @userdata: Specific data of user
*
* This function is callback passed to __hal_mempool_create to create memory
* pool for RxD block
*/
static vxge_hal_status_e
__hal_ring_mempool_item_alloc(
vxge_hal_mempool_h mempoolh,
void *memblock,
u32 memblock_index,
vxge_hal_mempool_dma_t *dma_object,
void *item,
u32 item_index,
u32 is_last,
void *userdata)
{
u32 i;
__hal_ring_t *ring = (__hal_ring_t *) userdata;
__hal_device_t *hldev;
vxge_assert((item != NULL) && (ring != NULL));
hldev = (__hal_device_t *) ring->channel.devh;
vxge_hal_trace_log_pool("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_pool(
"mempoolh = 0x"VXGE_OS_STXFMT", memblock = 0x"VXGE_OS_STXFMT", "
"memblock_index = %d, dma_object = 0x"VXGE_OS_STXFMT", "
"item = 0x"VXGE_OS_STXFMT", item_index = %d, is_last = %d, "
"userdata = 0x"VXGE_OS_STXFMT, (ptr_t) mempoolh, (ptr_t) memblock,
memblock_index, (ptr_t) dma_object, (ptr_t) item, item_index, is_last,
(ptr_t) userdata);
/* format rxds array */
for (i = 0; i < ring->rxds_per_block; i++) {
void *uld_priv;
void *rxdblock_priv;
__hal_ring_rxd_priv_t *rxd_priv;
vxge_hal_ring_rxd_1_t *rxdp;
u32 memblock_item_idx;
u32 dtr_index = item_index * ring->rxds_per_block + i;
ring->channel.dtr_arr[dtr_index].dtr =
((u8 *) item) + i * ring->rxd_size;
/*
* Note: memblock_item_idx is index of the item within
* the memblock. For instance, in case of three RxD-blocks
* per memblock this value can be 0, 1 or 2.
*/
rxdblock_priv = __hal_mempool_item_priv(
(vxge_hal_mempool_t *) mempoolh,
memblock_index,
item,
&memblock_item_idx);
rxdp = (vxge_hal_ring_rxd_1_t *)
ring->channel.dtr_arr[dtr_index].dtr;
uld_priv = ((u8 *) rxdblock_priv + ring->rxd_priv_size * i);
rxd_priv =
(__hal_ring_rxd_priv_t *) ((void *)(((char *) uld_priv) +
ring->per_rxd_space));
((vxge_hal_ring_rxd_5_t *) rxdp)->host_control = dtr_index;
ring->channel.dtr_arr[dtr_index].uld_priv = (void *)uld_priv;
ring->channel.dtr_arr[dtr_index].hal_priv = (void *)rxd_priv;
/* pre-format per-RxD Ring's private */
/* LINTED */
rxd_priv->dma_offset = (u8 *) rxdp - (u8 *) memblock;
rxd_priv->dma_addr = dma_object->addr + rxd_priv->dma_offset;
rxd_priv->dma_handle = dma_object->handle;
#if defined(VXGE_DEBUG_ASSERT)
rxd_priv->dma_object = dma_object;
#endif
rxd_priv->db_bytes = ring->rxd_size;
if (i == (ring->rxds_per_block - 1)) {
rxd_priv->db_bytes +=
(((vxge_hal_mempool_t *) mempoolh)->memblock_size -
(ring->rxds_per_block * ring->rxd_size));
}
}
__hal_ring_block_memblock_idx_set((u8 *) item, memblock_index);
if (is_last) {
/* link last one with first one */
__hal_ring_rxdblock_link(mempoolh, ring, item_index, 0);
}
if (item_index > 0) {
/* link this RxD block with previous one */
__hal_ring_rxdblock_link(mempoolh, ring, item_index - 1, item_index);
}
vxge_hal_trace_log_pool("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
return (VXGE_HAL_OK);
}
/*
* __hal_ring_mempool_item_free - Free RxD blockt callback
* @mempoolh: Handle to memory pool
* @memblock: Address of this memory block
* @memblock_index: Index of this memory block
* @dma_object: dma object for this block
* @item: Pointer to this item
* @index: Index of this item in memory block
* @is_last: If this is last item in the block
* @userdata: Specific data of user
*
* This function is callback passed to __hal_mempool_free to destroy memory
* pool for RxD block
*/
static vxge_hal_status_e
__hal_ring_mempool_item_free(
vxge_hal_mempool_h mempoolh,
void *memblock,
u32 memblock_index,
vxge_hal_mempool_dma_t *dma_object,
void *item,
u32 item_index,
u32 is_last,
void *userdata)
{
__hal_ring_t *ring = (__hal_ring_t *) userdata;
__hal_device_t *hldev;
vxge_assert((item != NULL) && (ring != NULL));
hldev = (__hal_device_t *) ring->channel.devh;
vxge_hal_trace_log_pool("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_pool(
"mempoolh = 0x"VXGE_OS_STXFMT", memblock = 0x"VXGE_OS_STXFMT", "
"memblock_index = %d, dma_object = 0x"VXGE_OS_STXFMT", "
"item = 0x"VXGE_OS_STXFMT", item_index = %d, is_last = %d, "
"userdata = 0x"VXGE_OS_STXFMT, (ptr_t) mempoolh, (ptr_t) memblock,
memblock_index, (ptr_t) dma_object, (ptr_t) item, item_index, is_last,
(ptr_t) userdata);
vxge_hal_trace_log_pool("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
return (VXGE_HAL_OK);
}
/*
* __hal_ring_initial_replenish - Initial replenish of RxDs
* @ring: ring
* @reopen: Flag to denote if it is open or repopen
*
* This function replenishes the RxDs from reserve array to work array
*/
static vxge_hal_status_e
__hal_ring_initial_replenish(
__hal_ring_t *ring,
vxge_hal_reopen_e reopen)
{
vxge_hal_rxd_h rxd;
void *uld_priv;
__hal_device_t *hldev;
vxge_hal_status_e status;
vxge_assert(ring != NULL);
hldev = (__hal_device_t *) ring->channel.devh;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring("ring = 0x"VXGE_OS_STXFMT", reopen = %d",
(ptr_t) ring, reopen);
while (vxge_hal_ring_rxd_reserve(ring->channel.vph, &rxd, &uld_priv) ==
VXGE_HAL_OK) {
if (ring->rxd_init) {
status = ring->rxd_init(ring->channel.vph,
rxd,
uld_priv,
VXGE_HAL_RING_RXD_INDEX(rxd),
ring->channel.userdata,
reopen);
if (status != VXGE_HAL_OK) {
vxge_hal_ring_rxd_free(ring->channel.vph, rxd);
vxge_hal_trace_log_ring("<== %s:%s:%d \
Result: %d",
__FILE__, __func__, __LINE__, status);
return (status);
}
}
vxge_hal_ring_rxd_post(ring->channel.vph, rxd);
}
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
return (VXGE_HAL_OK);
}
/*
* __hal_ring_create - Create a Ring
* @vpath_handle: Handle returned by virtual path open
* @attr: Ring configuration parameters structure
*
* This function creates Ring and initializes it.
*
*/
vxge_hal_status_e
__hal_ring_create(
vxge_hal_vpath_h vpath_handle,
vxge_hal_ring_attr_t *attr)
{
vxge_hal_status_e status;
__hal_ring_t *ring;
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
vxge_hal_ring_config_t *config;
__hal_device_t *hldev;
vxge_assert((vpath_handle != NULL) && (attr != NULL));
hldev = (__hal_device_t *) vp->vpath->hldev;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"vpath_handle = 0x"VXGE_OS_STXFMT", attr = 0x"VXGE_OS_STXFMT,
(ptr_t) vpath_handle, (ptr_t) attr);
if ((vpath_handle == NULL) || (attr == NULL)) {
vxge_hal_err_log_ring("null pointer passed == > %s : %d",
__func__, __LINE__);
vxge_hal_trace_log_ring("<== %s:%s:%d Result:1",
__FILE__, __func__, __LINE__);
return (VXGE_HAL_FAIL);
}
config =
&vp->vpath->hldev->header.config.vp_config[vp->vpath->vp_id].ring;
config->ring_length = ((config->ring_length +
vxge_hal_ring_rxds_per_block_get(config->buffer_mode) - 1) /
vxge_hal_ring_rxds_per_block_get(config->buffer_mode)) *
vxge_hal_ring_rxds_per_block_get(config->buffer_mode);
ring = (__hal_ring_t *) vxge_hal_channel_allocate(
(vxge_hal_device_h) vp->vpath->hldev,
vpath_handle,
VXGE_HAL_CHANNEL_TYPE_RING,
config->ring_length,
attr->per_rxd_space,
attr->userdata);
if (ring == NULL) {
vxge_hal_err_log_ring("Memory allocation failed == > %s : %d",
__func__, __LINE__);
vxge_hal_trace_log_ring("<== %s:%s:%d Result: %d",
__FILE__, __func__, __LINE__,
VXGE_HAL_ERR_OUT_OF_MEMORY);
return (VXGE_HAL_ERR_OUT_OF_MEMORY);
}
vp->vpath->ringh = (vxge_hal_ring_h) ring;
ring->stats = &vp->vpath->sw_stats->ring_stats;
ring->config = config;
ring->callback = attr->callback;
ring->rxd_init = attr->rxd_init;
ring->rxd_term = attr->rxd_term;
ring->indicate_max_pkts = config->indicate_max_pkts;
ring->buffer_mode = config->buffer_mode;
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_lock_init(&ring->channel.post_lock, hldev->pdev);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_lock_init_irq(&ring->channel.post_lock, hldev->irqh);
#endif
ring->rxd_size = vxge_hal_ring_rxd_size_get(config->buffer_mode);
ring->rxd_priv_size =
sizeof(__hal_ring_rxd_priv_t) + attr->per_rxd_space;
ring->per_rxd_space = attr->per_rxd_space;
ring->rxd_priv_size =
((ring->rxd_priv_size + __vxge_os_cacheline_size - 1) /
__vxge_os_cacheline_size) * __vxge_os_cacheline_size;
/*
* how many RxDs can fit into one block. Depends on configured
* buffer_mode.
*/
ring->rxds_per_block =
vxge_hal_ring_rxds_per_block_get(config->buffer_mode);
/* calculate actual RxD block private size */
ring->rxdblock_priv_size = ring->rxd_priv_size * ring->rxds_per_block;
ring->rxd_mem_avail =
((__hal_vpath_handle_t *) ring->channel.vph)->vpath->rxd_mem_size;
ring->db_byte_count = 0;
ring->mempool = vxge_hal_mempool_create(
(vxge_hal_device_h) vp->vpath->hldev,
VXGE_OS_HOST_PAGE_SIZE,
VXGE_OS_HOST_PAGE_SIZE,
ring->rxdblock_priv_size,
ring->config->ring_length / ring->rxds_per_block,
ring->config->ring_length / ring->rxds_per_block,
__hal_ring_mempool_item_alloc,
__hal_ring_mempool_item_free,
ring);
if (ring->mempool == NULL) {
__hal_ring_delete(vpath_handle);
vxge_hal_trace_log_ring("<== %s:%s:%d Result: %d",
__FILE__, __func__, __LINE__, VXGE_HAL_ERR_OUT_OF_MEMORY);
return (VXGE_HAL_ERR_OUT_OF_MEMORY);
}
status = vxge_hal_channel_initialize(&ring->channel);
if (status != VXGE_HAL_OK) {
__hal_ring_delete(vpath_handle);
vxge_hal_trace_log_ring("<== %s:%s:%d Result: %d",
__FILE__, __func__, __LINE__, status);
return (status);
}
/*
* Note:
* Specifying rxd_init callback means two things:
* 1) rxds need to be initialized by ULD at channel-open time;
* 2) rxds need to be posted at channel-open time
* (that's what the initial_replenish() below does)
* Currently we don't have a case when the 1) is done without the 2).
*/
if (ring->rxd_init) {
if ((status = __hal_ring_initial_replenish(
ring,
VXGE_HAL_OPEN_NORMAL))
!= VXGE_HAL_OK) {
__hal_ring_delete(vpath_handle);
vxge_hal_trace_log_ring("<== %s:%s:%d Result: %d",
__FILE__, __func__, __LINE__, status);
return (status);
}
}
/*
* initial replenish will increment the counter in its post() routine,
* we have to reset it
*/
ring->stats->common_stats.usage_cnt = 0;
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
return (VXGE_HAL_OK);
}
/*
* __hal_ring_abort - Returns the RxD
* @ringh: Ring to be reset
* @reopen: See vxge_hal_reopen_e {}.
*
* This function terminates the RxDs of ring
*/
void
__hal_ring_abort(
vxge_hal_ring_h ringh,
vxge_hal_reopen_e reopen)
{
u32 i = 0;
vxge_hal_rxd_h rxdh;
__hal_device_t *hldev;
__hal_ring_t *ring = (__hal_ring_t *) ringh;
vxge_assert(ringh != NULL);
hldev = (__hal_device_t *) ring->channel.devh;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring("ring = 0x"VXGE_OS_STXFMT", reopen = %d",
(ptr_t) ringh, reopen);
if (ring->rxd_term) {
__hal_channel_for_each_dtr(&ring->channel, rxdh, i) {
if (!__hal_channel_is_posted_dtr(&ring->channel, i)) {
ring->rxd_term(ring->channel.vph, rxdh,
VXGE_HAL_RING_ULD_PRIV(ring, rxdh),
VXGE_HAL_RXD_STATE_FREED,
ring->channel.userdata,
reopen);
}
}
}
for (;;) {
__hal_channel_dtr_try_complete(&ring->channel, &rxdh);
if (rxdh == NULL)
break;
__hal_channel_dtr_complete(&ring->channel);
if (ring->rxd_term) {
ring->rxd_term(ring->channel.vph, rxdh,
VXGE_HAL_RING_ULD_PRIV(ring, rxdh),
VXGE_HAL_RXD_STATE_POSTED,
ring->channel.userdata,
reopen);
}
__hal_channel_dtr_free(&ring->channel,
VXGE_HAL_RING_RXD_INDEX(rxdh));
}
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
}
/*
* __hal_ring_reset - Resets the ring
* @ringh: Ring to be reset
*
* This function resets the ring during vpath reset operation
*/
vxge_hal_status_e
__hal_ring_reset(
vxge_hal_ring_h ringh)
{
__hal_ring_t *ring = (__hal_ring_t *) ringh;
__hal_device_t *hldev;
vxge_hal_status_e status;
__hal_vpath_handle_t *vph = (__hal_vpath_handle_t *) ring->channel.vph;
vxge_assert(ringh != NULL);
hldev = (__hal_device_t *) ring->channel.devh;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring("ring = 0x"VXGE_OS_STXFMT,
(ptr_t) ringh);
__hal_ring_abort(ringh, VXGE_HAL_RESET_ONLY);
status = __hal_channel_reset(&ring->channel);
if (status != VXGE_HAL_OK) {
vxge_hal_trace_log_ring("<== %s:%s:%d Result: %d",
__FILE__, __func__, __LINE__, status);
return (status);
}
ring->rxd_mem_avail = vph->vpath->rxd_mem_size;
ring->db_byte_count = 0;
if (ring->rxd_init) {
if ((status = __hal_ring_initial_replenish(
ring,
VXGE_HAL_RESET_ONLY))
!= VXGE_HAL_OK) {
vxge_hal_trace_log_ring("<== %s:%s:%d Result: %d",
__FILE__, __func__, __LINE__, status);
return (status);
}
}
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
return (VXGE_HAL_OK);
}
/*
* __hal_ring_delete - Removes the ring
* @vpath_handle: Virtual path handle to which this queue belongs
*
* This function freeup the memory pool and removes the ring
*/
void
__hal_ring_delete(
vxge_hal_vpath_h vpath_handle)
{
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
__hal_device_t *hldev;
__hal_ring_t *ring;
vxge_assert(vpath_handle != NULL);
hldev = (__hal_device_t *) vp->vpath->hldev;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring("vpath_handle = 0x"VXGE_OS_STXFMT,
(ptr_t) vpath_handle);
ring = (__hal_ring_t *) vp->vpath->ringh;
vxge_assert(ring != NULL);
vxge_assert(ring->channel.pdev);
__hal_ring_abort(vp->vpath->ringh, VXGE_HAL_OPEN_NORMAL);
if (ring->mempool) {
vxge_hal_mempool_destroy(ring->mempool);
}
vxge_hal_channel_terminate(&ring->channel);
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_lock_destroy(&ring->channel.post_lock, hldev->pdev);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_lock_destroy_irq(&ring->channel.post_lock, hldev->pdev);
#endif
vxge_hal_channel_free(&ring->channel);
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
}
/*
* __hal_ring_frame_length_set - Set the maximum frame length of recv frames.
* @vpath: virtual Path
* @new_frmlen: New frame length
*
*
* Returns: VXGE_HAL_OK - success.
* VXGE_HAL_INF_OUT_OF_DESCRIPTORS - Currently no descriptors available.
*
*/
vxge_hal_status_e
__hal_ring_frame_length_set(
__hal_virtualpath_t *vpath,
u32 new_frmlen)
{
u64 val64;
__hal_device_t *hldev;
vxge_assert(vpath != NULL);
hldev = (__hal_device_t *) vpath->hldev;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"vpath = 0x"VXGE_OS_STXFMT", new_frmlen = %d",
(ptr_t) vpath, new_frmlen);
if (vpath->vp_open == VXGE_HAL_VP_NOT_OPEN) {
vxge_hal_trace_log_ring("<== %s:%s:%d Result: %d",
__FILE__, __func__, __LINE__,
VXGE_HAL_ERR_VPATH_NOT_OPEN);
return (VXGE_HAL_ERR_VPATH_NOT_OPEN);
}
val64 = vxge_os_pio_mem_read64(
vpath->hldev->header.pdev,
vpath->hldev->header.regh0,
&vpath->vp_reg->rxmac_vcfg0);
val64 &= ~VXGE_HAL_RXMAC_VCFG0_RTS_MAX_FRM_LEN(0x3fff);
if (vpath->vp_config->ring.max_frm_len !=
VXGE_HAL_MAX_RING_FRM_LEN_USE_MTU) {
val64 |= VXGE_HAL_RXMAC_VCFG0_RTS_MAX_FRM_LEN(
vpath->vp_config->ring.max_frm_len +
VXGE_HAL_MAC_HEADER_MAX_SIZE);
} else {
val64 |= VXGE_HAL_RXMAC_VCFG0_RTS_MAX_FRM_LEN(new_frmlen +
VXGE_HAL_MAC_HEADER_MAX_SIZE);
}
vxge_os_pio_mem_write64(
vpath->hldev->header.pdev,
vpath->hldev->header.regh0,
val64,
&vpath->vp_reg->rxmac_vcfg0);
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
return (VXGE_HAL_OK);
}
/*
* vxge_hal_ring_rxd_reserve - Reserve ring descriptor.
* @vpath_handle: virtual Path handle.
* @rxdh: Reserved descriptor. On success HAL fills this "out" parameter
* with a valid handle.
* @rxd_priv: Buffer to return pointer to per rxd private space
*
* Reserve Rx descriptor for the subsequent filling-in (by upper layer
* driver (ULD)) and posting on the corresponding channel (@channelh)
* via vxge_hal_ring_rxd_post().
*
* Returns: VXGE_HAL_OK - success.
* VXGE_HAL_INF_OUT_OF_DESCRIPTORS - Currently no descriptors available.
*
*/
vxge_hal_status_e
vxge_hal_ring_rxd_reserve(
vxge_hal_vpath_h vpath_handle,
vxge_hal_rxd_h * rxdh,
void **rxd_priv)
{
vxge_hal_status_e status;
#if defined(VXGE_HAL_RX_MULTI_POST_IRQ)
unsigned long flags;
#endif
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
__hal_device_t *hldev;
__hal_ring_t *ring;
vxge_assert((vpath_handle != NULL) && (rxdh != NULL) &&
(rxd_priv != NULL));
hldev = (__hal_device_t *) vp->vpath->hldev;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"vpath_handle = 0x"VXGE_OS_STXFMT", rxdh = 0x"VXGE_OS_STXFMT", "
"rxd_priv = 0x"VXGE_OS_STXFMT, (ptr_t) vpath_handle,
(ptr_t) rxdh, (ptr_t) rxd_priv);
ring = (__hal_ring_t *) vp->vpath->ringh;
vxge_assert(ring != NULL);
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_lock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_lock_irq(&ring->channel.post_lock, flags);
#endif
status = __hal_channel_dtr_reserve(&ring->channel, rxdh);
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_unlock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_unlock_irq(&ring->channel.post_lock, flags);
#endif
if (status == VXGE_HAL_OK) {
vxge_hal_ring_rxd_1_t *rxdp = (vxge_hal_ring_rxd_1_t *)*rxdh;
/* instead of memset: reset this RxD */
rxdp->control_0 = rxdp->control_1 = 0;
*rxd_priv = VXGE_HAL_RING_ULD_PRIV(ring, rxdp);
#if defined(VXGE_OS_MEMORY_CHECK)
VXGE_HAL_RING_HAL_PRIV(ring, rxdp)->allocated = 1;
#endif
}
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
return (status);
}
/*
* vxge_hal_ring_rxd_pre_post - Prepare rxd and post
* @vpath_handle: virtual Path handle.
* @rxdh: Descriptor handle.
*
* This routine prepares a rxd and posts
*/
void
vxge_hal_ring_rxd_pre_post(
vxge_hal_vpath_h vpath_handle,
vxge_hal_rxd_h rxdh)
{
#if defined(VXGE_DEBUG_ASSERT)
vxge_hal_ring_rxd_1_t *rxdp = (vxge_hal_ring_rxd_1_t *) rxdh;
#endif
#if defined(VXGE_HAL_RX_MULTI_POST_IRQ)
unsigned long flags;
#endif
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
__hal_device_t *hldev;
__hal_ring_t *ring;
vxge_assert((vpath_handle != NULL) && (rxdh != NULL));
hldev = (__hal_device_t *) vp->vpath->hldev;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"vpath_handle = 0x"VXGE_OS_STXFMT", rxdh = 0x"VXGE_OS_STXFMT,
(ptr_t) vpath_handle, (ptr_t) rxdh);
ring = (__hal_ring_t *) vp->vpath->ringh;
vxge_assert(ring != NULL);
#if defined(VXGE_DEBUG_ASSERT)
/* make sure device overwrites the (illegal) t_code on completion */
rxdp->control_0 |=
VXGE_HAL_RING_RXD_T_CODE(VXGE_HAL_RING_RXD_T_CODE_UNUSED);
#endif
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_lock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_lock_irq(&ring->channel.post_lock, flags);
#endif
#if defined(VXGE_DEBUG_ASSERT) && defined(VXGE_HAL_RING_ENFORCE_ORDER)
if (TRUE) {
if (VXGE_HAL_RING_RXD_INDEX(rxdp) != 0) {
vxge_hal_rxd_h prev_rxdh;
__hal_ring_rxd_priv_t *rxdp_priv;
u32 index;
rxdp_priv = VXGE_HAL_RING_HAL_PRIV(ring, rxdp);
if (VXGE_HAL_RING_RXD_INDEX(rxdp) == 0)
index = ring->channel.length;
else
index = VXGE_HAL_RING_RXD_INDEX(rxdp) - 1;
prev_rxdh = ring->channel.dtr_arr[index].dtr;
if (prev_rxdh != NULL &&
(rxdp_priv->dma_offset & (~0xFFF)) !=
rxdp_priv->dma_offset) {
vxge_assert((char *) prev_rxdh +
ring->rxd_size == rxdh);
}
}
}
#endif
__hal_channel_dtr_post(&ring->channel, VXGE_HAL_RING_RXD_INDEX(rxdh));
ring->db_byte_count +=
VXGE_HAL_RING_HAL_PRIV(ring, rxdh)->db_bytes;
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_unlock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_unlock_irq(&ring->channel.post_lock, flags);
#endif
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
}
/*
* vxge_hal_ring_rxd_post_post - Process rxd after post.
* @vpath_handle: virtual Path handle.
* @rxdh: Descriptor handle.
*
* Processes rxd after post
*/
void
vxge_hal_ring_rxd_post_post(
vxge_hal_vpath_h vpath_handle,
vxge_hal_rxd_h rxdh)
{
vxge_hal_ring_rxd_1_t *rxdp = (vxge_hal_ring_rxd_1_t *) rxdh;
#if defined(VXGE_OS_DMA_REQUIRES_SYNC) && defined(VXGE_HAL_DMA_RXD_STREAMING)
__hal_ring_rxd_priv_t *priv;
#endif
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
__hal_device_t *hldev;
__hal_ring_t *ring;
vxge_assert((vpath_handle != NULL) && (rxdh != NULL));
hldev = (__hal_device_t *) vp->vpath->hldev;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"vpath_handle = 0x"VXGE_OS_STXFMT", rxdh = 0x"VXGE_OS_STXFMT,
(ptr_t) vpath_handle, (ptr_t) rxdh);
ring = (__hal_ring_t *) vp->vpath->ringh;
vxge_assert(ring != NULL);
/* do POST */
rxdp->control_0 |= VXGE_HAL_RING_RXD_LIST_OWN_ADAPTER;
rxdp->control_1 |= VXGE_HAL_RING_RXD_LIST_TAIL_OWN_ADAPTER;
#if defined(VXGE_OS_DMA_REQUIRES_SYNC) && defined(VXGE_HAL_DMA_RXD_STREAMING)
priv = __hal_ring_rxd_priv(ring, rxdp);
vxge_os_dma_sync(ring->channel.pdev,
priv->dma_handle,
priv->dma_addr,
priv->dma_offset,
ring->rxd_size,
VXGE_OS_DMA_DIR_TODEVICE);
#endif
if (ring->stats->common_stats.usage_cnt > 0)
ring->stats->common_stats.usage_cnt--;
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
}
/*
* vxge_hal_ring_rxd_post - Post descriptor on the ring.
* @vpath_handle: virtual Path handle.
* @rxdh: Descriptor obtained via vxge_hal_ring_rxd_reserve().
*
* Post descriptor on the ring.
* Prior to posting the descriptor should be filled in accordance with
* Host/X3100 interface specification for a given service (LL, etc.).
*
*/
void
vxge_hal_ring_rxd_post(
vxge_hal_vpath_h vpath_handle,
vxge_hal_rxd_h rxdh)
{
vxge_hal_ring_rxd_1_t *rxdp = (vxge_hal_ring_rxd_1_t *) rxdh;
#if defined(VXGE_HAL_RX_MULTI_POST_IRQ)
unsigned long flags;
#endif
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
__hal_device_t *hldev;
__hal_ring_t *ring;
vxge_assert((vpath_handle != NULL) && (rxdh != NULL));
hldev = (__hal_device_t *) vp->vpath->hldev;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"vpath_handle = 0x"VXGE_OS_STXFMT", rxdh = 0x"VXGE_OS_STXFMT,
(ptr_t) vpath_handle, (ptr_t) rxdh);
ring = (__hal_ring_t *) vp->vpath->ringh;
vxge_assert(ring != NULL);
/* Based on Titan HW bugzilla # 3039, we need to reset the tcode */
rxdp->control_0 = 0;
#if defined(VXGE_DEBUG_ASSERT)
/* make sure device overwrites the (illegal) t_code on completion */
rxdp->control_0 |=
VXGE_HAL_RING_RXD_T_CODE(VXGE_HAL_RING_RXD_T_CODE_UNUSED);
#endif
rxdp->control_1 |= VXGE_HAL_RING_RXD_LIST_TAIL_OWN_ADAPTER;
rxdp->control_0 |= VXGE_HAL_RING_RXD_LIST_OWN_ADAPTER;
#if defined(VXGE_OS_DMA_REQUIRES_SYNC) && defined(VXGE_HAL_DMA_RXD_STREAMING)
{
__hal_ring_rxd_priv_t *rxdp_temp1;
rxdp_temp1 = VXGE_HAL_RING_HAL_PRIV(ring, rxdp);
vxge_os_dma_sync(ring->channel.pdev,
rxdp_temp1->dma_handle,
rxdp_temp1->dma_addr,
rxdp_temp1->dma_offset,
ring->rxd_size,
VXGE_OS_DMA_DIR_TODEVICE);
}
#endif
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_lock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_lock_irq(&ring->channel.post_lock, flags);
#endif
#if defined(VXGE_DEBUG_ASSERT) && defined(VXGE_HAL_RING_ENFORCE_ORDER)
if (TRUE) {
if (VXGE_HAL_RING_RXD_INDEX(rxdp) != 0) {
vxge_hal_rxd_h prev_rxdh;
__hal_ring_rxd_priv_t *rxdp_temp2;
rxdp_temp2 = VXGE_HAL_RING_HAL_PRIV(ring, rxdp);
prev_rxdh =
ring->channel.dtr_arr[VXGE_HAL_RING_RXD_INDEX(rxdp) - 1].dtr;
if (prev_rxdh != NULL &&
(rxdp_temp2->dma_offset & (~0xFFF)) != rxdp_temp2->dma_offset)
vxge_assert((char *) prev_rxdh + ring->rxd_size == rxdh);
}
}
#endif
__hal_channel_dtr_post(&ring->channel, VXGE_HAL_RING_RXD_INDEX(rxdh));
ring->db_byte_count +=
VXGE_HAL_RING_HAL_PRIV(ring, rxdp)->db_bytes;
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_unlock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_unlock_irq(&ring->channel.post_lock, flags);
#endif
if (ring->stats->common_stats.usage_cnt > 0)
ring->stats->common_stats.usage_cnt--;
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
}
/*
* vxge_hal_ring_rxd_post_post_wmb - Process rxd after post with memory barrier
* @vpath_handle: virtual Path handle.
* @rxdh: Descriptor handle.
*
* Processes rxd after post with memory barrier.
*/
void
vxge_hal_ring_rxd_post_post_wmb(
vxge_hal_vpath_h vpath_handle,
vxge_hal_rxd_h rxdh)
{
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
__hal_device_t *hldev;
vxge_assert((vpath_handle != NULL) && (rxdh != NULL));
hldev = (__hal_device_t *) vp->vpath->hldev;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"vpath_handle = 0x"VXGE_OS_STXFMT", rxdh = 0x"VXGE_OS_STXFMT,
(ptr_t) vpath_handle, (ptr_t) rxdh);
/* Do memory barrier before changing the ownership */
vxge_os_wmb();
vxge_hal_ring_rxd_post_post(vpath_handle, rxdh);
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
}
/*
* vxge_hal_ring_rxd_post_post_db - Post Doorbell after posting the rxd(s).
* @vpath_handle: virtual Path handle.
*
* Post Doorbell after posting the rxd(s).
*/
void
vxge_hal_ring_rxd_post_post_db(
vxge_hal_vpath_h vpath_handle)
{
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
__hal_device_t *hldev;
__hal_ring_t *ring;
vxge_assert(vpath_handle != NULL);
hldev = (__hal_device_t *) vp->vpath->hldev;
ring = (__hal_ring_t *) vp->vpath->ringh;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring("vpath_handle = 0x"VXGE_OS_STXFMT,
(ptr_t) vpath_handle);
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_lock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_lock_irq(&ring->channel.post_lock, flags);
#endif
if (ring->db_byte_count <= ring->rxd_mem_avail) {
__hal_rxd_db_post(vpath_handle, ring->db_byte_count);
ring->rxd_mem_avail -= ring->db_byte_count;
ring->db_byte_count = 0;
} else {
__hal_rxd_db_post(vpath_handle, ring->rxd_mem_avail);
ring->db_byte_count -= ring->rxd_mem_avail;
ring->rxd_mem_avail = 0;
}
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_unlock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_unlock_irq(&ring->channel.post_lock, flags);
#endif
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
}
/*
* vxge_hal_ring_is_next_rxd_completed - Check if the next rxd is completed
* @vpath_handle: Virtual Path handle.
*
* Checks if the _next_ completed descriptor is in host memory
*
* Returns: VXGE_HAL_OK - success.
* VXGE_HAL_INF_NO_MORE_COMPLETED_DESCRIPTORS - No completed descriptors
* are currently available for processing.
*/
vxge_hal_status_e
vxge_hal_ring_is_next_rxd_completed(
vxge_hal_vpath_h vpath_handle)
{
__hal_ring_t *ring;
vxge_hal_rxd_h rxdh;
vxge_hal_ring_rxd_1_t *rxdp; /* doesn't matter 1, 3 or 5... */
__hal_device_t *hldev;
vxge_hal_status_e status = VXGE_HAL_INF_NO_MORE_COMPLETED_DESCRIPTORS;
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
vxge_assert(vpath_handle != NULL);
hldev = (__hal_device_t *) vp->vpath->hldev;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring("vpath_handle = 0x"VXGE_OS_STXFMT,
(ptr_t) vpath_handle);
ring = (__hal_ring_t *) vp->vpath->ringh;
vxge_assert(ring != NULL);
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_lock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_lock_irq(&ring->channel.post_lock, flags);
#endif
__hal_channel_dtr_try_complete(&ring->channel, &rxdh);
rxdp = (vxge_hal_ring_rxd_1_t *) rxdh;
if (rxdp != NULL) {
/* check whether it is not the end */
if ((!(rxdp->control_0 & VXGE_HAL_RING_RXD_LIST_OWN_ADAPTER)) &&
(!(rxdp->control_1 &
VXGE_HAL_RING_RXD_LIST_TAIL_OWN_ADAPTER))) {
status = VXGE_HAL_OK;
}
}
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_unlock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_unlock_irq(&ring->channel.post_lock, flags);
#endif
vxge_hal_trace_log_ring("<== %s:%s:%d Result: %d",
__FILE__, __func__, __LINE__, status);
return (status);
}
/*
* vxge_hal_ring_rxd_next_completed - Get the _next_ completed descriptor.
* @channelh: Channel handle.
* @rxdh: Descriptor handle. Returned by HAL.
* @rxd_priv: Buffer to return a pointer to the per rxd space allocated
* @t_code: Transfer code, as per X3100 User Guide,
* Receive Descriptor Format. Returned by HAL.
*
* Retrieve the _next_ completed descriptor.
* HAL uses ring callback (*vxge_hal_ring_callback_f) to notifiy
* upper-layer driver (ULD) of new completed descriptors. After that
* the ULD can use vxge_hal_ring_rxd_next_completed to retrieve the rest
* completions (the very first completion is passed by HAL via
* vxge_hal_ring_callback_f).
*
* Implementation-wise, the upper-layer driver is free to call
* vxge_hal_ring_rxd_next_completed either immediately from inside the
* ring callback, or in a deferred fashion and separate (from HAL)
* context.
*
* Non-zero @t_code means failure to fill-in receive buffer(s)
* of the descriptor.
* For instance, parity error detected during the data transfer.
* In this case X3100 will complete the descriptor and indicate
* for the host that the received data is not to be used.
* For details please refer to X3100 User Guide.
*
* Returns: VXGE_HAL_OK - success.
* VXGE_HAL_INF_NO_MORE_COMPLETED_DESCRIPTORS - No completed descriptors
* are currently available for processing.
*
* See also: vxge_hal_ring_callback_f {},
* vxge_hal_fifo_rxd_next_completed(), vxge_hal_status_e {}.
*/
vxge_hal_status_e
vxge_hal_ring_rxd_next_completed(
vxge_hal_vpath_h vpath_handle,
vxge_hal_rxd_h *rxdh,
void **rxd_priv,
u8 *t_code)
{
__hal_ring_t *ring;
vxge_hal_ring_rxd_5_t *rxdp; /* doesn't matter 1, 3 or 5... */
#if defined(VXGE_OS_DMA_REQUIRES_SYNC) && defined(VXGE_HAL_DMA_RXD_STREAMING)
__hal_ring_rxd_priv_t *priv;
#endif
__hal_device_t *hldev;
vxge_hal_status_e status = VXGE_HAL_INF_NO_MORE_COMPLETED_DESCRIPTORS;
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
u64 own, control_0, control_1;
vxge_assert((vpath_handle != NULL) && (rxdh != NULL) &&
(rxd_priv != NULL) && (t_code != NULL));
hldev = (__hal_device_t *) vp->vpath->hldev;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"vpath_handle = 0x"VXGE_OS_STXFMT", rxdh = 0x"VXGE_OS_STXFMT", "
"rxd_priv = 0x"VXGE_OS_STXFMT", t_code = 0x"VXGE_OS_STXFMT,
(ptr_t) vpath_handle, (ptr_t) rxdh, (ptr_t) rxd_priv,
(ptr_t) t_code);
ring = (__hal_ring_t *) vp->vpath->ringh;
vxge_assert(ring != NULL);
*rxdh = 0;
*rxd_priv = NULL;
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_lock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_lock_irq(&ring->channel.post_lock, flags);
#endif
__hal_channel_dtr_try_complete(&ring->channel, rxdh);
rxdp = (vxge_hal_ring_rxd_5_t *)*rxdh;
if (rxdp != NULL) {
#if defined(VXGE_OS_DMA_REQUIRES_SYNC) && defined(VXGE_HAL_DMA_RXD_STREAMING)
/*
* Note: 24 bytes at most means:
* - Control_3 in case of 5-buffer mode
* - Control_1 and Control_2
*
* This is the only length needs to be invalidated
* type of channels.
*/
priv = __hal_ring_rxd_priv(ring, rxdp);
vxge_os_dma_sync(ring->channel.pdev,
priv->dma_handle,
priv->dma_addr,
priv->dma_offset,
24,
VXGE_OS_DMA_DIR_FROMDEVICE);
#endif
*t_code = (u8) VXGE_HAL_RING_RXD_T_CODE_GET(rxdp->control_0);
control_0 = rxdp->control_0;
control_1 = rxdp->control_1;
own = control_0 & VXGE_HAL_RING_RXD_LIST_OWN_ADAPTER;
/* check whether it is not the end */
if ((!own && !(control_1 & VXGE_HAL_RING_RXD_LIST_TAIL_OWN_ADAPTER)) ||
(*t_code == VXGE_HAL_RING_RXD_T_CODE_FRM_DROP)) {
#ifndef VXGE_HAL_IRQ_POLLING
if (++ring->cmpl_cnt > ring->indicate_max_pkts) {
/*
* reset it. since we don't want to return
* garbage to the ULD
*/
*rxdh = 0;
status = VXGE_HAL_COMPLETIONS_REMAIN;
} else {
#endif
__hal_channel_dtr_complete(&ring->channel);
*rxd_priv = VXGE_HAL_RING_ULD_PRIV(ring, rxdp);
ring->rxd_mem_avail +=
(VXGE_HAL_RING_HAL_PRIV(ring, rxdp))->db_bytes;
ring->stats->common_stats.usage_cnt++;
if (ring->stats->common_stats.usage_max <
ring->stats->common_stats.usage_cnt)
ring->stats->common_stats.usage_max =
ring->stats->common_stats.usage_cnt;
switch (ring->buffer_mode) {
case VXGE_HAL_RING_RXD_BUFFER_MODE_1:
ring->channel.poll_bytes +=
(u32) VXGE_HAL_RING_RXD_1_BUFFER0_SIZE_GET(
rxdp->control_1);
break;
case VXGE_HAL_RING_RXD_BUFFER_MODE_3:
ring->channel.poll_bytes +=
(u32) VXGE_HAL_RING_RXD_3_BUFFER0_SIZE_GET(
rxdp->control_1) +
(u32) VXGE_HAL_RING_RXD_3_BUFFER1_SIZE_GET(
rxdp->control_1) +
(u32) VXGE_HAL_RING_RXD_3_BUFFER2_SIZE_GET(
rxdp->control_1);
break;
case VXGE_HAL_RING_RXD_BUFFER_MODE_5:
ring->channel.poll_bytes +=
(u32) VXGE_HAL_RING_RXD_5_BUFFER0_SIZE_GET(
rxdp->control_1) +
(u32) VXGE_HAL_RING_RXD_5_BUFFER1_SIZE_GET(
rxdp->control_1) +
(u32) VXGE_HAL_RING_RXD_5_BUFFER2_SIZE_GET(
rxdp->control_1) +
(u32) VXGE_HAL_RING_RXD_5_BUFFER3_SIZE_GET(
rxdp->control_2) +
(u32) VXGE_HAL_RING_RXD_5_BUFFER4_SIZE_GET(
rxdp->control_2);
break;
}
status = VXGE_HAL_OK;
#ifndef VXGE_HAL_IRQ_POLLING
}
#endif
}
}
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_unlock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_unlock_irq(&ring->channel.post_lock, flags);
#endif
vxge_hal_trace_log_ring("<== %s:%s:%d Result: %d",
__FILE__, __func__, __LINE__, status);
return (status);
}
/*
* vxge_hal_ring_handle_tcode - Handle transfer code.
* @vpath_handle: Virtual Path handle.
* @rxdh: Descriptor handle.
* @t_code: One of the enumerated (and documented in the X3100 user guide)
* "transfer codes".
*
* Handle descriptor's transfer code. The latter comes with each completed
* descriptor.
*
* Returns: one of the vxge_hal_status_e {} enumerated types.
* VXGE_HAL_OK - for success.
* VXGE_HAL_ERR_CRITICAL - when encounters critical error.
*/
vxge_hal_status_e
vxge_hal_ring_handle_tcode(
vxge_hal_vpath_h vpath_handle,
vxge_hal_rxd_h rxdh,
u8 t_code)
{
__hal_device_t *hldev;
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
vxge_assert((vpath_handle != NULL) && (rxdh != NULL));
hldev = (__hal_device_t *) vp->vpath->hldev;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"vpath_handle = 0x"VXGE_OS_STXFMT", "
"rxdh = 0x"VXGE_OS_STXFMT", t_code = 0x%d",
(ptr_t) vpath_handle, (ptr_t) rxdh, t_code);
switch (t_code) {
case 0x0:
/* 0x0: Transfer ok. */
break;
case 0x1:
/*
* 0x1: Layer 3 checksum presentation
* configuration mismatch.
*/
break;
case 0x2:
/*
* 0x2: Layer 4 checksum presentation
* configuration mismatch.
*/
break;
case 0x3:
/*
* 0x3: Layer 3 and Layer 4 checksum
* presentation configuration mismatch.
*/
break;
case 0x4:
/* 0x4: Reserved. */
break;
case 0x5:
/*
* 0x5: Layer 3 error unparseable packet,
* such as unknown IPv6 header.
*/
break;
case 0x6:
/*
* 0x6: Layer 2 error frame integrity
* error, such as FCS or ECC).
*/
break;
case 0x7:
/*
* 0x7: Buffer size error the RxD buffer(s)
* were not appropriately sized and
* data loss occurred.
*/
break;
case 0x8:
/* 0x8: Internal ECC error RxD corrupted. */
__hal_device_handle_error(vp->vpath->hldev,
vp->vpath->vp_id, VXGE_HAL_EVENT_ECCERR);
break;
case 0x9:
/*
* 0x9: Benign overflow the contents of
* Segment1 exceeded the capacity of
* Buffer1 and the remainder was placed
* in Buffer2. Segment2 now starts in
* Buffer3. No data loss or errors occurred.
*/
break;
case 0xA:
/*
* 0xA: Buffer size 0 one of the RxDs
* assigned buffers has a size of 0 bytes.
*/
break;
case 0xB:
/* 0xB: Reserved. */
break;
case 0xC:
/*
* 0xC: Frame dropped either due to
* VPath Reset or because of a VPIN mismatch.
*/
break;
case 0xD:
/* 0xD: Reserved. */
break;
case 0xE:
/* 0xE: Reserved. */
break;
case 0xF:
/*
* 0xF: Multiple errors more than one
* transfer code condition occurred.
*/
break;
default:
vxge_hal_trace_log_ring("<== %s:%s:%d Result: %d",
__FILE__, __func__, __LINE__, VXGE_HAL_ERR_INVALID_TCODE);
return (VXGE_HAL_ERR_INVALID_TCODE);
}
vp->vpath->sw_stats->ring_stats.rxd_t_code_err_cnt[t_code]++;
vxge_hal_trace_log_ring("<== %s:%s:%d Result: %d",
__FILE__, __func__, __LINE__, VXGE_HAL_OK);
return (VXGE_HAL_OK);
}
/*
* vxge_hal_ring_rxd_private_get - Get ULD private per-descriptor data.
* @vpath_handle: Virtual Path handle.
* @rxdh: Descriptor handle.
*
* Returns: private ULD info associated with the descriptor.
* ULD requests per-descriptor space via vxge_hal_ring_attr.
*
*/
void *
vxge_hal_ring_rxd_private_get(
vxge_hal_vpath_h vpath_handle,
vxge_hal_rxd_h rxdh)
{
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
return (VXGE_HAL_RING_ULD_PRIV(
((__hal_ring_t *) vp->vpath->ringh), rxdh));
}
/*
* vxge_hal_ring_rxd_free - Free descriptor.
* @vpath_handle: Virtual Path handle.
* @rxdh: Descriptor handle.
*
* Free the reserved descriptor. This operation is "symmetrical" to
* vxge_hal_ring_rxd_reserve. The "free-ing" completes the descriptor's
* lifecycle.
*
* After free-ing (see vxge_hal_ring_rxd_free()) the descriptor again can
* be:
*
* - reserved (vxge_hal_ring_rxd_reserve);
*
* - posted (vxge_hal_ring_rxd_post);
*
* - completed (vxge_hal_ring_rxd_next_completed);
*
* - and recycled again (vxge_hal_ring_rxd_free).
*
* For alternative state transitions and more details please refer to
* the design doc.
*
*/
void
vxge_hal_ring_rxd_free(
vxge_hal_vpath_h vpath_handle,
vxge_hal_rxd_h rxdh)
{
#if defined(VXGE_HAL_RX_MULTI_POST_IRQ)
unsigned long flags;
#endif
__hal_ring_t *ring;
__hal_device_t *hldev;
__hal_vpath_handle_t *vp = (__hal_vpath_handle_t *) vpath_handle;
vxge_assert((vpath_handle != NULL) && (rxdh != NULL));
hldev = (__hal_device_t *) vp->vpath->hldev;
vxge_hal_trace_log_ring("==> %s:%s:%d",
__FILE__, __func__, __LINE__);
vxge_hal_trace_log_ring(
"vpath_handle = 0x"VXGE_OS_STXFMT", rxdh = 0x"VXGE_OS_STXFMT,
(ptr_t) vpath_handle, (ptr_t) rxdh);
ring = (__hal_ring_t *) vp->vpath->ringh;
vxge_assert(ring != NULL);
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_lock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_lock_irq(&ring->channel.post_lock, flags);
#endif
__hal_channel_dtr_free(&ring->channel, VXGE_HAL_RING_RXD_INDEX(rxdh));
#if defined(VXGE_OS_MEMORY_CHECK)
VXGE_HAL_RING_HAL_PRIV(ring, rxdh)->allocated = 0;
#endif
#if defined(VXGE_HAL_RX_MULTI_POST)
vxge_os_spin_unlock(&ring->channel.post_lock);
#elif defined(VXGE_HAL_RX_MULTI_POST_IRQ)
vxge_os_spin_unlock_irq(&ring->channel.post_lock, flags);
#endif
vxge_hal_trace_log_ring("<== %s:%s:%d Result: 0",
__FILE__, __func__, __LINE__);
}