There is no reason for the DPDK libraries to all have 'librte_' prefix on the directory names. This prefix makes the directory names longer and also makes it awkward to add features referring to individual libraries in the build - should the lib names be specified with or without the prefix. Therefore, we can just remove the library prefix and use the library's unique name as the directory name, i.e. 'eal' rather than 'librte_eal' Signed-off-by: Bruce Richardson <bruce.richardson@intel.com>
182 lines
3.4 KiB
C
182 lines
3.4 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(c) 2010-2014 Intel Corporation
|
|
*/
|
|
|
|
#ifndef _RTE_SPINLOCK_X86_64_H_
|
|
#define _RTE_SPINLOCK_X86_64_H_
|
|
|
|
#ifdef __cplusplus
|
|
extern "C" {
|
|
#endif
|
|
|
|
#include "generic/rte_spinlock.h"
|
|
#include "rte_rtm.h"
|
|
#include "rte_cpuflags.h"
|
|
#include "rte_branch_prediction.h"
|
|
#include "rte_common.h"
|
|
#include "rte_pause.h"
|
|
#include "rte_cycles.h"
|
|
|
|
#define RTE_RTM_MAX_RETRIES (20)
|
|
#define RTE_XABORT_LOCK_BUSY (0xff)
|
|
|
|
#ifndef RTE_FORCE_INTRINSICS
|
|
static inline void
|
|
rte_spinlock_lock(rte_spinlock_t *sl)
|
|
{
|
|
int lock_val = 1;
|
|
asm volatile (
|
|
"1:\n"
|
|
"xchg %[locked], %[lv]\n"
|
|
"test %[lv], %[lv]\n"
|
|
"jz 3f\n"
|
|
"2:\n"
|
|
"pause\n"
|
|
"cmpl $0, %[locked]\n"
|
|
"jnz 2b\n"
|
|
"jmp 1b\n"
|
|
"3:\n"
|
|
: [locked] "=m" (sl->locked), [lv] "=q" (lock_val)
|
|
: "[lv]" (lock_val)
|
|
: "memory");
|
|
}
|
|
|
|
static inline void
|
|
rte_spinlock_unlock (rte_spinlock_t *sl)
|
|
{
|
|
int unlock_val = 0;
|
|
asm volatile (
|
|
"xchg %[locked], %[ulv]\n"
|
|
: [locked] "=m" (sl->locked), [ulv] "=q" (unlock_val)
|
|
: "[ulv]" (unlock_val)
|
|
: "memory");
|
|
}
|
|
|
|
static inline int
|
|
rte_spinlock_trylock (rte_spinlock_t *sl)
|
|
{
|
|
int lockval = 1;
|
|
|
|
asm volatile (
|
|
"xchg %[locked], %[lockval]"
|
|
: [locked] "=m" (sl->locked), [lockval] "=q" (lockval)
|
|
: "[lockval]" (lockval)
|
|
: "memory");
|
|
|
|
return lockval == 0;
|
|
}
|
|
#endif
|
|
|
|
extern uint8_t rte_rtm_supported;
|
|
|
|
static inline int rte_tm_supported(void)
|
|
{
|
|
return rte_rtm_supported;
|
|
}
|
|
|
|
static inline int
|
|
rte_try_tm(volatile int *lock)
|
|
{
|
|
int i, retries;
|
|
|
|
if (!rte_rtm_supported)
|
|
return 0;
|
|
|
|
retries = RTE_RTM_MAX_RETRIES;
|
|
|
|
while (likely(retries--)) {
|
|
|
|
unsigned int status = rte_xbegin();
|
|
|
|
if (likely(RTE_XBEGIN_STARTED == status)) {
|
|
if (unlikely(*lock))
|
|
rte_xabort(RTE_XABORT_LOCK_BUSY);
|
|
else
|
|
return 1;
|
|
}
|
|
while (*lock)
|
|
rte_pause();
|
|
|
|
if ((status & RTE_XABORT_CONFLICT) ||
|
|
((status & RTE_XABORT_EXPLICIT) &&
|
|
(RTE_XABORT_CODE(status) == RTE_XABORT_LOCK_BUSY))) {
|
|
/* add a small delay before retrying, basing the
|
|
* delay on the number of times we've already tried,
|
|
* to give a back-off type of behaviour. We
|
|
* randomize trycount by taking bits from the tsc count
|
|
*/
|
|
int try_count = RTE_RTM_MAX_RETRIES - retries;
|
|
int pause_count = (rte_rdtsc() & 0x7) | 1;
|
|
pause_count <<= try_count;
|
|
for (i = 0; i < pause_count; i++)
|
|
rte_pause();
|
|
continue;
|
|
}
|
|
|
|
if ((status & RTE_XABORT_RETRY) == 0) /* do not retry */
|
|
break;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static inline void
|
|
rte_spinlock_lock_tm(rte_spinlock_t *sl)
|
|
{
|
|
if (likely(rte_try_tm(&sl->locked)))
|
|
return;
|
|
|
|
rte_spinlock_lock(sl); /* fall-back */
|
|
}
|
|
|
|
static inline int
|
|
rte_spinlock_trylock_tm(rte_spinlock_t *sl)
|
|
{
|
|
if (likely(rte_try_tm(&sl->locked)))
|
|
return 1;
|
|
|
|
return rte_spinlock_trylock(sl);
|
|
}
|
|
|
|
static inline void
|
|
rte_spinlock_unlock_tm(rte_spinlock_t *sl)
|
|
{
|
|
if (unlikely(sl->locked))
|
|
rte_spinlock_unlock(sl);
|
|
else
|
|
rte_xend();
|
|
}
|
|
|
|
static inline void
|
|
rte_spinlock_recursive_lock_tm(rte_spinlock_recursive_t *slr)
|
|
{
|
|
if (likely(rte_try_tm(&slr->sl.locked)))
|
|
return;
|
|
|
|
rte_spinlock_recursive_lock(slr); /* fall-back */
|
|
}
|
|
|
|
static inline void
|
|
rte_spinlock_recursive_unlock_tm(rte_spinlock_recursive_t *slr)
|
|
{
|
|
if (unlikely(slr->sl.locked))
|
|
rte_spinlock_recursive_unlock(slr);
|
|
else
|
|
rte_xend();
|
|
}
|
|
|
|
static inline int
|
|
rte_spinlock_recursive_trylock_tm(rte_spinlock_recursive_t *slr)
|
|
{
|
|
if (likely(rte_try_tm(&slr->sl.locked)))
|
|
return 1;
|
|
|
|
return rte_spinlock_recursive_trylock(slr);
|
|
}
|
|
|
|
|
|
#ifdef __cplusplus
|
|
}
|
|
#endif
|
|
|
|
#endif /* _RTE_SPINLOCK_X86_64_H_ */
|