20894340f2
This path updates ocf library to version 19.06 and also introduces all necessary changes required to integrate ocf bdev adapter with new version of ocf. Summary of changes introduces with new OCF in ocf bdev: - ocf_env.h increased limit for memory operations, changed behaviour of strncpy to less restrictive, both changes are required to run new OCF - ctx.c changed functions to new from new OCF - added new cache mode "write only" - added missed cache modes wa and wi to RPC scripts - rewritten spdk_rpc_bdev_ocf_get_stats function to use asynhronus ocf_mngt_cache_read_lock - used new asynhronus ocf_mngt_cache_flush instead of waiting for request - removed no longer valid filed - cfg->device.min_free_ram - changed expected result in metadata_probe_cb Signed-off-by: Marcin Dziegielewski <marcin.dziegielewski@intel.com> Change-Id: I83e4335e16600e4d22e6bb517931102de42d39e9 Reviewed-on: https://review.gerrithub.io/c/spdk/spdk/+/468132 Reviewed-by: Tomasz Zawadzki <tomasz.zawadzki@intel.com> Reviewed-by: Jim Harris <james.r.harris@intel.com> Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
779 lines
16 KiB
C
779 lines
16 KiB
C
/*-
|
|
* BSD LICENSE
|
|
*
|
|
* Copyright (c) Intel Corporation.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* * Neither the name of Intel Corporation nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
|
|
#ifndef __LIBOCF_ENV_H__
|
|
#define __LIBOCF_ENV_H__
|
|
|
|
#ifndef _GNU_SOURCE
|
|
#define _GNU_SOURCE
|
|
#endif
|
|
#ifndef __USE_GNU
|
|
#define __USE_GNU
|
|
#endif
|
|
|
|
#include <linux/limits.h>
|
|
#include <linux/stddef.h>
|
|
|
|
#include "spdk/stdinc.h"
|
|
#include "spdk/likely.h"
|
|
#include "spdk/env.h"
|
|
#include "spdk/util.h"
|
|
#include "spdk_internal/log.h"
|
|
|
|
#include "ocf_env_list.h"
|
|
#include "ocf/ocf_err.h"
|
|
|
|
typedef uint8_t u8;
|
|
typedef uint16_t u16;
|
|
typedef uint32_t u32;
|
|
typedef uint64_t u64;
|
|
|
|
typedef uint64_t sector_t;
|
|
|
|
#define __packed __attribute__((packed))
|
|
#define __aligned(x) __attribute__((aligned(x)))
|
|
|
|
/* linux sector 512-bytes */
|
|
#define ENV_SECTOR_SHIFT 9
|
|
#define ENV_SECTOR_SIZE (1<<ENV_SECTOR_SHIFT)
|
|
#define BYTES_TO_SECTOR(x) ((x) >> ENV_SECTOR_SHIFT)
|
|
|
|
/* *** MEMORY MANAGEMENT *** */
|
|
|
|
#define ENV_MEM_NORMAL 0
|
|
#define ENV_MEM_NOIO 0
|
|
#define ENV_MEM_ATOMIC 0
|
|
|
|
#define likely spdk_likely
|
|
#define unlikely spdk_unlikely
|
|
|
|
#define min(x, y) MIN(x, y)
|
|
#ifndef MIN
|
|
#define MIN(x, y) spdk_min(x, y)
|
|
#endif
|
|
|
|
#define ARRAY_SIZE(x) SPDK_COUNTOF(x)
|
|
|
|
/* LOGGING */
|
|
#define ENV_PRIu64 PRIu64
|
|
|
|
#define ENV_WARN(cond, fmt, args...) ({ \
|
|
if (spdk_unlikely((uintptr_t)(cond))) \
|
|
SPDK_NOTICELOG("WARNING" fmt, ##args); \
|
|
})
|
|
|
|
#define ENV_WARN_ON(cond) ({ \
|
|
if (spdk_unlikely((uintptr_t)(cond))) \
|
|
SPDK_NOTICELOG("WARNING\n"); \
|
|
})
|
|
|
|
#define ENV_BUG() ({ \
|
|
SPDK_ERRLOG("BUG\n"); \
|
|
assert(0); \
|
|
abort(); \
|
|
})
|
|
|
|
#define ENV_BUG_ON(cond) ({ \
|
|
if (spdk_unlikely((uintptr_t)(cond))) { \
|
|
SPDK_ERRLOG("BUG\n"); \
|
|
assert(0); \
|
|
abort(); \
|
|
} \
|
|
})
|
|
|
|
#define container_of(ptr, type, member) SPDK_CONTAINEROF(ptr, type, member)
|
|
|
|
static inline void *env_malloc(size_t size, int flags)
|
|
{
|
|
return spdk_malloc(size, 0, NULL, SPDK_ENV_LCORE_ID_ANY,
|
|
SPDK_MALLOC_DMA);
|
|
}
|
|
|
|
static inline void *env_zalloc(size_t size, int flags)
|
|
{
|
|
return spdk_zmalloc(size, 0, NULL, SPDK_ENV_LCORE_ID_ANY,
|
|
SPDK_MALLOC_DMA);
|
|
}
|
|
|
|
static inline void env_free(const void *ptr)
|
|
{
|
|
return spdk_free((void *)ptr);
|
|
}
|
|
|
|
static inline void *env_vmalloc(size_t size)
|
|
{
|
|
return spdk_malloc(size, 0, NULL, SPDK_ENV_LCORE_ID_ANY,
|
|
SPDK_MALLOC_DMA);
|
|
}
|
|
|
|
static inline void *env_vzalloc(size_t size)
|
|
{
|
|
/* TODO: raw_ram init can request huge amount of memory to store
|
|
* hashtable in it. need to ensure that allocation succedds */
|
|
return spdk_zmalloc(size, 0, NULL, SPDK_ENV_LCORE_ID_ANY,
|
|
SPDK_MALLOC_DMA);
|
|
}
|
|
|
|
static inline void *env_secure_alloc(size_t size)
|
|
{
|
|
return spdk_zmalloc(size, 0, NULL, SPDK_ENV_LCORE_ID_ANY,
|
|
SPDK_MALLOC_DMA);
|
|
}
|
|
|
|
static inline void env_secure_free(const void *ptr, size_t size)
|
|
{
|
|
return spdk_free((void *)ptr);
|
|
}
|
|
|
|
static inline void env_vfree(const void *ptr)
|
|
{
|
|
return spdk_free((void *)ptr);
|
|
}
|
|
|
|
static inline uint64_t env_get_free_memory(void)
|
|
{
|
|
/* TODO: do we need implementation for this function? */
|
|
return sysconf(_SC_PAGESIZE) * sysconf(_SC_AVPHYS_PAGES);
|
|
}
|
|
|
|
/* *** ALLOCATOR *** */
|
|
|
|
#define OCF_ALLOCATOR_NAME_MAX 128
|
|
|
|
typedef struct {
|
|
struct spdk_mempool *mempool;
|
|
size_t element_size;
|
|
} env_allocator;
|
|
|
|
env_allocator *env_allocator_create(uint32_t size, const char *name);
|
|
|
|
void env_allocator_destroy(env_allocator *allocator);
|
|
|
|
void *env_allocator_new(env_allocator *allocator);
|
|
|
|
void env_allocator_del(env_allocator *allocator, void *item);
|
|
|
|
uint32_t env_allocator_item_count(env_allocator *allocator);
|
|
|
|
/* *** MUTEX *** */
|
|
|
|
typedef struct {
|
|
pthread_mutex_t m;
|
|
} env_mutex;
|
|
|
|
static inline int env_mutex_init(env_mutex *mutex)
|
|
{
|
|
return !!pthread_mutex_init(&mutex->m, NULL);
|
|
}
|
|
|
|
static inline void env_mutex_lock(env_mutex *mutex)
|
|
{
|
|
ENV_BUG_ON(pthread_mutex_lock(&mutex->m));
|
|
}
|
|
|
|
static inline int env_mutex_lock_interruptible(env_mutex *mutex)
|
|
{
|
|
env_mutex_lock(mutex);
|
|
return 0;
|
|
}
|
|
|
|
static inline int env_mutex_trylock(env_mutex *mutex)
|
|
{
|
|
return pthread_mutex_trylock(&mutex->m) ? -OCF_ERR_NO_LOCK : 0;
|
|
}
|
|
|
|
static inline void env_mutex_unlock(env_mutex *mutex)
|
|
{
|
|
ENV_BUG_ON(pthread_mutex_unlock(&mutex->m));
|
|
}
|
|
|
|
static inline int env_mutex_is_locked(env_mutex *mutex)
|
|
{
|
|
if (env_mutex_trylock(mutex) == 0) {
|
|
env_mutex_unlock(mutex);
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/* *** RECURSIVE MUTEX *** */
|
|
|
|
typedef env_mutex env_rmutex;
|
|
|
|
static inline int env_rmutex_init(env_rmutex *rmutex)
|
|
{
|
|
return env_mutex_init(rmutex);
|
|
}
|
|
|
|
static inline void env_rmutex_lock(env_rmutex *rmutex)
|
|
{
|
|
env_mutex_lock(rmutex);
|
|
}
|
|
|
|
static inline int env_rmutex_lock_interruptible(env_rmutex *rmutex)
|
|
{
|
|
return env_mutex_lock_interruptible(rmutex);
|
|
}
|
|
|
|
static inline int env_rmutex_trylock(env_rmutex *rmutex)
|
|
{
|
|
return env_mutex_trylock(rmutex);
|
|
}
|
|
|
|
static inline void env_rmutex_unlock(env_rmutex *rmutex)
|
|
{
|
|
env_mutex_unlock(rmutex);
|
|
}
|
|
|
|
static inline int env_rmutex_is_locked(env_rmutex *rmutex)
|
|
{
|
|
return env_mutex_is_locked(rmutex);
|
|
}
|
|
|
|
/* *** RW SEMAPHORE *** */
|
|
typedef struct {
|
|
pthread_rwlock_t lock;
|
|
} env_rwsem;
|
|
|
|
static inline int env_rwsem_init(env_rwsem *s)
|
|
{
|
|
return !!pthread_rwlock_init(&s->lock, NULL);
|
|
}
|
|
|
|
static inline void env_rwsem_up_read(env_rwsem *s)
|
|
{
|
|
ENV_BUG_ON(pthread_rwlock_unlock(&s->lock));
|
|
}
|
|
|
|
static inline void env_rwsem_down_read(env_rwsem *s)
|
|
{
|
|
ENV_BUG_ON(pthread_rwlock_rdlock(&s->lock));
|
|
}
|
|
|
|
static inline int env_rwsem_down_read_trylock(env_rwsem *s)
|
|
{
|
|
return pthread_rwlock_tryrdlock(&s->lock) ? -OCF_ERR_NO_LOCK : 0;
|
|
}
|
|
|
|
static inline void env_rwsem_up_write(env_rwsem *s)
|
|
{
|
|
ENV_BUG_ON(pthread_rwlock_unlock(&s->lock));
|
|
}
|
|
|
|
static inline void env_rwsem_down_write(env_rwsem *s)
|
|
{
|
|
ENV_BUG_ON(pthread_rwlock_wrlock(&s->lock));
|
|
}
|
|
|
|
static inline int env_rwsem_down_write_trylock(env_rwsem *s)
|
|
{
|
|
return pthread_rwlock_trywrlock(&s->lock) ? -OCF_ERR_NO_LOCK : 0;
|
|
}
|
|
|
|
static inline int env_rwsem_is_locked(env_rwsem *s)
|
|
{
|
|
if (env_rwsem_down_read_trylock(s) == 0) {
|
|
env_rwsem_up_read(s);
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static inline int env_rwsem_down_read_interruptible(env_rwsem *s)
|
|
{
|
|
return pthread_rwlock_rdlock(&s->lock);
|
|
}
|
|
static inline int env_rwsem_down_write_interruptible(env_rwsem *s)
|
|
{
|
|
return pthread_rwlock_wrlock(&s->lock);
|
|
}
|
|
|
|
/* *** ATOMIC VARIABLES *** */
|
|
|
|
typedef int env_atomic;
|
|
|
|
typedef long env_atomic64;
|
|
|
|
#ifndef atomic_read
|
|
#define atomic_read(ptr) (*(__typeof__(*ptr) *volatile) (ptr))
|
|
#endif
|
|
|
|
#ifndef atomic_set
|
|
#define atomic_set(ptr, i) ((*(__typeof__(*ptr) *volatile) (ptr)) = (i))
|
|
#endif
|
|
|
|
#define atomic_inc(ptr) ((void) __sync_fetch_and_add(ptr, 1))
|
|
#define atomic_dec(ptr) ((void) __sync_fetch_and_add(ptr, -1))
|
|
#define atomic_add(ptr, n) ((void) __sync_fetch_and_add(ptr, n))
|
|
#define atomic_sub(ptr, n) ((void) __sync_fetch_and_sub(ptr, n))
|
|
|
|
#define atomic_cmpxchg __sync_val_compare_and_swap
|
|
|
|
static inline int env_atomic_read(const env_atomic *a)
|
|
{
|
|
return atomic_read(a);
|
|
}
|
|
|
|
static inline void env_atomic_set(env_atomic *a, int i)
|
|
{
|
|
atomic_set(a, i);
|
|
}
|
|
|
|
static inline void env_atomic_add(int i, env_atomic *a)
|
|
{
|
|
atomic_add(a, i);
|
|
}
|
|
|
|
static inline void env_atomic_sub(int i, env_atomic *a)
|
|
{
|
|
atomic_sub(a, i);
|
|
}
|
|
|
|
static inline bool env_atomic_sub_and_test(int i, env_atomic *a)
|
|
{
|
|
return __sync_sub_and_fetch(a, i) == 0;
|
|
}
|
|
|
|
static inline void env_atomic_inc(env_atomic *a)
|
|
{
|
|
atomic_inc(a);
|
|
}
|
|
|
|
static inline void env_atomic_dec(env_atomic *a)
|
|
{
|
|
atomic_dec(a);
|
|
}
|
|
|
|
static inline bool env_atomic_dec_and_test(env_atomic *a)
|
|
{
|
|
return __sync_sub_and_fetch(a, 1) == 0;
|
|
}
|
|
|
|
static inline bool env_atomic_inc_and_test(env_atomic *a)
|
|
{
|
|
return __sync_add_and_fetch(a, 1) == 0;
|
|
}
|
|
|
|
static inline int env_atomic_add_return(int i, env_atomic *a)
|
|
{
|
|
return __sync_add_and_fetch(a, i);
|
|
}
|
|
|
|
static inline int env_atomic_sub_return(int i, env_atomic *a)
|
|
{
|
|
return __sync_sub_and_fetch(a, i);
|
|
}
|
|
|
|
static inline int env_atomic_inc_return(env_atomic *a)
|
|
{
|
|
return env_atomic_add_return(1, a);
|
|
}
|
|
|
|
static inline int env_atomic_dec_return(env_atomic *a)
|
|
{
|
|
return env_atomic_sub_return(1, a);
|
|
}
|
|
|
|
static inline int env_atomic_cmpxchg(env_atomic *a, int old, int new_value)
|
|
{
|
|
return atomic_cmpxchg(a, old, new_value);
|
|
}
|
|
|
|
static inline int env_atomic_add_unless(env_atomic *a, int i, int u)
|
|
{
|
|
int c, old;
|
|
c = env_atomic_read(a);
|
|
for (;;) {
|
|
if (spdk_unlikely(c == (u))) {
|
|
break;
|
|
}
|
|
old = env_atomic_cmpxchg((a), c, c + (i));
|
|
if (spdk_likely(old == c)) {
|
|
break;
|
|
}
|
|
c = old;
|
|
}
|
|
return c != (u);
|
|
}
|
|
|
|
static inline long env_atomic64_read(const env_atomic64 *a)
|
|
{
|
|
return atomic_read(a);
|
|
}
|
|
|
|
static inline void env_atomic64_set(env_atomic64 *a, long i)
|
|
{
|
|
atomic_set(a, i);
|
|
}
|
|
|
|
static inline void env_atomic64_add(long i, env_atomic64 *a)
|
|
{
|
|
atomic_add(a, i);
|
|
}
|
|
|
|
static inline void env_atomic64_sub(long i, env_atomic64 *a)
|
|
{
|
|
atomic_sub(a, i);
|
|
}
|
|
|
|
static inline void env_atomic64_inc(env_atomic64 *a)
|
|
{
|
|
atomic_inc(a);
|
|
}
|
|
|
|
static inline void env_atomic64_dec(env_atomic64 *a)
|
|
{
|
|
atomic_dec(a);
|
|
}
|
|
|
|
static inline int env_atomic64_add_return(int i, env_atomic *a)
|
|
{
|
|
return __sync_add_and_fetch(a, i);
|
|
}
|
|
|
|
static inline int env_atomic64_sub_return(int i, env_atomic *a)
|
|
{
|
|
return __sync_sub_and_fetch(a, i);
|
|
}
|
|
|
|
static inline int env_atomic64_inc_return(env_atomic *a)
|
|
{
|
|
return env_atomic64_add_return(1, a);
|
|
}
|
|
|
|
static inline int env_atomic64_dec_return(env_atomic *a)
|
|
{
|
|
return env_atomic_sub_return(1, a);
|
|
}
|
|
|
|
static inline long env_atomic64_cmpxchg(env_atomic64 *a, long old, long new)
|
|
{
|
|
return atomic_cmpxchg(a, old, new);
|
|
}
|
|
|
|
/* *** COMPLETION *** */
|
|
struct completion {
|
|
env_atomic atom;
|
|
};
|
|
|
|
typedef struct completion env_completion;
|
|
|
|
void env_completion_init(env_completion *completion);
|
|
void env_completion_wait(env_completion *completion);
|
|
void env_completion_complete(env_completion *completion);
|
|
|
|
/* *** SPIN LOCKS *** */
|
|
|
|
typedef env_mutex env_spinlock;
|
|
|
|
static inline void env_spinlock_init(env_spinlock *l)
|
|
{
|
|
env_mutex_init(l);
|
|
}
|
|
|
|
static inline void env_spinlock_lock(env_spinlock *l)
|
|
{
|
|
env_mutex_lock(l);
|
|
}
|
|
|
|
static inline void env_spinlock_unlock(env_spinlock *l)
|
|
{
|
|
env_mutex_unlock(l);
|
|
}
|
|
|
|
static inline void env_spinlock_lock_irq(env_spinlock *l)
|
|
{
|
|
env_spinlock_lock(l);
|
|
}
|
|
|
|
static inline void env_spinlock_unlock_irq(env_spinlock *l)
|
|
{
|
|
env_spinlock_unlock(l);
|
|
}
|
|
|
|
static inline void env_spinlock_lock_irqsave(env_spinlock *l, int flags)
|
|
{
|
|
env_spinlock_lock(l);
|
|
(void)flags;
|
|
}
|
|
|
|
static inline void env_spinlock_unlock_irqrestore(env_spinlock *l, int flags)
|
|
{
|
|
env_spinlock_unlock(l);
|
|
(void)flags;
|
|
}
|
|
|
|
/* *** RW LOCKS *** */
|
|
|
|
typedef env_rwsem env_rwlock;
|
|
|
|
static inline void env_rwlock_init(env_rwlock *l)
|
|
{
|
|
env_rwsem_init(l);
|
|
}
|
|
|
|
static inline void env_rwlock_read_lock(env_rwlock *l)
|
|
{
|
|
env_rwsem_down_read(l);
|
|
}
|
|
|
|
static inline void env_rwlock_read_unlock(env_rwlock *l)
|
|
{
|
|
env_rwsem_up_read(l);
|
|
}
|
|
|
|
static inline void env_rwlock_write_lock(env_rwlock *l)
|
|
{
|
|
env_rwsem_down_write(l);
|
|
}
|
|
|
|
static inline void env_rwlock_write_unlock(env_rwlock *l)
|
|
{
|
|
env_rwsem_up_write(l);
|
|
}
|
|
|
|
static inline void env_bit_set(int nr, volatile void *addr)
|
|
{
|
|
char *byte = (char *)addr + (nr >> 3);
|
|
char mask = 1 << (nr & 7);
|
|
|
|
__sync_or_and_fetch(byte, mask);
|
|
}
|
|
|
|
static inline void env_bit_clear(int nr, volatile void *addr)
|
|
{
|
|
char *byte = (char *)addr + (nr >> 3);
|
|
char mask = 1 << (nr & 7);
|
|
|
|
mask = ~mask;
|
|
__sync_and_and_fetch(byte, mask);
|
|
}
|
|
|
|
static inline bool env_bit_test(int nr, const volatile unsigned long *addr)
|
|
{
|
|
const char *byte = (char *)addr + (nr >> 3);
|
|
char mask = 1 << (nr & 7);
|
|
|
|
return !!(*byte & mask);
|
|
}
|
|
|
|
/* *** WAITQUEUE *** */
|
|
|
|
typedef struct {
|
|
sem_t sem;
|
|
} env_waitqueue;
|
|
|
|
static inline void env_waitqueue_init(env_waitqueue *w)
|
|
{
|
|
sem_init(&w->sem, 0, 0);
|
|
}
|
|
|
|
static inline void env_waitqueue_wake_up(env_waitqueue *w)
|
|
{
|
|
sem_post(&w->sem);
|
|
}
|
|
|
|
#define env_waitqueue_wait(w, condition) \
|
|
({ \
|
|
int __ret = 0; \
|
|
if (!(condition)) \
|
|
sem_wait(&w.sem); \
|
|
__ret = __ret; \
|
|
})
|
|
|
|
/* *** SCHEDULING *** */
|
|
|
|
/* CAS does not need this while in user-space */
|
|
static inline void env_schedule(void)
|
|
{
|
|
}
|
|
|
|
#define env_cond_resched env_schedule
|
|
|
|
static inline int env_in_interrupt(void)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline uint64_t env_get_tick_count(void)
|
|
{
|
|
return spdk_get_ticks();
|
|
}
|
|
|
|
static inline uint64_t env_ticks_to_secs(uint64_t j)
|
|
{
|
|
return j / spdk_get_ticks_hz();
|
|
}
|
|
|
|
static inline uint64_t env_ticks_to_msecs(uint64_t j)
|
|
{
|
|
return env_ticks_to_secs(j) * 1000;
|
|
}
|
|
|
|
static inline uint64_t env_ticks_to_nsecs(uint64_t j)
|
|
{
|
|
return env_ticks_to_secs(j) * 1000 * 1000;
|
|
}
|
|
|
|
static inline uint64_t env_ticks_to_usecs(uint64_t j)
|
|
{
|
|
return env_ticks_to_secs(j) * 1000 * 1000 * 1000;
|
|
}
|
|
|
|
static inline uint64_t env_secs_to_ticks(uint64_t j)
|
|
{
|
|
return j * spdk_get_ticks_hz();
|
|
}
|
|
|
|
/* *** STRING OPERATIONS *** */
|
|
|
|
/* 512 KB is sufficient amount of memory for OCF operations */
|
|
#define ENV_MAX_MEM (512 * 1024)
|
|
|
|
static inline int env_memset(void *dest, size_t len, uint8_t value)
|
|
{
|
|
if (dest == NULL || len == 0) {
|
|
return 1;
|
|
}
|
|
|
|
memset(dest, value, len);
|
|
return 0;
|
|
}
|
|
|
|
static inline int env_memcpy(void *dest, size_t dmax, const void *src, size_t len)
|
|
{
|
|
if (dest == NULL || src == NULL) {
|
|
return 1;
|
|
}
|
|
if (dmax == 0 || dmax > ENV_MAX_MEM) {
|
|
return 1;
|
|
}
|
|
if (len == 0 || len > dmax) {
|
|
return 1;
|
|
}
|
|
|
|
memcpy(dest, src, len);
|
|
return 0;
|
|
}
|
|
|
|
static inline int env_memcmp(const void *aptr, size_t dmax, const void *bptr, size_t len,
|
|
int *diff)
|
|
{
|
|
if (diff == NULL || aptr == NULL || bptr == NULL) {
|
|
return 1;
|
|
}
|
|
if (dmax == 0 || dmax > ENV_MAX_MEM) {
|
|
return 1;
|
|
}
|
|
if (len == 0 || len > dmax) {
|
|
return 1;
|
|
}
|
|
|
|
*diff = memcmp(aptr, bptr, len);
|
|
return 0;
|
|
}
|
|
|
|
/* 4096 is sufficient max length for any OCF operation on string */
|
|
#define ENV_MAX_STR (4 * 1024)
|
|
|
|
static inline size_t env_strnlen(const char *src, size_t dmax)
|
|
{
|
|
return strnlen(src, dmax);
|
|
}
|
|
|
|
static inline int env_strncpy(char *dest, size_t dmax, const char *src, size_t len)
|
|
{
|
|
if (dest == NULL || src == NULL) {
|
|
return 1;
|
|
}
|
|
if (dmax == 0 || dmax > ENV_MAX_STR) {
|
|
return 1;
|
|
}
|
|
if (len == 0) {
|
|
return 1;
|
|
}
|
|
/* Just copy as many characters as we can instead of return failure */
|
|
len = min(len, dmax);
|
|
|
|
strncpy(dest, src, len);
|
|
return 0;
|
|
}
|
|
|
|
#define env_strncmp strncmp
|
|
|
|
static inline char *env_strdup(const char *src, int flags)
|
|
{
|
|
int len;
|
|
char *ret;
|
|
|
|
if (src == NULL) {
|
|
return NULL;
|
|
}
|
|
|
|
len = env_strnlen(src, ENV_MAX_STR) + 1;
|
|
ret = env_malloc(len, flags);
|
|
|
|
if (env_strncpy(ret, ENV_MAX_STR, src, len)) {
|
|
return NULL;
|
|
} else {
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
/* *** SORTING *** */
|
|
|
|
static inline void env_sort(void *base, size_t num, size_t size,
|
|
int (*cmp_fn)(const void *, const void *),
|
|
void (*swap_fn)(void *, void *, int size))
|
|
{
|
|
qsort(base, num, size, cmp_fn);
|
|
}
|
|
|
|
static inline void env_msleep(uint64_t n)
|
|
{
|
|
usleep(n * 1000);
|
|
}
|
|
|
|
static inline void env_touch_softlockup_wd(void)
|
|
{
|
|
}
|
|
|
|
/* *** CRC *** */
|
|
|
|
uint32_t env_crc32(uint32_t crc, uint8_t const *data, size_t len);
|
|
|
|
#endif /* __OCF_ENV_H__ */
|