dbba7c9efb
The string copy api rte_strscpy() did not set rte_errno during failures, instead it just returned negative error number. Set rte_errrno if the destination buffer is too small. Signed-off-by: Thomas Monjalon <thomas@monjalon.net> Signed-off-by: Xueming Li <xuemingl@nvidia.com> Reviewed-by: David Marchand <david.marchand@redhat.com>
533 lines
12 KiB
C
533 lines
12 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(C) 2020 Marvell International Ltd.
|
|
*/
|
|
|
|
#include <fnmatch.h>
|
|
#include <inttypes.h>
|
|
#include <sys/queue.h>
|
|
#include <regex.h>
|
|
|
|
#include <rte_common.h>
|
|
#include <rte_errno.h>
|
|
#include <rte_lcore.h>
|
|
#include <rte_per_lcore.h>
|
|
#include <rte_string_fns.h>
|
|
|
|
#include "eal_trace.h"
|
|
|
|
RTE_DEFINE_PER_LCORE(volatile int, trace_point_sz);
|
|
RTE_DEFINE_PER_LCORE(void *, trace_mem);
|
|
static RTE_DEFINE_PER_LCORE(char *, ctf_field);
|
|
|
|
static struct trace_point_head tp_list = STAILQ_HEAD_INITIALIZER(tp_list);
|
|
static struct trace trace = { .args = STAILQ_HEAD_INITIALIZER(trace.args), };
|
|
|
|
struct trace *
|
|
trace_obj_get(void)
|
|
{
|
|
return &trace;
|
|
}
|
|
|
|
struct trace_point_head *
|
|
trace_list_head_get(void)
|
|
{
|
|
return &tp_list;
|
|
}
|
|
|
|
int
|
|
eal_trace_init(void)
|
|
{
|
|
struct trace_arg *arg;
|
|
|
|
/* Trace memory should start with 8B aligned for natural alignment */
|
|
RTE_BUILD_BUG_ON((offsetof(struct __rte_trace_header, mem) % 8) != 0);
|
|
|
|
/* One of the trace point registration failed */
|
|
if (trace.register_errno) {
|
|
rte_errno = trace.register_errno;
|
|
goto fail;
|
|
}
|
|
|
|
if (!STAILQ_EMPTY(&trace.args))
|
|
trace.status = true;
|
|
|
|
if (!rte_trace_is_enabled())
|
|
return 0;
|
|
|
|
rte_spinlock_init(&trace.lock);
|
|
|
|
/* Is duplicate trace name registered */
|
|
if (trace_has_duplicate_entry())
|
|
goto fail;
|
|
|
|
/* Generate UUID ver 4 with total size of events and number of
|
|
* events
|
|
*/
|
|
trace_uuid_generate();
|
|
|
|
/* Apply buffer size configuration for trace output */
|
|
trace_bufsz_args_apply();
|
|
|
|
/* Generate CTF TDSL metadata */
|
|
if (trace_metadata_create() < 0)
|
|
goto fail;
|
|
|
|
/* Create trace directory */
|
|
if (trace_mkdir())
|
|
goto free_meta;
|
|
|
|
/* Save current epoch timestamp for future use */
|
|
if (trace_epoch_time_save() < 0)
|
|
goto fail;
|
|
|
|
/* Apply global configurations */
|
|
STAILQ_FOREACH(arg, &trace.args, next)
|
|
trace_args_apply(arg->val);
|
|
|
|
rte_trace_mode_set(trace.mode);
|
|
|
|
return 0;
|
|
|
|
free_meta:
|
|
trace_metadata_destroy();
|
|
fail:
|
|
trace_err("failed to initialize trace [%s]", rte_strerror(rte_errno));
|
|
return -rte_errno;
|
|
}
|
|
|
|
void
|
|
eal_trace_fini(void)
|
|
{
|
|
if (!rte_trace_is_enabled())
|
|
return;
|
|
trace_mem_free();
|
|
trace_metadata_destroy();
|
|
eal_trace_args_free();
|
|
}
|
|
|
|
bool
|
|
rte_trace_is_enabled(void)
|
|
{
|
|
return trace.status;
|
|
}
|
|
|
|
static void
|
|
trace_mode_set(rte_trace_point_t *trace, enum rte_trace_mode mode)
|
|
{
|
|
if (mode == RTE_TRACE_MODE_OVERWRITE)
|
|
__atomic_and_fetch(trace, ~__RTE_TRACE_FIELD_ENABLE_DISCARD,
|
|
__ATOMIC_RELEASE);
|
|
else
|
|
__atomic_or_fetch(trace, __RTE_TRACE_FIELD_ENABLE_DISCARD,
|
|
__ATOMIC_RELEASE);
|
|
}
|
|
|
|
void
|
|
rte_trace_mode_set(enum rte_trace_mode mode)
|
|
{
|
|
struct trace_point *tp;
|
|
|
|
if (!rte_trace_is_enabled())
|
|
return;
|
|
|
|
STAILQ_FOREACH(tp, &tp_list, next)
|
|
trace_mode_set(tp->handle, mode);
|
|
|
|
trace.mode = mode;
|
|
}
|
|
|
|
enum
|
|
rte_trace_mode rte_trace_mode_get(void)
|
|
{
|
|
return trace.mode;
|
|
}
|
|
|
|
static bool
|
|
trace_point_is_invalid(rte_trace_point_t *t)
|
|
{
|
|
return (t == NULL) || (trace_id_get(t) >= trace.nb_trace_points);
|
|
}
|
|
|
|
bool
|
|
rte_trace_point_is_enabled(rte_trace_point_t *trace)
|
|
{
|
|
uint64_t val;
|
|
|
|
if (trace_point_is_invalid(trace))
|
|
return false;
|
|
|
|
val = __atomic_load_n(trace, __ATOMIC_ACQUIRE);
|
|
return (val & __RTE_TRACE_FIELD_ENABLE_MASK) != 0;
|
|
}
|
|
|
|
int
|
|
rte_trace_point_enable(rte_trace_point_t *trace)
|
|
{
|
|
if (trace_point_is_invalid(trace))
|
|
return -ERANGE;
|
|
|
|
__atomic_or_fetch(trace, __RTE_TRACE_FIELD_ENABLE_MASK,
|
|
__ATOMIC_RELEASE);
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
rte_trace_point_disable(rte_trace_point_t *trace)
|
|
{
|
|
if (trace_point_is_invalid(trace))
|
|
return -ERANGE;
|
|
|
|
__atomic_and_fetch(trace, ~__RTE_TRACE_FIELD_ENABLE_MASK,
|
|
__ATOMIC_RELEASE);
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
rte_trace_pattern(const char *pattern, bool enable)
|
|
{
|
|
struct trace_point *tp;
|
|
int rc = 0, found = 0;
|
|
|
|
STAILQ_FOREACH(tp, &tp_list, next) {
|
|
if (fnmatch(pattern, tp->name, 0) == 0) {
|
|
if (enable)
|
|
rc = rte_trace_point_enable(tp->handle);
|
|
else
|
|
rc = rte_trace_point_disable(tp->handle);
|
|
found = 1;
|
|
}
|
|
if (rc < 0)
|
|
return rc;
|
|
}
|
|
|
|
return rc | found;
|
|
}
|
|
|
|
int
|
|
rte_trace_regexp(const char *regex, bool enable)
|
|
{
|
|
struct trace_point *tp;
|
|
int rc = 0, found = 0;
|
|
regex_t r;
|
|
|
|
if (regcomp(&r, regex, 0) != 0)
|
|
return -EINVAL;
|
|
|
|
STAILQ_FOREACH(tp, &tp_list, next) {
|
|
if (regexec(&r, tp->name, 0, NULL, 0) == 0) {
|
|
if (enable)
|
|
rc = rte_trace_point_enable(tp->handle);
|
|
else
|
|
rc = rte_trace_point_disable(tp->handle);
|
|
found = 1;
|
|
}
|
|
if (rc < 0)
|
|
return rc;
|
|
}
|
|
regfree(&r);
|
|
|
|
return rc | found;
|
|
}
|
|
|
|
rte_trace_point_t *
|
|
rte_trace_point_lookup(const char *name)
|
|
{
|
|
struct trace_point *tp;
|
|
|
|
if (name == NULL)
|
|
return NULL;
|
|
|
|
STAILQ_FOREACH(tp, &tp_list, next)
|
|
if (strncmp(tp->name, name, TRACE_POINT_NAME_SIZE) == 0)
|
|
return tp->handle;
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void
|
|
trace_point_dump(FILE *f, struct trace_point *tp)
|
|
{
|
|
rte_trace_point_t *handle = tp->handle;
|
|
|
|
fprintf(f, "\tid %d, %s, size is %d, %s\n",
|
|
trace_id_get(handle), tp->name,
|
|
(uint16_t)(*handle & __RTE_TRACE_FIELD_SIZE_MASK),
|
|
rte_trace_point_is_enabled(handle) ? "enabled" : "disabled");
|
|
}
|
|
|
|
static void
|
|
trace_lcore_mem_dump(FILE *f)
|
|
{
|
|
struct trace *trace = trace_obj_get();
|
|
struct __rte_trace_header *header;
|
|
uint32_t count;
|
|
|
|
if (trace->nb_trace_mem_list == 0)
|
|
return;
|
|
|
|
rte_spinlock_lock(&trace->lock);
|
|
fprintf(f, "nb_trace_mem_list = %d\n", trace->nb_trace_mem_list);
|
|
fprintf(f, "\nTrace mem info\n--------------\n");
|
|
for (count = 0; count < trace->nb_trace_mem_list; count++) {
|
|
header = trace->lcore_meta[count].mem;
|
|
fprintf(f, "\tid %d, mem=%p, area=%s, lcore_id=%d, name=%s\n",
|
|
count, header,
|
|
trace_area_to_string(trace->lcore_meta[count].area),
|
|
header->stream_header.lcore_id,
|
|
header->stream_header.thread_name);
|
|
}
|
|
rte_spinlock_unlock(&trace->lock);
|
|
}
|
|
|
|
void
|
|
rte_trace_dump(FILE *f)
|
|
{
|
|
struct trace_point_head *tp_list = trace_list_head_get();
|
|
struct trace *trace = trace_obj_get();
|
|
struct trace_point *tp;
|
|
|
|
fprintf(f, "\nGlobal info\n-----------\n");
|
|
fprintf(f, "status = %s\n",
|
|
rte_trace_is_enabled() ? "enabled" : "disabled");
|
|
fprintf(f, "mode = %s\n",
|
|
trace_mode_to_string(rte_trace_mode_get()));
|
|
fprintf(f, "dir = %s\n", trace->dir);
|
|
fprintf(f, "buffer len = %d\n", trace->buff_len);
|
|
fprintf(f, "number of trace points = %d\n", trace->nb_trace_points);
|
|
|
|
trace_lcore_mem_dump(f);
|
|
fprintf(f, "\nTrace point info\n----------------\n");
|
|
STAILQ_FOREACH(tp, tp_list, next)
|
|
trace_point_dump(f, tp);
|
|
}
|
|
|
|
void
|
|
__rte_trace_mem_per_thread_alloc(void)
|
|
{
|
|
struct trace *trace = trace_obj_get();
|
|
struct __rte_trace_header *header;
|
|
uint32_t count;
|
|
|
|
if (!rte_trace_is_enabled())
|
|
return;
|
|
|
|
if (RTE_PER_LCORE(trace_mem))
|
|
return;
|
|
|
|
rte_spinlock_lock(&trace->lock);
|
|
|
|
count = trace->nb_trace_mem_list;
|
|
|
|
/* Allocate room for storing the thread trace mem meta */
|
|
trace->lcore_meta = realloc(trace->lcore_meta,
|
|
sizeof(trace->lcore_meta[0]) * (count + 1));
|
|
|
|
/* Provide dummy space for fast path to consume */
|
|
if (trace->lcore_meta == NULL) {
|
|
trace_crit("trace mem meta memory realloc failed");
|
|
header = NULL;
|
|
goto fail;
|
|
}
|
|
|
|
/* First attempt from huge page */
|
|
header = eal_malloc_no_trace(NULL, trace_mem_sz(trace->buff_len), 8);
|
|
if (header) {
|
|
trace->lcore_meta[count].area = TRACE_AREA_HUGEPAGE;
|
|
goto found;
|
|
}
|
|
|
|
/* Second attempt from heap */
|
|
header = malloc(trace_mem_sz(trace->buff_len));
|
|
if (header == NULL) {
|
|
trace_crit("trace mem malloc attempt failed");
|
|
header = NULL;
|
|
goto fail;
|
|
|
|
}
|
|
|
|
/* Second attempt from heap is success */
|
|
trace->lcore_meta[count].area = TRACE_AREA_HEAP;
|
|
|
|
/* Initialize the trace header */
|
|
found:
|
|
header->offset = 0;
|
|
header->len = trace->buff_len;
|
|
header->stream_header.magic = TRACE_CTF_MAGIC;
|
|
rte_uuid_copy(header->stream_header.uuid, trace->uuid);
|
|
header->stream_header.lcore_id = rte_lcore_id();
|
|
|
|
/* Store the thread name */
|
|
char *name = header->stream_header.thread_name;
|
|
memset(name, 0, __RTE_TRACE_EMIT_STRING_LEN_MAX);
|
|
rte_thread_getname(pthread_self(), name,
|
|
__RTE_TRACE_EMIT_STRING_LEN_MAX);
|
|
|
|
trace->lcore_meta[count].mem = header;
|
|
trace->nb_trace_mem_list++;
|
|
fail:
|
|
RTE_PER_LCORE(trace_mem) = header;
|
|
rte_spinlock_unlock(&trace->lock);
|
|
}
|
|
|
|
static void
|
|
trace_mem_per_thread_free_unlocked(struct thread_mem_meta *meta)
|
|
{
|
|
if (meta->area == TRACE_AREA_HUGEPAGE)
|
|
eal_free_no_trace(meta->mem);
|
|
else if (meta->area == TRACE_AREA_HEAP)
|
|
free(meta->mem);
|
|
}
|
|
|
|
void
|
|
trace_mem_per_thread_free(void)
|
|
{
|
|
struct trace *trace = trace_obj_get();
|
|
struct __rte_trace_header *header;
|
|
uint32_t count;
|
|
|
|
header = RTE_PER_LCORE(trace_mem);
|
|
if (header == NULL)
|
|
return;
|
|
|
|
rte_spinlock_lock(&trace->lock);
|
|
for (count = 0; count < trace->nb_trace_mem_list; count++) {
|
|
if (trace->lcore_meta[count].mem == header)
|
|
break;
|
|
}
|
|
if (count != trace->nb_trace_mem_list) {
|
|
struct thread_mem_meta *meta = &trace->lcore_meta[count];
|
|
|
|
trace_mem_per_thread_free_unlocked(meta);
|
|
if (count != trace->nb_trace_mem_list - 1) {
|
|
memmove(meta, meta + 1,
|
|
sizeof(*meta) *
|
|
(trace->nb_trace_mem_list - count - 1));
|
|
}
|
|
trace->nb_trace_mem_list--;
|
|
}
|
|
rte_spinlock_unlock(&trace->lock);
|
|
}
|
|
|
|
void
|
|
trace_mem_free(void)
|
|
{
|
|
struct trace *trace = trace_obj_get();
|
|
uint32_t count;
|
|
|
|
if (!rte_trace_is_enabled())
|
|
return;
|
|
|
|
rte_spinlock_lock(&trace->lock);
|
|
for (count = 0; count < trace->nb_trace_mem_list; count++) {
|
|
trace_mem_per_thread_free_unlocked(&trace->lcore_meta[count]);
|
|
}
|
|
trace->nb_trace_mem_list = 0;
|
|
rte_spinlock_unlock(&trace->lock);
|
|
}
|
|
|
|
void
|
|
__rte_trace_point_emit_field(size_t sz, const char *in, const char *datatype)
|
|
{
|
|
char *field;
|
|
char *fixup;
|
|
int rc;
|
|
|
|
fixup = trace_metadata_fixup_field(in);
|
|
if (fixup != NULL)
|
|
in = fixup;
|
|
rc = asprintf(&field, "%s %s %s;\n",
|
|
RTE_PER_LCORE(ctf_field) != NULL ?
|
|
RTE_PER_LCORE(ctf_field) : "",
|
|
datatype, in);
|
|
free(RTE_PER_LCORE(ctf_field));
|
|
free(fixup);
|
|
if (rc == -1) {
|
|
RTE_PER_LCORE(trace_point_sz) = 0;
|
|
RTE_PER_LCORE(ctf_field) = NULL;
|
|
trace_crit("could not allocate CTF field");
|
|
return;
|
|
}
|
|
RTE_PER_LCORE(trace_point_sz) += sz;
|
|
RTE_PER_LCORE(ctf_field) = field;
|
|
}
|
|
|
|
int
|
|
__rte_trace_point_register(rte_trace_point_t *handle, const char *name,
|
|
void (*register_fn)(void))
|
|
{
|
|
struct trace_point *tp;
|
|
uint16_t sz;
|
|
|
|
/* Sanity checks of arguments */
|
|
if (name == NULL || register_fn == NULL || handle == NULL) {
|
|
trace_err("invalid arguments");
|
|
rte_errno = EINVAL;
|
|
goto fail;
|
|
}
|
|
|
|
/* Check the size of the trace point object */
|
|
RTE_PER_LCORE(trace_point_sz) = 0;
|
|
register_fn();
|
|
if (RTE_PER_LCORE(trace_point_sz) == 0) {
|
|
trace_err("missing rte_trace_emit_header() in register fn");
|
|
rte_errno = EBADF;
|
|
goto fail;
|
|
}
|
|
|
|
/* Is size overflowed */
|
|
if (RTE_PER_LCORE(trace_point_sz) > UINT16_MAX) {
|
|
trace_err("trace point size overflowed");
|
|
rte_errno = ENOSPC;
|
|
goto fail;
|
|
}
|
|
|
|
/* Are we running out of space to store trace points? */
|
|
if (trace.nb_trace_points > UINT16_MAX) {
|
|
trace_err("trace point exceeds the max count");
|
|
rte_errno = ENOSPC;
|
|
goto fail;
|
|
}
|
|
|
|
/* Get the size of the trace point */
|
|
sz = RTE_PER_LCORE(trace_point_sz);
|
|
tp = calloc(1, sizeof(struct trace_point));
|
|
if (tp == NULL) {
|
|
trace_err("fail to allocate trace point memory");
|
|
rte_errno = ENOMEM;
|
|
goto fail;
|
|
}
|
|
|
|
/* Initialize the trace point */
|
|
if (rte_strscpy(tp->name, name, TRACE_POINT_NAME_SIZE) < 0) {
|
|
trace_err("name is too long");
|
|
goto free;
|
|
}
|
|
|
|
/* Copy the accumulated fields description and clear it for the next
|
|
* trace point.
|
|
*/
|
|
tp->ctf_field = RTE_PER_LCORE(ctf_field);
|
|
RTE_PER_LCORE(ctf_field) = NULL;
|
|
|
|
/* Form the trace handle */
|
|
*handle = sz;
|
|
*handle |= trace.nb_trace_points << __RTE_TRACE_FIELD_ID_SHIFT;
|
|
|
|
trace.nb_trace_points++;
|
|
tp->handle = handle;
|
|
|
|
/* Add the trace point at tail */
|
|
STAILQ_INSERT_TAIL(&tp_list, tp, next);
|
|
__atomic_thread_fence(__ATOMIC_RELEASE);
|
|
|
|
/* All Good !!! */
|
|
return 0;
|
|
free:
|
|
free(tp);
|
|
fail:
|
|
if (trace.register_errno == 0)
|
|
trace.register_errno = rte_errno;
|
|
|
|
return -rte_errno;
|
|
}
|