@@ -3,7 +3,9 @@
*/
#include <inttypes.h>
+#include <fnmatch.h>
#include <sys/queue.h>
+#include <regex.h>
#include <rte_common.h>
#include <rte_errno.h>
@@ -20,6 +22,247 @@ RTE_DEFINE_PER_LCORE(int, ctf_count);
static struct trace_point_head tp_list = STAILQ_HEAD_INITIALIZER(tp_list);
static struct trace trace;
+bool
+rte_trace_global_is_enabled(void)
+{
+ return trace.global_status;
+}
+
+bool
+rte_trace_global_is_disabled(void)
+{
+ return !trace.global_status;
+}
+
+void
+rte_trace_global_level_set(uint32_t level)
+{
+ struct trace_point *tp;
+
+ trace.level = level;
+
+ if (rte_trace_global_is_disabled())
+ return;
+
+ STAILQ_FOREACH(tp, &tp_list, next) {
+ if (level >= rte_trace_level_get(tp->handle))
+ rte_trace_enable(tp->handle);
+ else
+ rte_trace_disable(tp->handle);
+ }
+}
+
+uint32_t
+rte_trace_global_level_get(void)
+{
+ return trace.level;
+}
+
+void
+rte_trace_global_mode_set(enum rte_trace_mode mode)
+{
+ struct trace_point *tp;
+
+ if (rte_trace_global_is_disabled())
+ return;
+
+ STAILQ_FOREACH(tp, &tp_list, next)
+ rte_trace_mode_set(tp->handle, mode);
+
+ trace.mode = mode;
+}
+
+enum
+rte_trace_mode rte_trace_global_mode_get(void)
+{
+ return trace.mode;
+}
+
+bool
+rte_trace_is_id_invalid(rte_trace_t *t)
+{
+ if (trace_id_get(t) >= trace.nb_trace_points)
+ return true;
+
+ return false;
+}
+
+bool
+rte_trace_is_enabled(rte_trace_t *trace)
+{
+ uint64_t val;
+
+ if (rte_trace_is_id_invalid(trace))
+ return false;
+
+ val = __atomic_load_n(trace, __ATOMIC_ACQUIRE);
+ return val & __RTE_TRACE_FIELD_ENABLE_MASK;
+}
+
+bool
+rte_trace_is_disabled(rte_trace_t *trace)
+{
+ uint64_t val;
+
+ if (rte_trace_is_id_invalid(trace))
+ return true;
+
+ val = __atomic_load_n(trace, __ATOMIC_ACQUIRE);
+ return !(val & __RTE_TRACE_FIELD_ENABLE_MASK);
+}
+
+int
+rte_trace_enable(rte_trace_t *trace)
+{
+ if (rte_trace_is_id_invalid(trace))
+ return -ERANGE;
+
+ if (rte_trace_level_get(trace) > rte_trace_global_level_get())
+ return -EACCES;
+
+ __atomic_or_fetch(trace, __RTE_TRACE_FIELD_ENABLE_MASK,
+ __ATOMIC_RELEASE);
+ return 0;
+}
+
+int
+rte_trace_disable(rte_trace_t *trace)
+{
+ if (rte_trace_is_id_invalid(trace))
+ return -ERANGE;
+
+ __atomic_and_fetch(trace, ~__RTE_TRACE_FIELD_ENABLE_MASK,
+ __ATOMIC_RELEASE);
+ return 0;
+}
+
+uint32_t
+rte_trace_level_get(rte_trace_t *trace)
+{
+ uint64_t val;
+
+ if (rte_trace_is_id_invalid(trace))
+ return 0;
+
+ val = __atomic_load_n(trace, __ATOMIC_ACQUIRE);
+ val &= __RTE_TRACE_FIELD_LEVEL_MASK;
+ val = val >> __RTE_TRACE_FIELD_LEVEL_SHIFT;
+
+ return val;
+}
+
+int
+rte_trace_level_set(rte_trace_t *trace, uint32_t level)
+{
+ uint64_t val;
+
+ if (rte_trace_is_id_invalid(trace) || level > RTE_LOG_DEBUG)
+ return -EINVAL;
+
+ val = __atomic_load_n(trace, __ATOMIC_ACQUIRE);
+ val &= ~__RTE_TRACE_FIELD_LEVEL_MASK;
+ val |= (uint64_t)level << __RTE_TRACE_FIELD_LEVEL_SHIFT;
+ __atomic_store_n(trace, val, __ATOMIC_RELEASE);
+
+ if (level <= rte_trace_global_level_get())
+ rte_trace_enable(trace);
+ else
+ rte_trace_disable(trace);
+
+ return 0;
+}
+
+int
+rte_trace_mode_get(rte_trace_t *trace)
+{
+ uint64_t val;
+
+ if (rte_trace_is_id_invalid(trace))
+ return -EINVAL;
+
+ val = __atomic_load_n(trace, __ATOMIC_ACQUIRE);
+
+ return !!(val & __RTE_TRACE_FIELD_ENABLE_DISCARD);
+}
+
+int
+rte_trace_mode_set(rte_trace_t *trace, enum rte_trace_mode mode)
+{
+ if (rte_trace_is_id_invalid(trace) || mode > RTE_TRACE_MODE_DISCARD)
+ return -EINVAL;
+
+ if (mode == RTE_TRACE_MODE_OVERWRITE)
+ __atomic_and_fetch(trace, ~__RTE_TRACE_FIELD_ENABLE_DISCARD,
+ __ATOMIC_RELEASE);
+ else
+ __atomic_or_fetch(trace, __RTE_TRACE_FIELD_ENABLE_DISCARD,
+ __ATOMIC_RELEASE);
+
+ return 0;
+}
+
+int
+rte_trace_pattern(const char *pattern, bool enable)
+{
+ struct trace_point *tp;
+ int rc = 0, found = 0;
+
+ STAILQ_FOREACH(tp, &tp_list, next) {
+ if (fnmatch(pattern, tp->name, 0) == 0) {
+ if (enable)
+ rc = rte_trace_enable(tp->handle);
+ else
+ rc = rte_trace_disable(tp->handle);
+ found = 1;
+ }
+ if (rc < 0)
+ return rc;
+ }
+
+ return rc | found;
+}
+
+int
+rte_trace_regexp(const char *regex, bool enable)
+{
+ struct trace_point *tp;
+ int rc = 0, found = 0;
+ regex_t r;
+
+ if (regcomp(&r, regex, 0) != 0)
+ return -EINVAL;
+
+ STAILQ_FOREACH(tp, &tp_list, next) {
+ if (regexec(&r, tp->name, 0, NULL, 0) == 0) {
+ if (enable)
+ rc = rte_trace_enable(tp->handle);
+ else
+ rc = rte_trace_disable(tp->handle);
+ found = 1;
+ }
+ if (rc < 0)
+ return rc;
+ }
+ regfree(&r);
+
+ return rc | found;
+}
+
+rte_trace_t *
+rte_trace_by_name(const char *name)
+{
+ struct trace_point *tp;
+
+ if (name == NULL)
+ return NULL;
+
+ STAILQ_FOREACH(tp, &tp_list, next)
+ if (strncmp(tp->name, name, TRACE_POINT_NAME_SIZE) == 0)
+ return tp->handle;
+
+ return NULL;
+}
+
int
__rte_trace_point_register(rte_trace_t *handle, const char *name,
uint32_t level, void (*register_fn)(void))
@@ -27,9 +27,20 @@ struct trace_point {
struct trace {
int register_errno;
+ bool global_status;
+ enum rte_trace_mode mode;
+ uint32_t level;
uint32_t nb_trace_points;
};
+/* Helper functions */
+static inline uint16_t
+trace_id_get(rte_trace_t *trace)
+{
+ return (*trace & __RTE_TRACE_FIELD_ID_MASK) >>
+ __RTE_TRACE_FIELD_ID_SHIFT;
+}
+
/* Trace point list functions */
STAILQ_HEAD(trace_point_head, trace_point);
@@ -340,4 +340,22 @@ EXPERIMENTAL {
rte_log_can_log;
rte_thread_getname;
__rte_trace_point_register;
+ rte_trace_global_is_enabled;
+ rte_trace_global_is_disabled;
+ rte_trace_is_id_invalid;
+ rte_trace_global_level_set;
+ rte_trace_global_level_get;
+ rte_trace_global_mode_set;
+ rte_trace_global_mode_get;
+ rte_trace_enable;
+ rte_trace_disable;
+ rte_trace_is_enabled;
+ rte_trace_is_disabled;
+ rte_trace_level_set;
+ rte_trace_level_get;
+ rte_trace_mode_set;
+ rte_trace_mode_get;
+ rte_trace_pattern;
+ rte_trace_regexp;
+ rte_trace_by_name;
};