[PATCH 4/5] perf trace: Remove unused code
From: Namhyung Kim
Date: Thu Aug 14 2025 - 03:19:51 EST
Now syscall init for augmented arguments is simplified. Let's get rid
of dead code.
Signed-off-by: Namhyung Kim <namhyung@xxxxxxxxxx>
---
tools/perf/builtin-trace.c | 110 -------------------------------------
1 file changed, 110 deletions(-)
diff --git a/tools/perf/builtin-trace.c b/tools/perf/builtin-trace.c
index e1caa82bc427b68b..a7a49d8997d55594 100644
--- a/tools/perf/builtin-trace.c
+++ b/tools/perf/builtin-trace.c
@@ -470,38 +470,6 @@ static int evsel__init_syscall_tp(struct evsel *evsel)
return -ENOMEM;
}
-static int evsel__init_augmented_syscall_tp(struct evsel *evsel, struct evsel *tp)
-{
- struct syscall_tp *sc = evsel__syscall_tp(evsel);
-
- if (sc != NULL) {
- struct tep_format_field *syscall_id = evsel__field(tp, "id");
- if (syscall_id == NULL)
- syscall_id = evsel__field(tp, "__syscall_nr");
- if (syscall_id == NULL ||
- __tp_field__init_uint(&sc->id, syscall_id->size, syscall_id->offset, evsel->needs_swap))
- return -EINVAL;
-
- return 0;
- }
-
- return -ENOMEM;
-}
-
-static int evsel__init_augmented_syscall_tp_args(struct evsel *evsel)
-{
- struct syscall_tp *sc = __evsel__syscall_tp(evsel);
-
- return __tp_field__init_ptr(&sc->args, sc->id.offset + sizeof(u64));
-}
-
-static int evsel__init_augmented_syscall_tp_ret(struct evsel *evsel)
-{
- struct syscall_tp *sc = __evsel__syscall_tp(evsel);
-
- return __tp_field__init_uint(&sc->ret, sizeof(u64), sc->id.offset + sizeof(u64), evsel->needs_swap);
-}
-
static int evsel__init_raw_syscall_tp(struct evsel *evsel, void *handler)
{
if (evsel__syscall_tp(evsel) != NULL) {
@@ -5506,7 +5474,6 @@ int cmd_trace(int argc, const char **argv)
};
bool __maybe_unused max_stack_user_set = true;
bool mmap_pages_user_set = true;
- struct evsel *evsel;
const char * const trace_subcommands[] = { "record", NULL };
int err = -1;
char bf[BUFSIZ];
@@ -5665,83 +5632,6 @@ int cmd_trace(int argc, const char **argv)
}
}
- /*
- * If we are augmenting syscalls, then combine what we put in the
- * __augmented_syscalls__ BPF map with what is in the
- * syscalls:sys_exit_FOO tracepoints, i.e. just like we do without BPF,
- * combining raw_syscalls:sys_enter with raw_syscalls:sys_exit.
- *
- * We'll switch to look at two BPF maps, one for sys_enter and the
- * other for sys_exit when we start augmenting the sys_exit paths with
- * buffers that are being copied from kernel to userspace, think 'read'
- * syscall.
- */
- if (trace.syscalls.events.bpf_output) {
- evlist__for_each_entry(trace.evlist, evsel) {
- bool raw_syscalls_sys_exit = evsel__name_is(evsel, "raw_syscalls:sys_exit");
-
- if (raw_syscalls_sys_exit) {
- trace.raw_augmented_syscalls = true;
- goto init_augmented_syscall_tp;
- }
-
- if (trace.syscalls.events.bpf_output->priv == NULL &&
- strstr(evsel__name(evsel), "syscalls:sys_enter")) {
- struct evsel *augmented = trace.syscalls.events.bpf_output;
- if (evsel__init_augmented_syscall_tp(augmented, evsel) ||
- evsel__init_augmented_syscall_tp_args(augmented))
- goto out;
- /*
- * Augmented is __augmented_syscalls__ BPF_OUTPUT event
- * Above we made sure we can get from the payload the tp fields
- * that we get from syscalls:sys_enter tracefs format file.
- */
- augmented->handler = trace__sys_enter;
- /*
- * Now we do the same for the *syscalls:sys_enter event so that
- * if we handle it directly, i.e. if the BPF prog returns 0 so
- * as not to filter it, then we'll handle it just like we would
- * for the BPF_OUTPUT one:
- */
- if (evsel__init_augmented_syscall_tp(evsel, evsel) ||
- evsel__init_augmented_syscall_tp_args(evsel))
- goto out;
- evsel->handler = trace__sys_enter;
- }
-
- if (strstarts(evsel__name(evsel), "syscalls:sys_exit_")) {
- struct syscall_tp *sc;
-init_augmented_syscall_tp:
- if (evsel__init_augmented_syscall_tp(evsel, evsel))
- goto out;
- sc = __evsel__syscall_tp(evsel);
- /*
- * For now with BPF raw_augmented we hook into
- * raw_syscalls:sys_enter and there we get all
- * 6 syscall args plus the tracepoint common
- * fields and the syscall_nr (another long).
- * So we check if that is the case and if so
- * don't look after the sc->args_size but
- * always after the full raw_syscalls:sys_enter
- * payload, which is fixed.
- *
- * We'll revisit this later to pass
- * s->args_size to the BPF augmenter (now
- * tools/perf/examples/bpf/augmented_raw_syscalls.c,
- * so that it copies only what we need for each
- * syscall, like what happens when we use
- * syscalls:sys_enter_NAME, so that we reduce
- * the kernel/userspace traffic to just what is
- * needed for each syscall.
- */
- if (trace.raw_augmented_syscalls)
- trace.raw_augmented_syscalls_args_size = (6 + 1) * sizeof(long) + sc->id.offset;
- evsel__init_augmented_syscall_tp_ret(evsel);
- evsel->handler = trace__sys_exit;
- }
- }
- }
-
if ((argc >= 1) && (strcmp(argv[0], "record") == 0)) {
err = trace__record(&trace, argc-1, &argv[1]);
goto out;
--
2.51.0.rc1.167.g924127e9c0-goog