// SPDX-License-Identifier: GPL-2.0 /* * builtin-test.c * * Builtin regression testing command: ever growing number of sanity tests */ #include #include #include #include #include #include #include #include #include #include "builtin.h" #include "config.h" #include "hist.h" #include "intlist.h" #include "tests.h" #include "debug.h" #include "color.h" #include #include "string2.h" #include "symbol.h" #include "util/rlimit.h" #include #include #include #include #include "builtin-test-list.h" static bool dont_fork; const char *dso_to_test; const char *test_objdump_path = "objdump"; /* * List of architecture specific tests. Not a weak symbol as the array length is * dependent on the initialization, as such GCC with LTO complains of * conflicting definitions with a weak symbol. */ #if defined(__i386__) || defined(__x86_64__) || defined(__aarch64__) || defined(__powerpc64__) extern struct test_suite *arch_tests[]; #else static struct test_suite *arch_tests[] = { NULL, }; #endif static struct test_suite *generic_tests[] = { &suite__vmlinux_matches_kallsyms, #ifdef HAVE_LIBTRACEEVENT &suite__openat_syscall_event, &suite__openat_syscall_event_on_all_cpus, &suite__basic_mmap, #endif &suite__mem, &suite__parse_events, &suite__expr, &suite__PERF_RECORD, &suite__pmu, &suite__pmu_events, &suite__dso_data, &suite__perf_evsel__roundtrip_name_test, #ifdef HAVE_LIBTRACEEVENT &suite__perf_evsel__tp_sched_test, &suite__syscall_openat_tp_fields, #endif &suite__attr, &suite__hists_link, &suite__python_use, &suite__bp_signal, &suite__bp_signal_overflow, &suite__bp_accounting, &suite__wp, &suite__task_exit, &suite__sw_clock_freq, &suite__code_reading, &suite__sample_parsing, &suite__keep_tracking, &suite__parse_no_sample_id_all, &suite__hists_filter, &suite__mmap_thread_lookup, &suite__thread_maps_share, &suite__hists_output, &suite__hists_cumulate, #ifdef HAVE_LIBTRACEEVENT &suite__switch_tracking, #endif &suite__fdarray__filter, &suite__fdarray__add, &suite__kmod_path__parse, &suite__thread_map, &suite__session_topology, &suite__thread_map_synthesize, &suite__thread_map_remove, &suite__cpu_map, &suite__synthesize_stat_config, &suite__synthesize_stat, &suite__synthesize_stat_round, &suite__event_update, &suite__event_times, &suite__backward_ring_buffer, &suite__sdt_event, &suite__is_printable_array, &suite__bitmap_print, &suite__perf_hooks, &suite__unit_number__scnprint, &suite__mem2node, &suite__time_utils, &suite__jit_write_elf, &suite__pfm, &suite__api_io, &suite__maps__merge_in, &suite__demangle_java, &suite__demangle_ocaml, &suite__parse_metric, &suite__pe_file_parsing, &suite__expand_cgroup_events, &suite__perf_time_to_tsc, &suite__dlfilter, &suite__sigtrap, &suite__event_groups, &suite__symbols, &suite__util, NULL, }; static struct test_suite **tests[] = { generic_tests, arch_tests, }; static struct test_workload *workloads[] = { &workload__noploop, &workload__thloop, &workload__leafloop, &workload__sqrtloop, &workload__brstack, &workload__datasym, }; static int num_subtests(const struct test_suite *t) { int num; if (!t->test_cases) return 0; num = 0; while (t->test_cases[num].name) num++; return num; } static bool has_subtests(const struct test_suite *t) { return num_subtests(t) > 1; } static const char *skip_reason(const struct test_suite *t, int subtest) { if (!t->test_cases) return NULL; return t->test_cases[subtest >= 0 ? subtest : 0].skip_reason; } static const char *test_description(const struct test_suite *t, int subtest) { if (t->test_cases && subtest >= 0) return t->test_cases[subtest].desc; return t->desc; } static test_fnptr test_function(const struct test_suite *t, int subtest) { if (subtest <= 0) return t->test_cases[0].run_case; return t->test_cases[subtest].run_case; } static bool perf_test__matches(const char *desc, int curr, int argc, const char *argv[]) { int i; if (argc == 0) return true; for (i = 0; i < argc; ++i) { char *end; long nr = strtoul(argv[i], &end, 10); if (*end == '\0') { if (nr == curr + 1) return true; continue; } if (strcasestr(desc, argv[i])) return true; } return false; } static int run_test(struct test_suite *test, int subtest) { int status, err = -1, child = dont_fork ? 0 : fork(); char sbuf[STRERR_BUFSIZE]; if (child < 0) { pr_err("failed to fork test: %s\n", str_error_r(errno, sbuf, sizeof(sbuf))); return -1; } if (!child) { if (!dont_fork) { pr_debug("test child forked, pid %d\n", getpid()); if (verbose <= 0) { int nullfd = open("/dev/null", O_WRONLY); if (nullfd >= 0) { close(STDERR_FILENO); close(STDOUT_FILENO); dup2(nullfd, STDOUT_FILENO); dup2(STDOUT_FILENO, STDERR_FILENO); close(nullfd); } } else { signal(SIGSEGV, sighandler_dump_stack); signal(SIGFPE, sighandler_dump_stack); } } err = test_function(test, subtest)(test, subtest); if (!dont_fork) exit(err); } if (!dont_fork) { wait(&status); if (WIFEXITED(status)) { err = (signed char)WEXITSTATUS(status); pr_debug("test child finished with %d\n", err); } else if (WIFSIGNALED(status)) { err = -1; pr_debug("test child interrupted\n"); } } return err; } #define for_each_test(j, k, t) \ for (j = 0, k = 0; j < ARRAY_SIZE(tests); j++, k = 0) \ while ((t = tests[j][k++]) != NULL) static int test_and_print(struct test_suite *t, int subtest) { int err; pr_debug("\n--- start ---\n"); err = run_test(t, subtest); pr_debug("---- end ----\n"); if (!has_subtests(t)) pr_debug("%s:", t->desc); else pr_debug("%s subtest %d:", t->desc, subtest + 1); switch (err) { case TEST_OK: pr_info(" Ok\n"); break; case TEST_SKIP: { const char *reason = skip_reason(t, subtest); if (reason) color_fprintf(stderr, PERF_COLOR_YELLOW, " Skip (%s)\n", reason); else color_fprintf(stderr, PERF_COLOR_YELLOW, " Skip\n"); } break; case TEST_FAIL: default: color_fprintf(stderr, PERF_COLOR_RED, " FAILED!\n"); break; } return err; } struct shell_test { const char *dir; const char *file; }; static int shell_test__run(struct test_suite *test, int subdir __maybe_unused) { int err; char script[PATH_MAX]; struct shell_test *st = test->priv; path__join(script, sizeof(script) - 3, st->dir, st->file); if (verbose > 0) strncat(script, " -v", sizeof(script) - strlen(script) - 1); err = system(script); if (!err) return TEST_OK; return WEXITSTATUS(err) == 2 ? TEST_SKIP : TEST_FAIL; } static int run_shell_tests(int argc, const char *argv[], int i, int width, struct intlist *skiplist) { struct shell_test st; const struct script_file *files, *file; files = list_script_files(); if (!files) return 0; for (file = files; file->dir; file++) { int curr = i++; struct test_case test_cases[] = { { .desc = file->desc, .run_case = shell_test__run, }, { .name = NULL, } }; struct test_suite test_suite = { .desc = test_cases[0].desc, .test_cases = test_cases, .priv = &st, }; st.dir = file->dir; if (test_suite.desc == NULL || !perf_test__matches(test_suite.desc, curr, argc, argv)) continue; st.file = file->file; pr_info("%3d: %-*s:", i, width, test_suite.desc); if (intlist__find(skiplist, i)) { color_fprintf(stderr, PERF_COLOR_YELLOW, " Skip (user override)\n"); continue; } test_and_print(&test_suite, 0); } return 0; } static int __cmd_test(int argc, const char *argv[], struct intlist *skiplist) { struct test_suite *t; unsigned int j, k; int i = 0; int width = list_script_max_width(); for_each_test(j, k, t) { int len = strlen(test_description(t, -1)); if (width < len) width = len; } for_each_test(j, k, t) { int curr = i++; int subi; if (!perf_test__matches(test_description(t, -1), curr, argc, argv)) { bool skip = true; int subn; subn = num_subtests(t); for (subi = 0; subi < subn; subi++) { if (perf_test__matches(test_description(t, subi), curr, argc, argv)) skip = false; } if (skip) continue; } pr_info("%3d: %-*s:", i, width, test_description(t, -1)); if (intlist__find(skiplist, i)) { color_fprintf(stderr, PERF_COLOR_YELLOW, " Skip (user override)\n"); continue; } if (!has_subtests(t)) { test_and_print(t, -1); } else { int subn = num_subtests(t); /* * minus 2 to align with normal testcases. * For subtest we print additional '.x' in number. * for example: * * 35: Test LLVM searching and compiling : * 35.1: Basic BPF llvm compiling test : Ok */ int subw = width > 2 ? width - 2 : width; if (subn <= 0) { color_fprintf(stderr, PERF_COLOR_YELLOW, " Skip (not compiled in)\n"); continue; } pr_info("\n"); for (subi = 0; subi < subn; subi++) { int len = strlen(test_description(t, subi)); if (subw < len) subw = len; } for (subi = 0; subi < subn; subi++) { if (!perf_test__matches(test_description(t, subi), curr, argc, argv)) continue; pr_info("%3d.%1d: %-*s:", i, subi + 1, subw, test_description(t, subi)); test_and_print(t, subi); } } } return run_shell_tests(argc, argv, i, width, skiplist); } static int perf_test__list_shell(int argc, const char **argv, int i) { const struct script_file *files, *file; files = list_script_files(); if (!files) return 0; for (file = files; file->dir; file++) { int curr = i++; struct test_suite t = { .desc = file->desc }; if (!perf_test__matches(t.desc, curr, argc, argv)) continue; pr_info("%3d: %s\n", i, t.desc); } return 0; } static int perf_test__list(int argc, const char **argv) { unsigned int j, k; struct test_suite *t; int i = 0; for_each_test(j, k, t) { int curr = i++; if (!perf_test__matches(test_description(t, -1), curr, argc, argv)) continue; pr_info("%3d: %s\n", i, test_description(t, -1)); if (has_subtests(t)) { int subn = num_subtests(t); int subi; for (subi = 0; subi < subn; subi++) pr_info("%3d:%1d: %s\n", i, subi + 1, test_description(t, subi)); } } perf_test__list_shell(argc, argv, i); return 0; } static int run_workload(const char *work, int argc, const char **argv) { unsigned int i = 0; struct test_workload *twl; for (i = 0; i < ARRAY_SIZE(workloads); i++) { twl = workloads[i]; if (!strcmp(twl->name, work)) return twl->func(argc, argv); } pr_info("No workload found: %s\n", work); return -1; } static int perf_test__config(const char *var, const char *value, void *data __maybe_unused) { if (!strcmp(var, "annotate.objdump")) test_objdump_path = value; return 0; } int cmd_test(int argc, const char **argv) { const char *test_usage[] = { "perf test [] [{list |[|]}]", NULL, }; const char *skip = NULL; const char *workload = NULL; const struct option test_options[] = { OPT_STRING('s', "skip", &skip, "tests", "tests to skip"), OPT_INCR('v', "verbose", &verbose, "be more verbose (show symbol address, etc)"), OPT_BOOLEAN('F', "dont-fork", &dont_fork, "Do not fork for testcase"), OPT_STRING('w', "workload", &workload, "work", "workload to run for testing"), OPT_STRING(0, "dso", &dso_to_test, "dso", "dso to test"), OPT_STRING(0, "objdump", &test_objdump_path, "path", "objdump binary to use for disassembly and annotations"), OPT_END() }; const char * const test_subcommands[] = { "list", NULL }; struct intlist *skiplist = NULL; int ret = hists__init(); if (ret < 0) return ret; perf_config(perf_test__config, NULL); /* Unbuffered output */ setvbuf(stdout, NULL, _IONBF, 0); argc = parse_options_subcommand(argc, argv, test_options, test_subcommands, test_usage, 0); if (argc >= 1 && !strcmp(argv[0], "list")) return perf_test__list(argc - 1, argv + 1); if (workload) return run_workload(workload, argc, argv); symbol_conf.priv_size = sizeof(int); symbol_conf.try_vmlinux_path = true; if (symbol__init(NULL) < 0) return -1; if (skip != NULL) skiplist = intlist__new(skip); /* * Tests that create BPF maps, for instance, need more than the 64K * default: */ rlimit__bump_memlock(); return __cmd_test(argc, argv, skiplist); }