Skip to content
Navigation Menu
Toggle navigation
Sign in
In this repository
All GitHub Enterprise
↵
Jump to
↵
No suggested jump to results
In this repository
All GitHub Enterprise
↵
Jump to
↵
In this organization
All GitHub Enterprise
↵
Jump to
↵
In this repository
All GitHub Enterprise
↵
Jump to
↵
Sign in
Reseting focus
You signed in with another tab or window.
Reload
to refresh your session.
You signed out in another tab or window.
Reload
to refresh your session.
You switched accounts on another tab or window.
Reload
to refresh your session.
Dismiss alert
{{ message }}
mariux64
/
linux
Public
Notifications
You must be signed in to change notification settings
Fork
0
Star
0
Code
Issues
2
Pull requests
0
Actions
Projects
0
Wiki
Security
Insights
Additional navigation options
Code
Issues
Pull requests
Actions
Projects
Wiki
Security
Insights
Files
936be50
Documentation
arch
block
crypto
drivers
firmware
fs
include
init
ipc
kernel
lib
mm
net
samples
scripts
security
sound
tools
firewire
perf
Documentation
arch
bench
config
python
scripts
util
.gitignore
CREDITS
MANIFEST
Makefile
builtin-annotate.c
builtin-bench.c
builtin-buildid-cache.c
builtin-buildid-list.c
builtin-diff.c
builtin-evlist.c
builtin-help.c
builtin-inject.c
builtin-kmem.c
builtin-kvm.c
builtin-list.c
builtin-lock.c
builtin-probe.c
builtin-record.c
builtin-report.c
builtin-sched.c
builtin-script.c
builtin-stat.c
builtin-test.c
builtin-timechart.c
builtin-top.c
builtin.h
command-list.txt
design.txt
perf-archive.sh
perf.c
perf.h
power
slub
testing
usb
virtio
usr
virt
.gitignore
.mailmap
COPYING
CREDITS
Kbuild
Kconfig
MAINTAINERS
Makefile
README
REPORTING-BUGS
Breadcrumbs
linux
/
tools
/
perf
/
builtin-test.c
Copy path
Blame
Blame
Latest commit
History
History
919 lines (791 loc) · 23.2 KB
Breadcrumbs
linux
/
tools
/
perf
/
builtin-test.c
Top
File metadata and controls
Code
Blame
919 lines (791 loc) · 23.2 KB
Raw
/* * builtin-test.c * * Builtin regression testing command: ever growing number of sanity tests */ #include "builtin.h" #include "util/cache.h" #include "util/debug.h" #include "util/evlist.h" #include "util/parse-options.h" #include "util/parse-events.h" #include "util/symbol.h" #include "util/thread_map.h" #include "../../include/linux/hw_breakpoint.h" static long page_size; static int vmlinux_matches_kallsyms_filter(struct map *map __used, struct symbol *sym) { bool *visited = symbol__priv(sym); *visited = true; return 0; } static int test__vmlinux_matches_kallsyms(void) { int err = -1; struct rb_node *nd; struct symbol *sym; struct map *kallsyms_map, *vmlinux_map; struct machine kallsyms, vmlinux; enum map_type type = MAP__FUNCTION; struct ref_reloc_sym ref_reloc_sym = { .name = "_stext", }; /* * Step 1: * * Init the machines that will hold kernel, modules obtained from * both vmlinux + .ko files and from /proc/kallsyms split by modules. */ machine__init(&kallsyms, "", HOST_KERNEL_ID); machine__init(&vmlinux, "", HOST_KERNEL_ID); /* * Step 2: * * Create the kernel maps for kallsyms and the DSO where we will then * load /proc/kallsyms. Also create the modules maps from /proc/modules * and find the .ko files that match them in /lib/modules/`uname -r`/. */ if (machine__create_kernel_maps(&kallsyms) < 0) { pr_debug("machine__create_kernel_maps "); return -1; } /* * Step 3: * * Load and split /proc/kallsyms into multiple maps, one per module. */ if (machine__load_kallsyms(&kallsyms, "/proc/kallsyms", type, NULL) <= 0) { pr_debug("dso__load_kallsyms "); goto out; } /* * Step 4: * * kallsyms will be internally on demand sorted by name so that we can * find the reference relocation * symbol, i.e. the symbol we will use * to see if the running kernel was relocated by checking if it has the * same value in the vmlinux file we load. */ kallsyms_map = machine__kernel_map(&kallsyms, type); sym = map__find_symbol_by_name(kallsyms_map, ref_reloc_sym.name, NULL); if (sym == NULL) { pr_debug("dso__find_symbol_by_name "); goto out; } ref_reloc_sym.addr = sym->start; /* * Step 5: * * Now repeat step 2, this time for the vmlinux file we'll auto-locate. */ if (machine__create_kernel_maps(&vmlinux) < 0) { pr_debug("machine__create_kernel_maps "); goto out; } vmlinux_map = machine__kernel_map(&vmlinux, type); map__kmap(vmlinux_map)->ref_reloc_sym = &ref_reloc_sym; /* * Step 6: * * Locate a vmlinux file in the vmlinux path that has a buildid that * matches the one of the running kernel. * * While doing that look if we find the ref reloc symbol, if we find it * we'll have its ref_reloc_symbol.unrelocated_addr and then * maps__reloc_vmlinux will notice and set proper ->[un]map_ip routines * to fixup the symbols. */ if (machine__load_vmlinux_path(&vmlinux, type, vmlinux_matches_kallsyms_filter) <= 0) { pr_debug("machine__load_vmlinux_path "); goto out; } err = 0; /* * Step 7: * * Now look at the symbols in the vmlinux DSO and check if we find all of them * in the kallsyms dso. For the ones that are in both, check its names and * end addresses too. */ for (nd = rb_first(&vmlinux_map->dso->symbols[type]); nd; nd = rb_next(nd)) { struct symbol *pair, *first_pair; bool backwards = true; sym = rb_entry(nd, struct symbol, rb_node); if (sym->start == sym->end) continue; first_pair = machine__find_kernel_symbol(&kallsyms, type, sym->start, NULL, NULL); pair = first_pair; if (pair && pair->start == sym->start) { next_pair: if (strcmp(sym->name, pair->name) == 0) { /* * kallsyms don't have the symbol end, so we * set that by using the next symbol start - 1, * in some cases we get this up to a page * wrong, trace_kmalloc when I was developing * this code was one such example, 2106 bytes * off the real size. More than that and we * _really_ have a problem. */ s64 skew = sym->end - pair->end; if (llabs(skew) < page_size) continue; pr_debug("%#" PRIx64 ": diff end addr for %s v: %#" PRIx64 " k: %#" PRIx64 "\n", sym->start, sym->name, sym->end, pair->end); } else { struct rb_node *nnd; detour: nnd = backwards ? rb_prev(&pair->rb_node) : rb_next(&pair->rb_node); if (nnd) { struct symbol *next = rb_entry(nnd, struct symbol, rb_node); if (next->start == sym->start) { pair = next; goto next_pair; } } if (backwards) { backwards = false; pair = first_pair; goto detour; } pr_debug("%#" PRIx64 ": diff name v: %s k: %s\n", sym->start, sym->name, pair->name); } } else pr_debug("%#" PRIx64 ": %s not on kallsyms\n", sym->start, sym->name); err = -1; } if (!verbose) goto out; pr_info("Maps only in vmlinux:\n"); for (nd = rb_first(&vmlinux.kmaps.maps[type]); nd; nd = rb_next(nd)) { struct map *pos = rb_entry(nd, struct map, rb_node), *pair; /* * If it is the kernel, kallsyms is always "[kernel.kallsyms]", while * the kernel will have the path for the vmlinux file being used, * so use the short name, less descriptive but the same ("[kernel]" in * both cases. */ pair = map_groups__find_by_name(&kallsyms.kmaps, type, (pos->dso->kernel ? pos->dso->short_name : pos->dso->name)); if (pair) pair->priv = 1; else map__fprintf(pos, stderr); } pr_info("Maps in vmlinux with a different name in kallsyms:\n"); for (nd = rb_first(&vmlinux.kmaps.maps[type]); nd; nd = rb_next(nd)) { struct map *pos = rb_entry(nd, struct map, rb_node), *pair; pair = map_groups__find(&kallsyms.kmaps, type, pos->start); if (pair == NULL || pair->priv) continue; if (pair->start == pos->start) { pair->priv = 1; pr_info(" %" PRIx64 "-%" PRIx64 " %" PRIx64 " %s in kallsyms as", pos->start, pos->end, pos->pgoff, pos->dso->name); if (pos->pgoff != pair->pgoff || pos->end != pair->end) pr_info(": \n*%" PRIx64 "-%" PRIx64 " %" PRIx64 "", pair->start, pair->end, pair->pgoff); pr_info(" %s\n", pair->dso->name); pair->priv = 1; } } pr_info("Maps only in kallsyms:\n"); for (nd = rb_first(&kallsyms.kmaps.maps[type]); nd; nd = rb_next(nd)) { struct map *pos = rb_entry(nd, struct map, rb_node); if (!pos->priv) map__fprintf(pos, stderr); } out: return err; } #include "util/cpumap.h" #include "util/evsel.h" #include <sys/types.h> static int trace_event__id(const char *evname) { char *filename; int err = -1, fd; if (asprintf(&filename, "%s/syscalls/%s/id", debugfs_path, evname) < 0) return -1; fd = open(filename, O_RDONLY); if (fd >= 0) { char id[16]; if (read(fd, id, sizeof(id)) > 0) err = atoi(id); close(fd); } free(filename); return err; } static int test__open_syscall_event(void) { int err = -1, fd; struct thread_map *threads; struct perf_evsel *evsel; struct perf_event_attr attr; unsigned int nr_open_calls = 111, i; int id = trace_event__id("sys_enter_open"); if (id < 0) { pr_debug("is debugfs mounted on /sys/kernel/debug?\n"); return -1; } threads = thread_map__new(-1, getpid()); if (threads == NULL) { pr_debug("thread_map__new\n"); return -1; } memset(&attr, 0, sizeof(attr)); attr.type = PERF_TYPE_TRACEPOINT; attr.config = id; evsel = perf_evsel__new(&attr, 0); if (evsel == NULL) { pr_debug("perf_evsel__new\n"); goto out_thread_map_delete; } if (perf_evsel__open_per_thread(evsel, threads, false) < 0) { pr_debug("failed to open counter: %s, " "tweak /proc/sys/kernel/perf_event_paranoid?\n", strerror(errno)); goto out_evsel_delete; } for (i = 0; i < nr_open_calls; ++i) { fd = open("/etc/passwd", O_RDONLY); close(fd); } if (perf_evsel__read_on_cpu(evsel, 0, 0) < 0) { pr_debug("perf_evsel__read_on_cpu\n"); goto out_close_fd; } if (evsel->counts->cpu[0].val != nr_open_calls) { pr_debug("perf_evsel__read_on_cpu: expected to intercept %d calls, got %" PRIu64 "\n", nr_open_calls, evsel->counts->cpu[0].val); goto out_close_fd; } err = 0; out_close_fd: perf_evsel__close_fd(evsel, 1, threads->nr); out_evsel_delete: perf_evsel__delete(evsel); out_thread_map_delete: thread_map__delete(threads); return err; } #include <sched.h> static int test__open_syscall_event_on_all_cpus(void) { int err = -1, fd, cpu; struct thread_map *threads; struct cpu_map *cpus; struct perf_evsel *evsel; struct perf_event_attr attr; unsigned int nr_open_calls = 111, i; cpu_set_t cpu_set; int id = trace_event__id("sys_enter_open"); if (id < 0) { pr_debug("is debugfs mounted on /sys/kernel/debug?\n"); return -1; } threads = thread_map__new(-1, getpid()); if (threads == NULL) { pr_debug("thread_map__new\n"); return -1; } cpus = cpu_map__new(NULL); if (cpus == NULL) { pr_debug("cpu_map__new\n"); goto out_thread_map_delete; } CPU_ZERO(&cpu_set); memset(&attr, 0, sizeof(attr)); attr.type = PERF_TYPE_TRACEPOINT; attr.config = id; evsel = perf_evsel__new(&attr, 0); if (evsel == NULL) { pr_debug("perf_evsel__new\n"); goto out_thread_map_delete; } if (perf_evsel__open(evsel, cpus, threads, false) < 0) { pr_debug("failed to open counter: %s, " "tweak /proc/sys/kernel/perf_event_paranoid?\n", strerror(errno)); goto out_evsel_delete; } for (cpu = 0; cpu < cpus->nr; ++cpu) { unsigned int ncalls = nr_open_calls + cpu; /* * XXX eventually lift this restriction in a way that * keeps perf building on older glibc installations * without CPU_ALLOC. 1024 cpus in 2010 still seems * a reasonable upper limit tho :-) */ if (cpus->map[cpu] >= CPU_SETSIZE) { pr_debug("Ignoring CPU %d\n", cpus->map[cpu]); continue; } CPU_SET(cpus->map[cpu], &cpu_set); if (sched_setaffinity(0, sizeof(cpu_set), &cpu_set) < 0) { pr_debug("sched_setaffinity() failed on CPU %d: %s ", cpus->map[cpu], strerror(errno)); goto out_close_fd; } for (i = 0; i < ncalls; ++i) { fd = open("/etc/passwd", O_RDONLY); close(fd); } CPU_CLR(cpus->map[cpu], &cpu_set); } /* * Here we need to explicitely preallocate the counts, as if * we use the auto allocation it will allocate just for 1 cpu, * as we start by cpu 0. */ if (perf_evsel__alloc_counts(evsel, cpus->nr) < 0) { pr_debug("perf_evsel__alloc_counts(ncpus=%d)\n", cpus->nr); goto out_close_fd; } err = 0; for (cpu = 0; cpu < cpus->nr; ++cpu) { unsigned int expected; if (cpus->map[cpu] >= CPU_SETSIZE) continue; if (perf_evsel__read_on_cpu(evsel, cpu, 0) < 0) { pr_debug("perf_evsel__read_on_cpu\n"); err = -1; break; } expected = nr_open_calls + cpu; if (evsel->counts->cpu[cpu].val != expected) { pr_debug("perf_evsel__read_on_cpu: expected to intercept %d calls on cpu %d, got %" PRIu64 "\n", expected, cpus->map[cpu], evsel->counts->cpu[cpu].val); err = -1; } } out_close_fd: perf_evsel__close_fd(evsel, 1, threads->nr); out_evsel_delete: perf_evsel__delete(evsel); out_thread_map_delete: thread_map__delete(threads); return err; } /* * This test will generate random numbers of calls to some getpid syscalls, * then establish an mmap for a group of events that are created to monitor * the syscalls. * * It will receive the events, using mmap, use its PERF_SAMPLE_ID generated * sample.id field to map back to its respective perf_evsel instance. * * Then it checks if the number of syscalls reported as perf events by * the kernel corresponds to the number of syscalls made. */ static int test__basic_mmap(void) { int err = -1; union perf_event *event; struct thread_map *threads; struct cpu_map *cpus; struct perf_evlist *evlist; struct perf_event_attr attr = { .type = PERF_TYPE_TRACEPOINT, .read_format = PERF_FORMAT_ID, .sample_type = PERF_SAMPLE_ID, .watermark = 0, }; cpu_set_t cpu_set; const char *syscall_names[] = { "getsid", "getppid", "getpgrp", "getpgid", }; pid_t (*syscalls[])(void) = { (void *)getsid, getppid, getpgrp, (void*)getpgid }; #define nsyscalls ARRAY_SIZE(syscall_names) int ids[nsyscalls]; unsigned int nr_events[nsyscalls], expected_nr_events[nsyscalls], i, j; struct perf_evsel *evsels[nsyscalls], *evsel; int sample_size = __perf_evsel__sample_size(attr.sample_type); for (i = 0; i < nsyscalls; ++i) { char name[64]; snprintf(name, sizeof(name), "sys_enter_%s", syscall_names[i]); ids[i] = trace_event__id(name); if (ids[i] < 0) { pr_debug("Is debugfs mounted on /sys/kernel/debug?\n"); return -1; } nr_events[i] = 0; expected_nr_events[i] = random() % 257; } threads = thread_map__new(-1, getpid()); if (threads == NULL) { pr_debug("thread_map__new\n"); return -1; } cpus = cpu_map__new(NULL); if (cpus == NULL) { pr_debug("cpu_map__new\n"); goto out_free_threads; } CPU_ZERO(&cpu_set); CPU_SET(cpus->map[0], &cpu_set); sched_setaffinity(0, sizeof(cpu_set), &cpu_set); if (sched_setaffinity(0, sizeof(cpu_set), &cpu_set) < 0) { pr_debug("sched_setaffinity() failed on CPU %d: %s ", cpus->map[0], strerror(errno)); goto out_free_cpus; } evlist = perf_evlist__new(cpus, threads); if (evlist == NULL) { pr_debug("perf_evlist__new\n"); goto out_free_cpus; } /* anonymous union fields, can't be initialized above */ attr.wakeup_events = 1; attr.sample_period = 1; for (i = 0; i < nsyscalls; ++i) { attr.config = ids[i]; evsels[i] = perf_evsel__new(&attr, i); if (evsels[i] == NULL) { pr_debug("perf_evsel__new\n"); goto out_free_evlist; } perf_evlist__add(evlist, evsels[i]); if (perf_evsel__open(evsels[i], cpus, threads, false) < 0) { pr_debug("failed to open counter: %s, " "tweak /proc/sys/kernel/perf_event_paranoid?\n", strerror(errno)); goto out_close_fd; } } if (perf_evlist__mmap(evlist, 128, true) < 0) { pr_debug("failed to mmap events: %d (%s)\n", errno, strerror(errno)); goto out_close_fd; } for (i = 0; i < nsyscalls; ++i) for (j = 0; j < expected_nr_events[i]; ++j) { int foo = syscalls[i](); ++foo; } while ((event = perf_evlist__mmap_read(evlist, 0)) != NULL) { struct perf_sample sample; if (event->header.type != PERF_RECORD_SAMPLE) { pr_debug("unexpected %s event\n", perf_event__name(event->header.type)); goto out_munmap; } err = perf_event__parse_sample(event, attr.sample_type, sample_size, false, &sample, false); if (err) { pr_err("Can't parse sample, err = %d\n", err); goto out_munmap; } evsel = perf_evlist__id2evsel(evlist, sample.id); if (evsel == NULL) { pr_debug("event with id %" PRIu64 " doesn't map to an evsel\n", sample.id); goto out_munmap; } nr_events[evsel->idx]++; } list_for_each_entry(evsel, &evlist->entries, node) { if (nr_events[evsel->idx] != expected_nr_events[evsel->idx]) { pr_debug("expected %d %s events, got %d\n", expected_nr_events[evsel->idx], event_name(evsel), nr_events[evsel->idx]); goto out_munmap; } } err = 0; out_munmap: perf_evlist__munmap(evlist); out_close_fd: for (i = 0; i < nsyscalls; ++i) perf_evsel__close_fd(evsels[i], 1, threads->nr); out_free_evlist: perf_evlist__delete(evlist); out_free_cpus: cpu_map__delete(cpus); out_free_threads: thread_map__delete(threads); return err; #undef nsyscalls } #define TEST_ASSERT_VAL(text, cond) \ do { \ if (!cond) { \ pr_debug("FAILED %s:%d %s\n", __FILE__, __LINE__, text); \ return -1; \ } \ } while (0) static int test__checkevent_tracepoint(struct perf_evlist *evlist) { struct perf_evsel *evsel = list_entry(evlist->entries.next, struct perf_evsel, node); TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); TEST_ASSERT_VAL("wrong type", PERF_TYPE_TRACEPOINT == evsel->attr.type); TEST_ASSERT_VAL("wrong sample_type", (PERF_SAMPLE_RAW | PERF_SAMPLE_TIME | PERF_SAMPLE_CPU) == evsel->attr.sample_type); TEST_ASSERT_VAL("wrong sample_period", 1 == evsel->attr.sample_period); return 0; } static int test__checkevent_tracepoint_multi(struct perf_evlist *evlist) { struct perf_evsel *evsel; TEST_ASSERT_VAL("wrong number of entries", evlist->nr_entries > 1); list_for_each_entry(evsel, &evlist->entries, node) { TEST_ASSERT_VAL("wrong type", PERF_TYPE_TRACEPOINT == evsel->attr.type); TEST_ASSERT_VAL("wrong sample_type", (PERF_SAMPLE_RAW | PERF_SAMPLE_TIME | PERF_SAMPLE_CPU) == evsel->attr.sample_type); TEST_ASSERT_VAL("wrong sample_period", 1 == evsel->attr.sample_period); } return 0; } static int test__checkevent_raw(struct perf_evlist *evlist) { struct perf_evsel *evsel = list_entry(evlist->entries.next, struct perf_evsel, node); TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); TEST_ASSERT_VAL("wrong type", PERF_TYPE_RAW == evsel->attr.type); TEST_ASSERT_VAL("wrong config", 1 == evsel->attr.config); return 0; } static int test__checkevent_numeric(struct perf_evlist *evlist) { struct perf_evsel *evsel = list_entry(evlist->entries.next, struct perf_evsel, node); TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); TEST_ASSERT_VAL("wrong type", 1 == evsel->attr.type); TEST_ASSERT_VAL("wrong config", 1 == evsel->attr.config); return 0; } static int test__checkevent_symbolic_name(struct perf_evlist *evlist) { struct perf_evsel *evsel = list_entry(evlist->entries.next, struct perf_evsel, node); TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); TEST_ASSERT_VAL("wrong type", PERF_TYPE_HARDWARE == evsel->attr.type); TEST_ASSERT_VAL("wrong config", PERF_COUNT_HW_INSTRUCTIONS == evsel->attr.config); return 0; } static int test__checkevent_symbolic_alias(struct perf_evlist *evlist) { struct perf_evsel *evsel = list_entry(evlist->entries.next, struct perf_evsel, node); TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); TEST_ASSERT_VAL("wrong type", PERF_TYPE_SOFTWARE == evsel->attr.type); TEST_ASSERT_VAL("wrong config", PERF_COUNT_SW_PAGE_FAULTS == evsel->attr.config); return 0; } static int test__checkevent_genhw(struct perf_evlist *evlist) { struct perf_evsel *evsel = list_entry(evlist->entries.next, struct perf_evsel, node); TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); TEST_ASSERT_VAL("wrong type", PERF_TYPE_HW_CACHE == evsel->attr.type); TEST_ASSERT_VAL("wrong config", (1 << 16) == evsel->attr.config); return 0; } static int test__checkevent_breakpoint(struct perf_evlist *evlist) { struct perf_evsel *evsel = list_entry(evlist->entries.next, struct perf_evsel, node); TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); TEST_ASSERT_VAL("wrong type", PERF_TYPE_BREAKPOINT == evsel->attr.type); TEST_ASSERT_VAL("wrong config", 0 == evsel->attr.config); TEST_ASSERT_VAL("wrong bp_type", (HW_BREAKPOINT_R | HW_BREAKPOINT_W) == evsel->attr.bp_type); TEST_ASSERT_VAL("wrong bp_len", HW_BREAKPOINT_LEN_4 == evsel->attr.bp_len); return 0; } static int test__checkevent_breakpoint_x(struct perf_evlist *evlist) { struct perf_evsel *evsel = list_entry(evlist->entries.next, struct perf_evsel, node); TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); TEST_ASSERT_VAL("wrong type", PERF_TYPE_BREAKPOINT == evsel->attr.type); TEST_ASSERT_VAL("wrong config", 0 == evsel->attr.config); TEST_ASSERT_VAL("wrong bp_type", HW_BREAKPOINT_X == evsel->attr.bp_type); TEST_ASSERT_VAL("wrong bp_len", sizeof(long) == evsel->attr.bp_len); return 0; } static int test__checkevent_breakpoint_r(struct perf_evlist *evlist) { struct perf_evsel *evsel = list_entry(evlist->entries.next, struct perf_evsel, node); TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); TEST_ASSERT_VAL("wrong type", PERF_TYPE_BREAKPOINT == evsel->attr.type); TEST_ASSERT_VAL("wrong config", 0 == evsel->attr.config); TEST_ASSERT_VAL("wrong bp_type", HW_BREAKPOINT_R == evsel->attr.bp_type); TEST_ASSERT_VAL("wrong bp_len", HW_BREAKPOINT_LEN_4 == evsel->attr.bp_len); return 0; } static int test__checkevent_breakpoint_w(struct perf_evlist *evlist) { struct perf_evsel *evsel = list_entry(evlist->entries.next, struct perf_evsel, node); TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); TEST_ASSERT_VAL("wrong type", PERF_TYPE_BREAKPOINT == evsel->attr.type); TEST_ASSERT_VAL("wrong config", 0 == evsel->attr.config); TEST_ASSERT_VAL("wrong bp_type", HW_BREAKPOINT_W == evsel->attr.bp_type); TEST_ASSERT_VAL("wrong bp_len", HW_BREAKPOINT_LEN_4 == evsel->attr.bp_len); return 0; } static struct test__event_st { const char *name; __u32 type; int (*check)(struct perf_evlist *evlist); } test__events[] = { { .name = "syscalls:sys_enter_open", .check = test__checkevent_tracepoint, }, { .name = "syscalls:*", .check = test__checkevent_tracepoint_multi, }, { .name = "r1", .check = test__checkevent_raw, }, { .name = "1:1", .check = test__checkevent_numeric, }, { .name = "instructions", .check = test__checkevent_symbolic_name, }, { .name = "faults", .check = test__checkevent_symbolic_alias, }, { .name = "L1-dcache-load-miss", .check = test__checkevent_genhw, }, { .name = "mem:0", .check = test__checkevent_breakpoint, }, { .name = "mem:0:x", .check = test__checkevent_breakpoint_x, }, { .name = "mem:0:r", .check = test__checkevent_breakpoint_r, }, { .name = "mem:0:w", .check = test__checkevent_breakpoint_w, }, }; #define TEST__EVENTS_CNT (sizeof(test__events) / sizeof(struct test__event_st)) static int test__parse_events(void) { struct perf_evlist *evlist; u_int i; int ret = 0; for (i = 0; i < TEST__EVENTS_CNT; i++) { struct test__event_st *e = &test__events[i]; evlist = perf_evlist__new(NULL, NULL); if (evlist == NULL) break; ret = parse_events(evlist, e->name, 0); if (ret) { pr_debug("failed to parse event '%s', err %d\n", e->name, ret); break; } ret = e->check(evlist); if (ret) break; perf_evlist__delete(evlist); } return ret; } static struct test { const char *desc; int (*func)(void); } tests[] = { { .desc = "vmlinux symtab matches kallsyms", .func = test__vmlinux_matches_kallsyms, }, { .desc = "detect open syscall event", .func = test__open_syscall_event, }, { .desc = "detect open syscall event on all cpus", .func = test__open_syscall_event_on_all_cpus, }, { .desc = "read samples using the mmap interface", .func = test__basic_mmap, }, { .desc = "parse events tests", .func = test__parse_events, }, { .func = NULL, }, }; static int __cmd_test(void) { int i = 0; page_size = sysconf(_SC_PAGE_SIZE); while (tests[i].func) { int err; pr_info("%2d: %s:", i + 1, tests[i].desc); pr_debug("\n--- start ---\n"); err = tests[i].func(); pr_debug("---- end ----\n%s:", tests[i].desc); pr_info(" %s\n", err ? "FAILED!\n" : "Ok"); ++i; } return 0; } static const char * const test_usage[] = { "perf test [<options>]", NULL, }; static const struct option test_options[] = { OPT_INTEGER('v', "verbose", &verbose, "be more verbose (show symbol address, etc)"), OPT_END() }; int cmd_test(int argc, const char **argv, const char *prefix __used) { argc = parse_options(argc, argv, test_options, test_usage, 0); if (argc) usage_with_options(test_usage, test_options); symbol_conf.priv_size = sizeof(int); symbol_conf.sort_by_name = true; symbol_conf.try_vmlinux_path = true; if (symbol__init() < 0) return -1; setup_pager(); return __cmd_test(); }
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
You can’t perform that action at this time.