Merge branch 'BPF test_progs tests improvement'
Mykola Lysenko says: ==================== First patch reduces the sample_freq to 1000 to ensure test will work even when kernel.perf_event_max_sample_rate was reduced to 1000. Patches for send_signal and find_vma tune the test implementation to make sure needed thread is scheduled. Also, both tests will finish as soon as possible after the test condition is met. ==================== Signed-off-by: Andrii Nakryiko <andrii@kernel.org>
This commit is contained in:
commit
3399dd9f37
@ -199,7 +199,7 @@ static void pe_subtest(struct test_bpf_cookie *skel)
|
||||
attr.type = PERF_TYPE_SOFTWARE;
|
||||
attr.config = PERF_COUNT_SW_CPU_CLOCK;
|
||||
attr.freq = 1;
|
||||
attr.sample_freq = 4000;
|
||||
attr.sample_freq = 1000;
|
||||
pfd = syscall(__NR_perf_event_open, &attr, -1, 0, -1, PERF_FLAG_FD_CLOEXEC);
|
||||
if (!ASSERT_GE(pfd, 0, "perf_fd"))
|
||||
goto cleanup;
|
||||
|
@ -7,12 +7,14 @@
|
||||
#include "find_vma_fail1.skel.h"
|
||||
#include "find_vma_fail2.skel.h"
|
||||
|
||||
static void test_and_reset_skel(struct find_vma *skel, int expected_find_zero_ret)
|
||||
static void test_and_reset_skel(struct find_vma *skel, int expected_find_zero_ret, bool need_test)
|
||||
{
|
||||
ASSERT_EQ(skel->bss->found_vm_exec, 1, "found_vm_exec");
|
||||
ASSERT_EQ(skel->data->find_addr_ret, 0, "find_addr_ret");
|
||||
ASSERT_EQ(skel->data->find_zero_ret, expected_find_zero_ret, "find_zero_ret");
|
||||
ASSERT_OK_PTR(strstr(skel->bss->d_iname, "test_progs"), "find_test_progs");
|
||||
if (need_test) {
|
||||
ASSERT_EQ(skel->bss->found_vm_exec, 1, "found_vm_exec");
|
||||
ASSERT_EQ(skel->data->find_addr_ret, 0, "find_addr_ret");
|
||||
ASSERT_EQ(skel->data->find_zero_ret, expected_find_zero_ret, "find_zero_ret");
|
||||
ASSERT_OK_PTR(strstr(skel->bss->d_iname, "test_progs"), "find_test_progs");
|
||||
}
|
||||
|
||||
skel->bss->found_vm_exec = 0;
|
||||
skel->data->find_addr_ret = -1;
|
||||
@ -30,17 +32,26 @@ static int open_pe(void)
|
||||
attr.type = PERF_TYPE_HARDWARE;
|
||||
attr.config = PERF_COUNT_HW_CPU_CYCLES;
|
||||
attr.freq = 1;
|
||||
attr.sample_freq = 4000;
|
||||
attr.sample_freq = 1000;
|
||||
pfd = syscall(__NR_perf_event_open, &attr, 0, -1, -1, PERF_FLAG_FD_CLOEXEC);
|
||||
|
||||
return pfd >= 0 ? pfd : -errno;
|
||||
}
|
||||
|
||||
static bool find_vma_pe_condition(struct find_vma *skel)
|
||||
{
|
||||
return skel->bss->found_vm_exec == 0 ||
|
||||
skel->data->find_addr_ret != 0 ||
|
||||
skel->data->find_zero_ret == -1 ||
|
||||
strcmp(skel->bss->d_iname, "test_progs") != 0;
|
||||
}
|
||||
|
||||
static void test_find_vma_pe(struct find_vma *skel)
|
||||
{
|
||||
struct bpf_link *link = NULL;
|
||||
volatile int j = 0;
|
||||
int pfd, i;
|
||||
const int one_bn = 1000000000;
|
||||
|
||||
pfd = open_pe();
|
||||
if (pfd < 0) {
|
||||
@ -57,10 +68,10 @@ static void test_find_vma_pe(struct find_vma *skel)
|
||||
if (!ASSERT_OK_PTR(link, "attach_perf_event"))
|
||||
goto cleanup;
|
||||
|
||||
for (i = 0; i < 1000000; ++i)
|
||||
for (i = 0; i < one_bn && find_vma_pe_condition(skel); ++i)
|
||||
++j;
|
||||
|
||||
test_and_reset_skel(skel, -EBUSY /* in nmi, irq_work is busy */);
|
||||
test_and_reset_skel(skel, -EBUSY /* in nmi, irq_work is busy */, i == one_bn);
|
||||
cleanup:
|
||||
bpf_link__destroy(link);
|
||||
close(pfd);
|
||||
@ -75,7 +86,7 @@ static void test_find_vma_kprobe(struct find_vma *skel)
|
||||
return;
|
||||
|
||||
getpgid(skel->bss->target_pid);
|
||||
test_and_reset_skel(skel, -ENOENT /* could not find vma for ptr 0 */);
|
||||
test_and_reset_skel(skel, -ENOENT /* could not find vma for ptr 0 */, true);
|
||||
}
|
||||
|
||||
static void test_illegal_write_vma(void)
|
||||
@ -108,7 +119,6 @@ void serial_test_find_vma(void)
|
||||
skel->bss->addr = (__u64)(uintptr_t)test_find_vma_pe;
|
||||
|
||||
test_find_vma_pe(skel);
|
||||
usleep(100000); /* allow the irq_work to finish */
|
||||
test_find_vma_kprobe(skel);
|
||||
|
||||
find_vma__destroy(skel);
|
||||
|
@ -110,7 +110,7 @@ static void test_perf_branches_hw(void)
|
||||
attr.type = PERF_TYPE_HARDWARE;
|
||||
attr.config = PERF_COUNT_HW_CPU_CYCLES;
|
||||
attr.freq = 1;
|
||||
attr.sample_freq = 4000;
|
||||
attr.sample_freq = 1000;
|
||||
attr.sample_type = PERF_SAMPLE_BRANCH_STACK;
|
||||
attr.branch_sample_type = PERF_SAMPLE_BRANCH_USER | PERF_SAMPLE_BRANCH_ANY;
|
||||
pfd = syscall(__NR_perf_event_open, &attr, -1, 0, -1, PERF_FLAG_FD_CLOEXEC);
|
||||
@ -151,7 +151,7 @@ static void test_perf_branches_no_hw(void)
|
||||
attr.type = PERF_TYPE_SOFTWARE;
|
||||
attr.config = PERF_COUNT_SW_CPU_CLOCK;
|
||||
attr.freq = 1;
|
||||
attr.sample_freq = 4000;
|
||||
attr.sample_freq = 1000;
|
||||
pfd = syscall(__NR_perf_event_open, &attr, -1, 0, -1, PERF_FLAG_FD_CLOEXEC);
|
||||
if (CHECK(pfd < 0, "perf_event_open", "err %d\n", pfd))
|
||||
return;
|
||||
|
@ -39,7 +39,7 @@ void serial_test_perf_link(void)
|
||||
attr.type = PERF_TYPE_SOFTWARE;
|
||||
attr.config = PERF_COUNT_SW_CPU_CLOCK;
|
||||
attr.freq = 1;
|
||||
attr.sample_freq = 4000;
|
||||
attr.sample_freq = 1000;
|
||||
pfd = syscall(__NR_perf_event_open, &attr, -1, 0, -1, PERF_FLAG_FD_CLOEXEC);
|
||||
if (!ASSERT_GE(pfd, 0, "perf_fd"))
|
||||
goto cleanup;
|
||||
|
@ -4,11 +4,11 @@
|
||||
#include <sys/resource.h>
|
||||
#include "test_send_signal_kern.skel.h"
|
||||
|
||||
int sigusr1_received = 0;
|
||||
static int sigusr1_received;
|
||||
|
||||
static void sigusr1_handler(int signum)
|
||||
{
|
||||
sigusr1_received++;
|
||||
sigusr1_received = 1;
|
||||
}
|
||||
|
||||
static void test_send_signal_common(struct perf_event_attr *attr,
|
||||
@ -40,9 +40,10 @@ static void test_send_signal_common(struct perf_event_attr *attr,
|
||||
|
||||
if (pid == 0) {
|
||||
int old_prio;
|
||||
volatile int j = 0;
|
||||
|
||||
/* install signal handler and notify parent */
|
||||
signal(SIGUSR1, sigusr1_handler);
|
||||
ASSERT_NEQ(signal(SIGUSR1, sigusr1_handler), SIG_ERR, "signal");
|
||||
|
||||
close(pipe_c2p[0]); /* close read */
|
||||
close(pipe_p2c[1]); /* close write */
|
||||
@ -63,9 +64,11 @@ static void test_send_signal_common(struct perf_event_attr *attr,
|
||||
ASSERT_EQ(read(pipe_p2c[0], buf, 1), 1, "pipe_read");
|
||||
|
||||
/* wait a little for signal handler */
|
||||
sleep(1);
|
||||
for (int i = 0; i < 100000000 && !sigusr1_received; i++)
|
||||
j /= i + 1;
|
||||
|
||||
buf[0] = sigusr1_received ? '2' : '0';
|
||||
ASSERT_EQ(sigusr1_received, 1, "sigusr1_received");
|
||||
ASSERT_EQ(write(pipe_c2p[1], buf, 1), 1, "pipe_write");
|
||||
|
||||
/* wait for parent notification and exit */
|
||||
@ -93,7 +96,7 @@ static void test_send_signal_common(struct perf_event_attr *attr,
|
||||
goto destroy_skel;
|
||||
}
|
||||
} else {
|
||||
pmu_fd = syscall(__NR_perf_event_open, attr, pid, -1,
|
||||
pmu_fd = syscall(__NR_perf_event_open, attr, pid, -1 /* cpu */,
|
||||
-1 /* group id */, 0 /* flags */);
|
||||
if (!ASSERT_GE(pmu_fd, 0, "perf_event_open")) {
|
||||
err = -1;
|
||||
@ -110,9 +113,9 @@ static void test_send_signal_common(struct perf_event_attr *attr,
|
||||
ASSERT_EQ(read(pipe_c2p[0], buf, 1), 1, "pipe_read");
|
||||
|
||||
/* trigger the bpf send_signal */
|
||||
skel->bss->pid = pid;
|
||||
skel->bss->sig = SIGUSR1;
|
||||
skel->bss->signal_thread = signal_thread;
|
||||
skel->bss->sig = SIGUSR1;
|
||||
skel->bss->pid = pid;
|
||||
|
||||
/* notify child that bpf program can send_signal now */
|
||||
ASSERT_EQ(write(pipe_p2c[1], buf, 1), 1, "pipe_write");
|
||||
|
@ -10,7 +10,7 @@ static __always_inline int bpf_send_signal_test(void *ctx)
|
||||
{
|
||||
int ret;
|
||||
|
||||
if (status != 0 || sig == 0 || pid == 0)
|
||||
if (status != 0 || pid == 0)
|
||||
return 0;
|
||||
|
||||
if ((bpf_get_current_pid_tgid() >> 32) == pid) {
|
||||
|
Loading…
Reference in New Issue
Block a user