When adapting the test to the kselftest framework, a few printf() calls indicating test progress were not updated.
Fix this by replacing these printf() calls by ksft_print_msg() calls.
Fixes: ce7d101750ff8450 ("selftests: timers: clocksource-switch: adapt to kselftest framework") Signed-off-by: Geert Uytterhoeven geert+renesas@glider.be --- When just running the test, the output looks like:
# Validating clocksource arch_sys_counter TAP version 13 1..12 ok 1 CLOCK_REALTIME ... # Validating clocksource ffca0000.timer TAP version 13 1..12 ok 1 CLOCK_REALTIME ...
When redirecting the test output to a file, the progress prints are not interspersed with the test output, but collated at the end:
TAP version 13 1..12 ok 1 CLOCK_REALTIME ... TAP version 13 1..12 ok 1 CLOCK_REALTIME ... # Totals: pass:6 fail:0 xfail:0 xpass:0 skip:6 error:0 # Validating clocksource arch_sys_counter # Validating clocksource ffca0000.timer ...
This makes it hard to match the test results with the timer under test. Is there a way to fix this? The test does use fork().
Thanks! --- tools/testing/selftests/timers/clocksource-switch.c | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-)
diff --git a/tools/testing/selftests/timers/clocksource-switch.c b/tools/testing/selftests/timers/clocksource-switch.c index c5264594064c8516..83faa4e354e389c2 100644 --- a/tools/testing/selftests/timers/clocksource-switch.c +++ b/tools/testing/selftests/timers/clocksource-switch.c @@ -156,8 +156,8 @@ int main(int argc, char **argv) /* Check everything is sane before we start switching asynchronously */ if (do_sanity_check) { for (i = 0; i < count; i++) { - printf("Validating clocksource %s\n", - clocksource_list[i]); + ksft_print_msg("Validating clocksource %s\n", + clocksource_list[i]); if (change_clocksource(clocksource_list[i])) { status = -1; goto out; @@ -169,7 +169,7 @@ int main(int argc, char **argv) } }
- printf("Running Asynchronous Switching Tests...\n"); + ksft_print_msg("Running Asynchronous Switching Tests...\n"); pid = fork(); if (!pid) return run_tests(runtime);