123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326 |
- /*
- * Copyright (c) 2022, David Tuin <davidot@serenityos.org>
- *
- * SPDX-License-Identifier: BSD-2-Clause
- */
- #include <AK/ByteString.h>
- #include <AK/Format.h>
- #include <AK/HashMap.h>
- #include <AK/JsonObject.h>
- #include <AK/JsonParser.h>
- #include <AK/LexicalPath.h>
- #include <AK/QuickSort.h>
- #include <AK/Vector.h>
- #include <LibCore/ArgsParser.h>
- #include <LibCore/Command.h>
- #include <LibCore/File.h>
- #include <LibFileSystem/FileSystem.h>
- #include <LibMain/Main.h>
- #include <LibTest/TestRunnerUtil.h>
- enum class TestResult {
- Passed,
- Failed,
- Skipped,
- MetadataError,
- HarnessError,
- TimeoutError,
- ProcessError,
- RunnerFailure,
- TodoError,
- };
- static StringView name_for_result(TestResult result)
- {
- switch (result) {
- case TestResult::Passed:
- return "PASSED"sv;
- case TestResult::Failed:
- return "FAILED"sv;
- case TestResult::Skipped:
- return "SKIPPED"sv;
- case TestResult::MetadataError:
- return "METADATA_ERROR"sv;
- case TestResult::HarnessError:
- return "HARNESS_ERROR"sv;
- case TestResult::TimeoutError:
- return "TIMEOUT_ERROR"sv;
- case TestResult::ProcessError:
- return "PROCESS_ERROR"sv;
- case TestResult::RunnerFailure:
- return "RUNNER_EXCEPTION"sv;
- case TestResult::TodoError:
- return "TODO_ERROR"sv;
- }
- VERIFY_NOT_REACHED();
- return ""sv;
- }
- static TestResult result_from_string(StringView string_view)
- {
- if (string_view == "passed"sv)
- return TestResult::Passed;
- if (string_view == "failed"sv)
- return TestResult::Failed;
- if (string_view == "skipped"sv)
- return TestResult::Skipped;
- if (string_view == "metadata_error"sv)
- return TestResult::MetadataError;
- if (string_view == "harness_error"sv)
- return TestResult::HarnessError;
- if (string_view == "timeout"sv)
- return TestResult::TimeoutError;
- if (string_view == "process_error"sv || string_view == "assert_fail"sv)
- return TestResult::ProcessError;
- if (string_view == "todo_error"sv)
- return TestResult::TodoError;
- return TestResult::RunnerFailure;
- }
- static StringView emoji_for_result(TestResult result)
- {
- switch (result) {
- case TestResult::Passed:
- return "✅"sv;
- case TestResult::Failed:
- return "❌"sv;
- case TestResult::Skipped:
- return "⚠"sv;
- case TestResult::MetadataError:
- return "📄"sv;
- case TestResult::HarnessError:
- return "⚙"sv;
- case TestResult::TimeoutError:
- return "💀"sv;
- case TestResult::ProcessError:
- return "💥"sv;
- case TestResult::RunnerFailure:
- return "🐍"sv;
- case TestResult::TodoError:
- return "📝"sv;
- }
- VERIFY_NOT_REACHED();
- return ""sv;
- }
- static constexpr StringView total_test_emoji = "🧪"sv;
- static ErrorOr<HashMap<size_t, TestResult>> run_test_files(Span<ByteString> files, size_t offset, StringView command, char const* const arguments[])
- {
- HashMap<size_t, TestResult> results {};
- TRY(results.try_ensure_capacity(files.size()));
- size_t test_index = 0;
- auto fail_all_after = [&] {
- for (; test_index < files.size(); ++test_index)
- results.set(offset + test_index, TestResult::RunnerFailure);
- };
- while (test_index < files.size()) {
- auto runner_process_or_error = Core::Command::create(command, arguments);
- if (runner_process_or_error.is_error()) {
- fail_all_after();
- return results;
- }
- auto& runner_process = runner_process_or_error.value();
- if (auto maybe_error = runner_process->write_lines(files.slice(test_index)); maybe_error.is_error()) {
- warnln("Runner process failed writing writing file input: {}", maybe_error.error());
- fail_all_after();
- return results;
- }
- auto output_or_error = runner_process->read_all();
- ByteString output;
- if (output_or_error.is_error())
- warnln("Got error: {} while reading runner output", output_or_error.error());
- else
- output = ByteString(output_or_error.release_value().standard_error.bytes(), Chomp);
- auto status_or_error = runner_process->status();
- bool failed = false;
- if (!status_or_error.is_error()) {
- VERIFY(status_or_error.value() != Core::Command::ProcessResult::Running);
- failed = status_or_error.value() != Core::Command::ProcessResult::DoneWithZeroExitCode;
- }
- for (StringView line : output.split_view('\n')) {
- if (!line.starts_with("RESULT "sv))
- break;
- auto test_for_line = test_index;
- ++test_index;
- if (test_for_line >= files.size())
- break;
- line = line.substring_view(7).trim("\n\0 "sv);
- JsonParser parser { line };
- TestResult result = TestResult::RunnerFailure;
- auto result_object_or_error = parser.parse();
- if (!result_object_or_error.is_error() && result_object_or_error.value().is_object()) {
- auto& result_object = result_object_or_error.value().as_object();
- if (auto result_string = result_object.get_byte_string("result"sv); result_string.has_value()) {
- auto const& view = result_string.value();
- // Timeout and assert fail already are the result of the stopping test
- if (view == "timeout"sv || view == "assert_fail"sv) {
- failed = false;
- }
- result = result_from_string(view);
- }
- }
- results.set(test_for_line + offset, result);
- }
- if (failed) {
- TestResult result = TestResult::ProcessError;
- if (!status_or_error.is_error() && status_or_error.value() == Core::Command::ProcessResult::FailedFromTimeout) {
- result = TestResult::TimeoutError;
- }
- // assume the last test failed, if by SIGALRM signal it's a timeout
- results.set(test_index + offset, result);
- ++test_index;
- }
- }
- return results;
- }
- void write_per_file(HashMap<size_t, TestResult> const& result_map, Vector<ByteString> const& paths, StringView per_file_name, double time_taken_in_ms);
- ErrorOr<int> serenity_main(Main::Arguments arguments)
- {
- size_t batch_size = 50;
- StringView per_file_location;
- StringView pass_through_parameters;
- StringView runner_command = "test262-runner"sv;
- StringView test_directory;
- bool dont_print_progress = false;
- bool dont_disable_core_dump = false;
- Core::ArgsParser args_parser;
- args_parser.add_positional_argument(test_directory, "Directory to search for tests", "tests");
- args_parser.add_option(per_file_location, "Output a per-file containing all results", "per-file", 'o', "filename");
- args_parser.add_option(batch_size, "Size of batches send to runner at once", "batch-size", 'b', "batch size");
- args_parser.add_option(runner_command, "Command to run", "runner-command", 'r', "command");
- args_parser.add_option(pass_through_parameters, "Parameters to pass through to the runner, will split on spaces", "pass-through", 'p', "parameters");
- args_parser.add_option(dont_print_progress, "Hide progress information", "quiet", 'q');
- args_parser.add_option(dont_disable_core_dump, "Enabled core dumps for runner (i.e. don't pass --disable-core-dump)", "enable-core-dumps");
- args_parser.parse(arguments);
- // Normalize the path to ensure filenames are consistent
- Vector<ByteString> paths;
- if (!FileSystem::is_directory(test_directory)) {
- paths.append(test_directory);
- } else {
- Test::iterate_directory_recursively(LexicalPath::canonicalized_path(test_directory), [&](ByteString const& file_path) {
- if (file_path.contains("_FIXTURE"sv))
- return;
- // FIXME: Add ignored file set
- paths.append(file_path);
- });
- quick_sort(paths);
- }
- outln("Found {} tests", paths.size());
- auto parameters = pass_through_parameters.split_view(' ');
- Vector<ByteString> args;
- args.ensure_capacity(parameters.size() + 2);
- args.append(runner_command);
- if (!dont_disable_core_dump)
- args.append("--disable-core-dump"sv);
- for (auto parameter : parameters)
- args.append(parameter);
- Vector<char const*> raw_args;
- raw_args.ensure_capacity(args.size() + 1);
- for (auto& arg : args)
- raw_args.append(arg.characters());
- raw_args.append(nullptr);
- dbgln("test262 runner command: {}", args);
- HashMap<size_t, TestResult> results;
- Array<size_t, 9> result_counts {};
- static_assert(result_counts.size() == static_cast<size_t>(TestResult::TodoError) + 1u);
- size_t index = 0;
- double start_time = Test::get_time_in_ms();
- auto print_progress = [&] {
- if (!dont_print_progress) {
- warn("\033]9;{};{};\033\\", index, paths.size());
- double percentage_done = (100. * index) / paths.size();
- warn("{:04.2f}% {:3.1f}s ", percentage_done, (Test::get_time_in_ms() - start_time) / 1000.);
- for (size_t i = 0; i < result_counts.size(); ++i) {
- auto result_type = static_cast<TestResult>(i);
- warn("{} {} ", emoji_for_result(result_type), result_counts[i]);
- }
- warn("\r");
- }
- };
- while (index < paths.size()) {
- print_progress();
- auto this_batch_size = min(batch_size, paths.size() - index);
- auto batch_results = TRY(run_test_files(paths.span().slice(index, this_batch_size), index, args[0], raw_args.data()));
- TRY(results.try_ensure_capacity(results.size() + batch_results.size()));
- for (auto& [key, value] : batch_results) {
- results.set(key, value);
- ++result_counts[static_cast<size_t>(value)];
- }
- index += this_batch_size;
- }
- double time_taken_in_ms = Test::get_time_in_ms() - start_time;
- print_progress();
- if (!dont_print_progress)
- warn("\n\033]9;-1;\033\\");
- outln("Took {} seconds", time_taken_in_ms / 1000.);
- outln("{}: {}", total_test_emoji, paths.size());
- for (size_t i = 0; i < result_counts.size(); ++i) {
- auto result_type = static_cast<TestResult>(i);
- outln("{}: {} ({:3.2f}%)", emoji_for_result(result_type), result_counts[i], 100. * static_cast<double>(result_counts[i]) / paths.size());
- }
- if (!per_file_location.is_empty())
- write_per_file(results, paths, per_file_location, time_taken_in_ms);
- return 0;
- }
- void write_per_file(HashMap<size_t, TestResult> const& result_map, Vector<ByteString> const& paths, StringView per_file_name, double time_taken_in_ms)
- {
- auto file_or_error = Core::File::open(per_file_name, Core::File::OpenMode::Write);
- if (file_or_error.is_error()) {
- warnln("Failed to open per file for writing at {}: {}", per_file_name, file_or_error.error().string_literal());
- return;
- }
- auto& file = file_or_error.value();
- JsonObject result_object;
- for (auto& [test, value] : result_map)
- result_object.set(paths[test], name_for_result(value));
- JsonObject complete_results {};
- complete_results.set("duration", time_taken_in_ms / 1000.);
- complete_results.set("results", result_object);
- if (file->write_until_depleted(complete_results.to_byte_string()).is_error())
- warnln("Failed to write per-file");
- file->close();
- }
|