mirror of
https://github.com/LadybirdBrowser/ladybird.git
synced 2024-11-22 07:30:19 +00:00
LibJS+LibTest: Move out the test-js test runner into LibTest
This commit is contained in:
parent
b6e5c76427
commit
f137c1bfaa
Notes:
sideshowbarker
2024-07-18 17:49:50 +09:00
Author: https://github.com/alimpfard Commit: https://github.com/SerenityOS/serenity/commit/f137c1bfaac Pull-request: https://github.com/SerenityOS/serenity/pull/7255 Reviewed-by: https://github.com/linusg
6 changed files with 776 additions and 733 deletions
|
@ -70,10 +70,14 @@ function(serenity_bin target_name)
|
||||||
endfunction()
|
endfunction()
|
||||||
|
|
||||||
function(serenity_test test_src sub_dir)
|
function(serenity_test test_src sub_dir)
|
||||||
cmake_parse_arguments(SERENITY_TEST "CUSTOM_MAIN" "" "LIBS" ${ARGN})
|
cmake_parse_arguments(SERENITY_TEST "MAIN_ALREADY_DEFINED" "CUSTOM_MAIN" "LIBS" ${ARGN})
|
||||||
set(TEST_SOURCES ${test_src})
|
set(TEST_SOURCES ${test_src})
|
||||||
if (NOT ${SERENITY_TEST_CUSTOM_MAIN})
|
if ("${SERENITY_TEST_CUSTOM_MAIN}" STREQUAL "")
|
||||||
list(APPEND TEST_SOURCES "${CMAKE_SOURCE_DIR}/Userland/Libraries/LibTest/TestMain.cpp")
|
set(SERENITY_TEST_CUSTOM_MAIN
|
||||||
|
"${CMAKE_SOURCE_DIR}/Userland/Libraries/LibTest/TestMain.cpp")
|
||||||
|
endif()
|
||||||
|
if (NOT ${SERENITY_TEST_MAIN_ALREADY_DEFINED})
|
||||||
|
list(PREPEND TEST_SOURCES "${SERENITY_TEST_CUSTOM_MAIN}")
|
||||||
endif()
|
endif()
|
||||||
get_filename_component(test_name ${test_src} NAME_WE)
|
get_filename_component(test_name ${test_src} NAME_WE)
|
||||||
add_executable(${test_name} ${TEST_SOURCES})
|
add_executable(${test_name} ${TEST_SOURCES})
|
||||||
|
@ -84,6 +88,19 @@ function(serenity_test test_src sub_dir)
|
||||||
install(TARGETS ${test_name} RUNTIME DESTINATION usr/Tests/${sub_dir})
|
install(TARGETS ${test_name} RUNTIME DESTINATION usr/Tests/${sub_dir})
|
||||||
endfunction()
|
endfunction()
|
||||||
|
|
||||||
|
|
||||||
|
function(serenity_testjs_test test_src sub_dir)
|
||||||
|
cmake_parse_arguments(SERENITY_TEST "" "CUSTOM_MAIN" "LIBS" ${ARGN})
|
||||||
|
if ("${SERENITY_TEST_CUSTOM_MAIN}" STREQUAL "")
|
||||||
|
set(SERENITY_TEST_CUSTOM_MAIN
|
||||||
|
"${CMAKE_SOURCE_DIR}/Userland/Libraries/LibTest/JavaScriptTestRunnerMain.cpp")
|
||||||
|
endif()
|
||||||
|
list(APPEND SERENITY_TEST_LIBS LibJS LibCore)
|
||||||
|
serenity_test(${test_src} ${sub_dir}
|
||||||
|
CUSTOM_MAIN "${SERENITY_TEST_CUSTOM_MAIN}"
|
||||||
|
LIBS ${SERENITY_TEST_LIBS})
|
||||||
|
endfunction()
|
||||||
|
|
||||||
function(serenity_app target_name)
|
function(serenity_app target_name)
|
||||||
cmake_parse_arguments(SERENITY_APP "" "ICON" "" ${ARGN})
|
cmake_parse_arguments(SERENITY_APP "" "ICON" "" ${ARGN})
|
||||||
|
|
||||||
|
|
|
@ -132,7 +132,9 @@ if (BUILD_LAGOM)
|
||||||
set_target_properties(ntpquery_lagom PROPERTIES OUTPUT_NAME ntpquery)
|
set_target_properties(ntpquery_lagom PROPERTIES OUTPUT_NAME ntpquery)
|
||||||
target_link_libraries(ntpquery_lagom Lagom)
|
target_link_libraries(ntpquery_lagom Lagom)
|
||||||
|
|
||||||
add_executable(test-js_lagom ../../Tests/LibJS/test-js.cpp)
|
add_executable(test-js_lagom
|
||||||
|
../../Tests/LibJS/test-js.cpp
|
||||||
|
../../Userland/Libraries/LibTest/JavaScriptTestRunnerMain.cpp)
|
||||||
set_target_properties(test-js_lagom PROPERTIES OUTPUT_NAME test-js)
|
set_target_properties(test-js_lagom PROPERTIES OUTPUT_NAME test-js)
|
||||||
target_link_libraries(test-js_lagom Lagom)
|
target_link_libraries(test-js_lagom Lagom)
|
||||||
target_link_libraries(test-js_lagom stdc++)
|
target_link_libraries(test-js_lagom stdc++)
|
||||||
|
|
|
@ -1,3 +1,2 @@
|
||||||
add_executable(test-js test-js.cpp)
|
serenity_testjs_test(test-js.cpp test-js)
|
||||||
target_link_libraries(test-js LibJS LibLine LibCore)
|
|
||||||
install(TARGETS test-js RUNTIME DESTINATION bin)
|
install(TARGETS test-js RUNTIME DESTINATION bin)
|
||||||
|
|
|
@ -5,135 +5,16 @@
|
||||||
* SPDX-License-Identifier: BSD-2-Clause
|
* SPDX-License-Identifier: BSD-2-Clause
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#include <AK/ByteBuffer.h>
|
#include <LibTest/JavaScriptTestRunner.h>
|
||||||
#include <AK/JsonObject.h>
|
|
||||||
#include <AK/JsonValue.h>
|
|
||||||
#include <AK/LexicalPath.h>
|
|
||||||
#include <AK/QuickSort.h>
|
|
||||||
#include <LibCore/ArgsParser.h>
|
|
||||||
#include <LibCore/DirIterator.h>
|
|
||||||
#include <LibCore/File.h>
|
|
||||||
#include <LibJS/Interpreter.h>
|
|
||||||
#include <LibJS/Lexer.h>
|
|
||||||
#include <LibJS/Parser.h>
|
|
||||||
#include <LibJS/Runtime/Array.h>
|
|
||||||
#include <LibJS/Runtime/GlobalObject.h>
|
|
||||||
#include <LibJS/Runtime/JSONObject.h>
|
|
||||||
#include <LibTest/Results.h>
|
|
||||||
#include <signal.h>
|
|
||||||
#include <stdlib.h>
|
|
||||||
#include <sys/time.h>
|
|
||||||
#include <unistd.h>
|
|
||||||
|
|
||||||
#define TOP_LEVEL_TEST_NAME "__$$TOP_LEVEL$$__"
|
TEST_ROOT("Userland/Libraries/LibJS/Tests");
|
||||||
|
|
||||||
RefPtr<JS::VM> vm;
|
TESTJS_GLOBAL_FUNCTION(is_strict_mode, isStrictMode, 0)
|
||||||
|
|
||||||
static bool collect_on_every_allocation = false;
|
|
||||||
static String currently_running_test;
|
|
||||||
|
|
||||||
struct ParserError {
|
|
||||||
JS::Parser::Error error;
|
|
||||||
String hint;
|
|
||||||
};
|
|
||||||
|
|
||||||
struct JSFileResult {
|
|
||||||
String name;
|
|
||||||
Optional<ParserError> error {};
|
|
||||||
double time_taken { 0 };
|
|
||||||
// A failed test takes precedence over a skipped test, which both have
|
|
||||||
// precedence over a passed test
|
|
||||||
Test::Result most_severe_test_result { Test::Result::Pass };
|
|
||||||
Vector<Test::Suite> suites {};
|
|
||||||
Vector<String> logged_messages {};
|
|
||||||
};
|
|
||||||
|
|
||||||
class TestRunnerGlobalObject final : public JS::GlobalObject {
|
|
||||||
JS_OBJECT(TestRunnerGlobalObject, JS::GlobalObject);
|
|
||||||
|
|
||||||
public:
|
|
||||||
TestRunnerGlobalObject();
|
|
||||||
virtual ~TestRunnerGlobalObject() override;
|
|
||||||
|
|
||||||
virtual void initialize_global_object() override;
|
|
||||||
|
|
||||||
private:
|
|
||||||
JS_DECLARE_NATIVE_FUNCTION(is_strict_mode);
|
|
||||||
JS_DECLARE_NATIVE_FUNCTION(can_parse_source);
|
|
||||||
JS_DECLARE_NATIVE_FUNCTION(run_queued_promise_jobs);
|
|
||||||
};
|
|
||||||
|
|
||||||
class TestRunner {
|
|
||||||
public:
|
|
||||||
static TestRunner* the()
|
|
||||||
{
|
|
||||||
return s_the;
|
|
||||||
}
|
|
||||||
|
|
||||||
TestRunner(String test_root, bool print_times, bool print_progress)
|
|
||||||
: m_test_root(move(test_root))
|
|
||||||
, m_print_times(print_times)
|
|
||||||
, m_print_progress(print_progress)
|
|
||||||
{
|
|
||||||
VERIFY(!s_the);
|
|
||||||
s_the = this;
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual ~TestRunner() { s_the = nullptr; };
|
|
||||||
|
|
||||||
Test::Counts run();
|
|
||||||
|
|
||||||
const Test::Counts& counts() const { return m_counts; }
|
|
||||||
|
|
||||||
bool is_printing_progress() const { return m_print_progress; }
|
|
||||||
|
|
||||||
protected:
|
|
||||||
static TestRunner* s_the;
|
|
||||||
|
|
||||||
virtual Vector<String> get_test_paths() const;
|
|
||||||
virtual JSFileResult run_file_test(const String& test_path);
|
|
||||||
void print_file_result(const JSFileResult& file_result) const;
|
|
||||||
void print_test_results() const;
|
|
||||||
|
|
||||||
String m_test_root;
|
|
||||||
bool m_print_times;
|
|
||||||
bool m_print_progress;
|
|
||||||
|
|
||||||
double m_total_elapsed_time_in_ms { 0 };
|
|
||||||
Test::Counts m_counts;
|
|
||||||
|
|
||||||
RefPtr<JS::Program> m_test_program;
|
|
||||||
};
|
|
||||||
|
|
||||||
TestRunner* TestRunner::s_the = nullptr;
|
|
||||||
|
|
||||||
TestRunnerGlobalObject::TestRunnerGlobalObject()
|
|
||||||
{
|
|
||||||
}
|
|
||||||
|
|
||||||
TestRunnerGlobalObject::~TestRunnerGlobalObject()
|
|
||||||
{
|
|
||||||
}
|
|
||||||
|
|
||||||
void TestRunnerGlobalObject::initialize_global_object()
|
|
||||||
{
|
|
||||||
Base::initialize_global_object();
|
|
||||||
static FlyString global_property_name { "global" };
|
|
||||||
static FlyString is_strict_mode_property_name { "isStrictMode" };
|
|
||||||
static FlyString can_parse_source_property_name { "canParseSource" };
|
|
||||||
static FlyString run_queued_promise_jobs_property_name { "runQueuedPromiseJobs" };
|
|
||||||
define_property(global_property_name, this, JS::Attribute::Enumerable);
|
|
||||||
define_native_function(is_strict_mode_property_name, is_strict_mode);
|
|
||||||
define_native_function(can_parse_source_property_name, can_parse_source);
|
|
||||||
define_native_function(run_queued_promise_jobs_property_name, run_queued_promise_jobs);
|
|
||||||
}
|
|
||||||
|
|
||||||
JS_DEFINE_NATIVE_FUNCTION(TestRunnerGlobalObject::is_strict_mode)
|
|
||||||
{
|
{
|
||||||
return JS::Value(vm.in_strict_mode());
|
return JS::Value(vm.in_strict_mode());
|
||||||
}
|
}
|
||||||
|
|
||||||
JS_DEFINE_NATIVE_FUNCTION(TestRunnerGlobalObject::can_parse_source)
|
TESTJS_GLOBAL_FUNCTION(can_parse_source, canParseSource)
|
||||||
{
|
{
|
||||||
auto source = vm.argument(0).to_string(global_object);
|
auto source = vm.argument(0).to_string(global_object);
|
||||||
if (vm.exception())
|
if (vm.exception())
|
||||||
|
@ -143,611 +24,8 @@ JS_DEFINE_NATIVE_FUNCTION(TestRunnerGlobalObject::can_parse_source)
|
||||||
return JS::Value(!parser.has_errors());
|
return JS::Value(!parser.has_errors());
|
||||||
}
|
}
|
||||||
|
|
||||||
JS_DEFINE_NATIVE_FUNCTION(TestRunnerGlobalObject::run_queued_promise_jobs)
|
TESTJS_GLOBAL_FUNCTION(run_queued_promise_jobs, runQueuedPromiseJobs)
|
||||||
{
|
{
|
||||||
vm.run_queued_promise_jobs();
|
vm.run_queued_promise_jobs();
|
||||||
return JS::js_undefined();
|
return JS::js_undefined();
|
||||||
}
|
}
|
||||||
|
|
||||||
static void cleanup_and_exit()
|
|
||||||
{
|
|
||||||
// Clear the taskbar progress.
|
|
||||||
if (TestRunner::the() && TestRunner::the()->is_printing_progress())
|
|
||||||
warn("\033]9;-1;\033\\");
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void handle_sigabrt(int)
|
|
||||||
{
|
|
||||||
dbgln("test-js: SIGABRT received, cleaning up.");
|
|
||||||
cleanup_and_exit();
|
|
||||||
}
|
|
||||||
|
|
||||||
static double get_time_in_ms()
|
|
||||||
{
|
|
||||||
struct timeval tv1;
|
|
||||||
auto return_code = gettimeofday(&tv1, nullptr);
|
|
||||||
VERIFY(return_code >= 0);
|
|
||||||
return static_cast<double>(tv1.tv_sec) * 1000.0 + static_cast<double>(tv1.tv_usec) / 1000.0;
|
|
||||||
}
|
|
||||||
|
|
||||||
template<typename Callback>
|
|
||||||
static void iterate_directory_recursively(const String& directory_path, Callback callback)
|
|
||||||
{
|
|
||||||
Core::DirIterator directory_iterator(directory_path, Core::DirIterator::Flags::SkipDots);
|
|
||||||
|
|
||||||
while (directory_iterator.has_next()) {
|
|
||||||
auto file_path = directory_iterator.next_full_path();
|
|
||||||
if (Core::File::is_directory(file_path)) {
|
|
||||||
iterate_directory_recursively(file_path, callback);
|
|
||||||
} else {
|
|
||||||
callback(move(file_path));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
Vector<String> TestRunner::get_test_paths() const
|
|
||||||
{
|
|
||||||
Vector<String> paths;
|
|
||||||
iterate_directory_recursively(m_test_root, [&](const String& file_path) {
|
|
||||||
if (!file_path.ends_with("test-common.js"))
|
|
||||||
paths.append(file_path);
|
|
||||||
});
|
|
||||||
quick_sort(paths);
|
|
||||||
return paths;
|
|
||||||
}
|
|
||||||
|
|
||||||
Test::Counts TestRunner::run()
|
|
||||||
{
|
|
||||||
size_t progress_counter = 0;
|
|
||||||
auto test_paths = get_test_paths();
|
|
||||||
for (auto& path : test_paths) {
|
|
||||||
++progress_counter;
|
|
||||||
print_file_result(run_file_test(path));
|
|
||||||
if (m_print_progress)
|
|
||||||
warn("\033]9;{};{};\033\\", progress_counter, test_paths.size());
|
|
||||||
}
|
|
||||||
|
|
||||||
if (m_print_progress)
|
|
||||||
warn("\033]9;-1;\033\\");
|
|
||||||
|
|
||||||
print_test_results();
|
|
||||||
|
|
||||||
return m_counts;
|
|
||||||
}
|
|
||||||
|
|
||||||
static Result<NonnullRefPtr<JS::Program>, ParserError> parse_file(const String& file_path)
|
|
||||||
{
|
|
||||||
auto file = Core::File::construct(file_path);
|
|
||||||
auto result = file->open(Core::OpenMode::ReadOnly);
|
|
||||||
if (!result) {
|
|
||||||
warnln("Failed to open the following file: \"{}\"", file_path);
|
|
||||||
cleanup_and_exit();
|
|
||||||
}
|
|
||||||
|
|
||||||
auto contents = file->read_all();
|
|
||||||
String test_file_string(reinterpret_cast<const char*>(contents.data()), contents.size());
|
|
||||||
file->close();
|
|
||||||
|
|
||||||
auto parser = JS::Parser(JS::Lexer(test_file_string));
|
|
||||||
auto program = parser.parse_program();
|
|
||||||
|
|
||||||
if (parser.has_errors()) {
|
|
||||||
auto error = parser.errors()[0];
|
|
||||||
return Result<NonnullRefPtr<JS::Program>, ParserError>(ParserError { error, error.source_location_hint(test_file_string) });
|
|
||||||
}
|
|
||||||
|
|
||||||
return Result<NonnullRefPtr<JS::Program>, ParserError>(program);
|
|
||||||
}
|
|
||||||
|
|
||||||
static Optional<JsonValue> get_test_results(JS::Interpreter& interpreter)
|
|
||||||
{
|
|
||||||
auto result = vm->get_variable("__TestResults__", interpreter.global_object());
|
|
||||||
auto json_string = JS::JSONObject::stringify_impl(interpreter.global_object(), result, JS::js_undefined(), JS::js_undefined());
|
|
||||||
|
|
||||||
auto json = JsonValue::from_string(json_string);
|
|
||||||
if (!json.has_value())
|
|
||||||
return {};
|
|
||||||
|
|
||||||
return json.value();
|
|
||||||
}
|
|
||||||
|
|
||||||
JSFileResult TestRunner::run_file_test(const String& test_path)
|
|
||||||
{
|
|
||||||
currently_running_test = test_path;
|
|
||||||
|
|
||||||
double start_time = get_time_in_ms();
|
|
||||||
auto interpreter = JS::Interpreter::create<TestRunnerGlobalObject>(*vm);
|
|
||||||
|
|
||||||
// FIXME: This is a hack while we're refactoring Interpreter/VM stuff.
|
|
||||||
JS::VM::InterpreterExecutionScope scope(*interpreter);
|
|
||||||
|
|
||||||
interpreter->heap().set_should_collect_on_every_allocation(collect_on_every_allocation);
|
|
||||||
|
|
||||||
if (!m_test_program) {
|
|
||||||
auto result = parse_file(String::formatted("{}/test-common.js", m_test_root));
|
|
||||||
if (result.is_error()) {
|
|
||||||
warnln("Unable to parse test-common.js");
|
|
||||||
warnln("{}", result.error().error.to_string());
|
|
||||||
warnln("{}", result.error().hint);
|
|
||||||
cleanup_and_exit();
|
|
||||||
}
|
|
||||||
m_test_program = result.value();
|
|
||||||
}
|
|
||||||
|
|
||||||
interpreter->run(interpreter->global_object(), *m_test_program);
|
|
||||||
|
|
||||||
auto file_program = parse_file(test_path);
|
|
||||||
if (file_program.is_error())
|
|
||||||
return { test_path, file_program.error() };
|
|
||||||
interpreter->run(interpreter->global_object(), *file_program.value());
|
|
||||||
|
|
||||||
auto test_json = get_test_results(*interpreter);
|
|
||||||
if (!test_json.has_value()) {
|
|
||||||
warnln("Received malformed JSON from test \"{}\"", test_path);
|
|
||||||
cleanup_and_exit();
|
|
||||||
}
|
|
||||||
|
|
||||||
JSFileResult file_result { test_path.substring(m_test_root.length() + 1, test_path.length() - m_test_root.length() - 1) };
|
|
||||||
|
|
||||||
// Collect logged messages
|
|
||||||
auto& arr = interpreter->vm().get_variable("__UserOutput__", interpreter->global_object()).as_array();
|
|
||||||
for (auto& entry : arr.indexed_properties()) {
|
|
||||||
auto message = entry.value_and_attributes(&interpreter->global_object()).value;
|
|
||||||
file_result.logged_messages.append(message.to_string_without_side_effects());
|
|
||||||
}
|
|
||||||
|
|
||||||
test_json.value().as_object().for_each_member([&](const String& suite_name, const JsonValue& suite_value) {
|
|
||||||
Test::Suite suite { suite_name };
|
|
||||||
|
|
||||||
VERIFY(suite_value.is_object());
|
|
||||||
|
|
||||||
suite_value.as_object().for_each_member([&](const String& test_name, const JsonValue& test_value) {
|
|
||||||
Test::Case test { test_name, Test::Result::Fail, "" };
|
|
||||||
|
|
||||||
VERIFY(test_value.is_object());
|
|
||||||
VERIFY(test_value.as_object().has("result"));
|
|
||||||
|
|
||||||
auto result = test_value.as_object().get("result");
|
|
||||||
VERIFY(result.is_string());
|
|
||||||
auto result_string = result.as_string();
|
|
||||||
if (result_string == "pass") {
|
|
||||||
test.result = Test::Result::Pass;
|
|
||||||
m_counts.tests_passed++;
|
|
||||||
} else if (result_string == "fail") {
|
|
||||||
test.result = Test::Result::Fail;
|
|
||||||
m_counts.tests_failed++;
|
|
||||||
suite.most_severe_test_result = Test::Result::Fail;
|
|
||||||
VERIFY(test_value.as_object().has("details"));
|
|
||||||
auto details = test_value.as_object().get("details");
|
|
||||||
VERIFY(result.is_string());
|
|
||||||
test.details = details.as_string();
|
|
||||||
} else {
|
|
||||||
test.result = Test::Result::Skip;
|
|
||||||
if (suite.most_severe_test_result == Test::Result::Pass)
|
|
||||||
suite.most_severe_test_result = Test::Result::Skip;
|
|
||||||
m_counts.tests_skipped++;
|
|
||||||
}
|
|
||||||
|
|
||||||
suite.tests.append(test);
|
|
||||||
});
|
|
||||||
|
|
||||||
if (suite.most_severe_test_result == Test::Result::Fail) {
|
|
||||||
m_counts.suites_failed++;
|
|
||||||
file_result.most_severe_test_result = Test::Result::Fail;
|
|
||||||
} else {
|
|
||||||
if (suite.most_severe_test_result == Test::Result::Skip && file_result.most_severe_test_result == Test::Result::Pass)
|
|
||||||
file_result.most_severe_test_result = Test::Result::Skip;
|
|
||||||
m_counts.suites_passed++;
|
|
||||||
}
|
|
||||||
|
|
||||||
file_result.suites.append(suite);
|
|
||||||
});
|
|
||||||
|
|
||||||
m_counts.files_total++;
|
|
||||||
|
|
||||||
file_result.time_taken = get_time_in_ms() - start_time;
|
|
||||||
m_total_elapsed_time_in_ms += file_result.time_taken;
|
|
||||||
|
|
||||||
return file_result;
|
|
||||||
}
|
|
||||||
|
|
||||||
enum Modifier {
|
|
||||||
BG_RED,
|
|
||||||
BG_GREEN,
|
|
||||||
FG_RED,
|
|
||||||
FG_GREEN,
|
|
||||||
FG_ORANGE,
|
|
||||||
FG_GRAY,
|
|
||||||
FG_BLACK,
|
|
||||||
FG_BOLD,
|
|
||||||
ITALIC,
|
|
||||||
CLEAR,
|
|
||||||
};
|
|
||||||
|
|
||||||
static void print_modifiers(Vector<Modifier> modifiers)
|
|
||||||
{
|
|
||||||
for (auto& modifier : modifiers) {
|
|
||||||
auto code = [&] {
|
|
||||||
switch (modifier) {
|
|
||||||
case BG_RED:
|
|
||||||
return "\033[48;2;255;0;102m";
|
|
||||||
case BG_GREEN:
|
|
||||||
return "\033[48;2;102;255;0m";
|
|
||||||
case FG_RED:
|
|
||||||
return "\033[38;2;255;0;102m";
|
|
||||||
case FG_GREEN:
|
|
||||||
return "\033[38;2;102;255;0m";
|
|
||||||
case FG_ORANGE:
|
|
||||||
return "\033[38;2;255;102;0m";
|
|
||||||
case FG_GRAY:
|
|
||||||
return "\033[38;2;135;139;148m";
|
|
||||||
case FG_BLACK:
|
|
||||||
return "\033[30m";
|
|
||||||
case FG_BOLD:
|
|
||||||
return "\033[1m";
|
|
||||||
case ITALIC:
|
|
||||||
return "\033[3m";
|
|
||||||
case CLEAR:
|
|
||||||
return "\033[0m";
|
|
||||||
}
|
|
||||||
VERIFY_NOT_REACHED();
|
|
||||||
}();
|
|
||||||
out("{}", code);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void TestRunner::print_file_result(const JSFileResult& file_result) const
|
|
||||||
{
|
|
||||||
if (file_result.most_severe_test_result == Test::Result::Fail || file_result.error.has_value()) {
|
|
||||||
print_modifiers({ BG_RED, FG_BLACK, FG_BOLD });
|
|
||||||
out(" FAIL ");
|
|
||||||
print_modifiers({ CLEAR });
|
|
||||||
} else {
|
|
||||||
if (m_print_times || file_result.most_severe_test_result != Test::Result::Pass) {
|
|
||||||
print_modifiers({ BG_GREEN, FG_BLACK, FG_BOLD });
|
|
||||||
out(" PASS ");
|
|
||||||
print_modifiers({ CLEAR });
|
|
||||||
} else {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
out(" {}", file_result.name);
|
|
||||||
|
|
||||||
if (m_print_times) {
|
|
||||||
print_modifiers({ CLEAR, ITALIC, FG_GRAY });
|
|
||||||
if (file_result.time_taken < 1000) {
|
|
||||||
outln(" ({}ms)", static_cast<int>(file_result.time_taken));
|
|
||||||
} else {
|
|
||||||
outln(" ({:3}s)", file_result.time_taken / 1000.0);
|
|
||||||
}
|
|
||||||
print_modifiers({ CLEAR });
|
|
||||||
} else {
|
|
||||||
outln();
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!file_result.logged_messages.is_empty()) {
|
|
||||||
print_modifiers({ FG_GRAY, FG_BOLD });
|
|
||||||
#ifdef __serenity__
|
|
||||||
outln(" ℹ Console output:");
|
|
||||||
#else
|
|
||||||
// This emoji has a second invisible byte after it. The one above does not
|
|
||||||
outln(" ℹ️ Console output:");
|
|
||||||
#endif
|
|
||||||
print_modifiers({ CLEAR, FG_GRAY });
|
|
||||||
for (auto& message : file_result.logged_messages)
|
|
||||||
outln(" {}", message);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (file_result.error.has_value()) {
|
|
||||||
auto test_error = file_result.error.value();
|
|
||||||
|
|
||||||
print_modifiers({ FG_RED });
|
|
||||||
#ifdef __serenity__
|
|
||||||
outln(" ❌ The file failed to parse");
|
|
||||||
#else
|
|
||||||
// No invisible byte here, but the spacing still needs to be altered on the host
|
|
||||||
outln(" ❌ The file failed to parse");
|
|
||||||
#endif
|
|
||||||
outln();
|
|
||||||
print_modifiers({ FG_GRAY });
|
|
||||||
for (auto& message : test_error.hint.split('\n', true)) {
|
|
||||||
outln(" {}", message);
|
|
||||||
}
|
|
||||||
print_modifiers({ FG_RED });
|
|
||||||
outln(" {}", test_error.error.to_string());
|
|
||||||
outln();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (file_result.most_severe_test_result != Test::Result::Pass) {
|
|
||||||
for (auto& suite : file_result.suites) {
|
|
||||||
if (suite.most_severe_test_result == Test::Result::Pass)
|
|
||||||
continue;
|
|
||||||
|
|
||||||
bool failed = suite.most_severe_test_result == Test::Result::Fail;
|
|
||||||
|
|
||||||
print_modifiers({ FG_GRAY, FG_BOLD });
|
|
||||||
|
|
||||||
if (failed) {
|
|
||||||
#ifdef __serenity__
|
|
||||||
out(" ❌ Suite: ");
|
|
||||||
#else
|
|
||||||
// No invisible byte here, but the spacing still needs to be altered on the host
|
|
||||||
out(" ❌ Suite: ");
|
|
||||||
#endif
|
|
||||||
} else {
|
|
||||||
#ifdef __serenity__
|
|
||||||
out(" ⚠ Suite: ");
|
|
||||||
#else
|
|
||||||
// This emoji has a second invisible byte after it. The one above does not
|
|
||||||
out(" ⚠️ Suite: ");
|
|
||||||
#endif
|
|
||||||
}
|
|
||||||
|
|
||||||
print_modifiers({ CLEAR, FG_GRAY });
|
|
||||||
|
|
||||||
if (suite.name == TOP_LEVEL_TEST_NAME) {
|
|
||||||
outln("<top-level>");
|
|
||||||
} else {
|
|
||||||
outln("{}", suite.name);
|
|
||||||
}
|
|
||||||
print_modifiers({ CLEAR });
|
|
||||||
|
|
||||||
for (auto& test : suite.tests) {
|
|
||||||
if (test.result == Test::Result::Pass)
|
|
||||||
continue;
|
|
||||||
|
|
||||||
print_modifiers({ FG_GRAY, FG_BOLD });
|
|
||||||
out(" Test: ");
|
|
||||||
if (test.result == Test::Result::Fail) {
|
|
||||||
print_modifiers({ CLEAR, FG_RED });
|
|
||||||
outln("{} (failed):", test.name);
|
|
||||||
outln(" {}", test.details);
|
|
||||||
} else {
|
|
||||||
print_modifiers({ CLEAR, FG_ORANGE });
|
|
||||||
outln("{} (skipped)", test.name);
|
|
||||||
}
|
|
||||||
print_modifiers({ CLEAR });
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void TestRunner::print_test_results() const
|
|
||||||
{
|
|
||||||
out("\nTest Suites: ");
|
|
||||||
if (m_counts.suites_failed) {
|
|
||||||
print_modifiers({ FG_RED });
|
|
||||||
out("{} failed, ", m_counts.suites_failed);
|
|
||||||
print_modifiers({ CLEAR });
|
|
||||||
}
|
|
||||||
if (m_counts.suites_passed) {
|
|
||||||
print_modifiers({ FG_GREEN });
|
|
||||||
out("{} passed, ", m_counts.suites_passed);
|
|
||||||
print_modifiers({ CLEAR });
|
|
||||||
}
|
|
||||||
outln("{} total", m_counts.suites_failed + m_counts.suites_passed);
|
|
||||||
|
|
||||||
out("Tests: ");
|
|
||||||
if (m_counts.tests_failed) {
|
|
||||||
print_modifiers({ FG_RED });
|
|
||||||
out("{} failed, ", m_counts.tests_failed);
|
|
||||||
print_modifiers({ CLEAR });
|
|
||||||
}
|
|
||||||
if (m_counts.tests_skipped) {
|
|
||||||
print_modifiers({ FG_ORANGE });
|
|
||||||
out("{} skipped, ", m_counts.tests_skipped);
|
|
||||||
print_modifiers({ CLEAR });
|
|
||||||
}
|
|
||||||
if (m_counts.tests_passed) {
|
|
||||||
print_modifiers({ FG_GREEN });
|
|
||||||
out("{} passed, ", m_counts.tests_passed);
|
|
||||||
print_modifiers({ CLEAR });
|
|
||||||
}
|
|
||||||
outln("{} total", m_counts.tests_failed + m_counts.tests_skipped + m_counts.tests_passed);
|
|
||||||
|
|
||||||
outln("Files: {} total", m_counts.files_total);
|
|
||||||
|
|
||||||
out("Time: ");
|
|
||||||
if (m_total_elapsed_time_in_ms < 1000.0) {
|
|
||||||
outln("{}ms", static_cast<int>(m_total_elapsed_time_in_ms));
|
|
||||||
} else {
|
|
||||||
outln("{:>.3}s", m_total_elapsed_time_in_ms / 1000.0);
|
|
||||||
}
|
|
||||||
outln();
|
|
||||||
}
|
|
||||||
|
|
||||||
class Test262ParserTestRunner final : public TestRunner {
|
|
||||||
public:
|
|
||||||
using TestRunner::TestRunner;
|
|
||||||
|
|
||||||
private:
|
|
||||||
virtual Vector<String> get_test_paths() const override;
|
|
||||||
virtual JSFileResult run_file_test(const String& test_path) override;
|
|
||||||
};
|
|
||||||
|
|
||||||
Vector<String> Test262ParserTestRunner::get_test_paths() const
|
|
||||||
{
|
|
||||||
Vector<String> paths;
|
|
||||||
iterate_directory_recursively(m_test_root, [&](const String& file_path) {
|
|
||||||
auto dirname = LexicalPath(file_path).dirname();
|
|
||||||
if (dirname.ends_with("early") || dirname.ends_with("fail") || dirname.ends_with("pass") || dirname.ends_with("pass-explicit"))
|
|
||||||
paths.append(file_path);
|
|
||||||
});
|
|
||||||
quick_sort(paths);
|
|
||||||
return paths;
|
|
||||||
}
|
|
||||||
|
|
||||||
JSFileResult Test262ParserTestRunner::run_file_test(const String& test_path)
|
|
||||||
{
|
|
||||||
currently_running_test = test_path;
|
|
||||||
|
|
||||||
auto dirname = LexicalPath(test_path).dirname();
|
|
||||||
bool expecting_file_to_parse;
|
|
||||||
if (dirname.ends_with("early") || dirname.ends_with("fail")) {
|
|
||||||
expecting_file_to_parse = false;
|
|
||||||
} else if (dirname.ends_with("pass") || dirname.ends_with("pass-explicit")) {
|
|
||||||
expecting_file_to_parse = true;
|
|
||||||
} else {
|
|
||||||
VERIFY_NOT_REACHED();
|
|
||||||
}
|
|
||||||
|
|
||||||
auto start_time = get_time_in_ms();
|
|
||||||
String details = "";
|
|
||||||
Test::Result test_result;
|
|
||||||
if (test_path.ends_with(".module.js")) {
|
|
||||||
test_result = Test::Result::Skip;
|
|
||||||
m_counts.tests_skipped++;
|
|
||||||
m_counts.suites_passed++;
|
|
||||||
} else {
|
|
||||||
auto parse_result = parse_file(test_path);
|
|
||||||
if (expecting_file_to_parse) {
|
|
||||||
if (!parse_result.is_error()) {
|
|
||||||
test_result = Test::Result::Pass;
|
|
||||||
} else {
|
|
||||||
test_result = Test::Result::Fail;
|
|
||||||
details = parse_result.error().error.to_string();
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
if (parse_result.is_error()) {
|
|
||||||
test_result = Test::Result::Pass;
|
|
||||||
} else {
|
|
||||||
test_result = Test::Result::Fail;
|
|
||||||
details = "File was expected to produce a parser error but didn't";
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// test262-parser-tests doesn't have "suites" and "tests" in the usual sense, it just has files
|
|
||||||
// and an expectation whether they should parse or not. We add one suite with one test nonetheless:
|
|
||||||
//
|
|
||||||
// - This makes interpreting skipped test easier as their file is shown as "PASS"
|
|
||||||
// - That way we can show additional information such as "file parsed but shouldn't have" or
|
|
||||||
// parser errors for files that should parse respectively
|
|
||||||
|
|
||||||
Test::Case test { expecting_file_to_parse ? "file should parse" : "file should not parse", test_result, details };
|
|
||||||
Test::Suite suite { "Parse file", test_result, { test } };
|
|
||||||
JSFileResult file_result {
|
|
||||||
test_path.substring(m_test_root.length() + 1, test_path.length() - m_test_root.length() - 1),
|
|
||||||
{},
|
|
||||||
get_time_in_ms() - start_time,
|
|
||||||
test_result,
|
|
||||||
{ suite }
|
|
||||||
};
|
|
||||||
|
|
||||||
if (test_result == Test::Result::Fail) {
|
|
||||||
m_counts.tests_failed++;
|
|
||||||
m_counts.suites_failed++;
|
|
||||||
} else {
|
|
||||||
m_counts.tests_passed++;
|
|
||||||
m_counts.suites_passed++;
|
|
||||||
}
|
|
||||||
m_counts.files_total++;
|
|
||||||
m_total_elapsed_time_in_ms += file_result.time_taken;
|
|
||||||
|
|
||||||
return file_result;
|
|
||||||
}
|
|
||||||
|
|
||||||
int main(int argc, char** argv)
|
|
||||||
{
|
|
||||||
struct sigaction act;
|
|
||||||
memset(&act, 0, sizeof(act));
|
|
||||||
act.sa_flags = SA_NOCLDWAIT;
|
|
||||||
act.sa_handler = handle_sigabrt;
|
|
||||||
int rc = sigaction(SIGABRT, &act, nullptr);
|
|
||||||
if (rc < 0) {
|
|
||||||
perror("sigaction");
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
#ifdef SIGINFO
|
|
||||||
signal(SIGINFO, [](int) {
|
|
||||||
static char buffer[4096];
|
|
||||||
auto& counts = TestRunner::the()->counts();
|
|
||||||
int len = snprintf(buffer, sizeof(buffer), "Pass: %d, Fail: %d, Skip: %d\nCurrent test: %s\n", counts.tests_passed, counts.tests_failed, counts.tests_skipped, currently_running_test.characters());
|
|
||||||
write(STDOUT_FILENO, buffer, len);
|
|
||||||
});
|
|
||||||
#endif
|
|
||||||
|
|
||||||
bool print_times = false;
|
|
||||||
bool print_progress =
|
|
||||||
#ifdef __serenity__
|
|
||||||
true; // Use OSC 9 to print progress
|
|
||||||
#else
|
|
||||||
false;
|
|
||||||
#endif
|
|
||||||
bool test262_parser_tests = false;
|
|
||||||
const char* specified_test_root = nullptr;
|
|
||||||
|
|
||||||
Core::ArgsParser args_parser;
|
|
||||||
args_parser.add_option(print_times, "Show duration of each test", "show-time", 't');
|
|
||||||
args_parser.add_option(Core::ArgsParser::Option {
|
|
||||||
.requires_argument = true,
|
|
||||||
.help_string = "Show progress with OSC 9 (true, false)",
|
|
||||||
.long_name = "show-progress",
|
|
||||||
.short_name = 'p',
|
|
||||||
.accept_value = [&](auto* str) {
|
|
||||||
if (StringView { "true" } == str)
|
|
||||||
print_progress = true;
|
|
||||||
else if (StringView { "false" } == str)
|
|
||||||
print_progress = false;
|
|
||||||
else
|
|
||||||
return false;
|
|
||||||
return true;
|
|
||||||
},
|
|
||||||
});
|
|
||||||
args_parser.add_option(collect_on_every_allocation, "Collect garbage after every allocation", "collect-often", 'g');
|
|
||||||
args_parser.add_option(test262_parser_tests, "Run test262 parser tests", "test262-parser-tests", 0);
|
|
||||||
args_parser.add_positional_argument(specified_test_root, "Tests root directory", "path", Core::ArgsParser::Required::No);
|
|
||||||
args_parser.parse(argc, argv);
|
|
||||||
|
|
||||||
if (test262_parser_tests) {
|
|
||||||
if (collect_on_every_allocation) {
|
|
||||||
warnln("--collect-often and --test262-parser-tests options must not be used together");
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
if (!specified_test_root) {
|
|
||||||
warnln("Test root is required with --test262-parser-tests");
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (getenv("DISABLE_DBG_OUTPUT")) {
|
|
||||||
AK::set_debug_enabled(false);
|
|
||||||
}
|
|
||||||
|
|
||||||
String test_root;
|
|
||||||
|
|
||||||
if (specified_test_root) {
|
|
||||||
test_root = String { specified_test_root };
|
|
||||||
} else {
|
|
||||||
#ifdef __serenity__
|
|
||||||
test_root = "/home/anon/js-tests";
|
|
||||||
#else
|
|
||||||
char* serenity_source_dir = getenv("SERENITY_SOURCE_DIR");
|
|
||||||
if (!serenity_source_dir) {
|
|
||||||
warnln("No test root given, test-js requires the SERENITY_SOURCE_DIR environment variable to be set");
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
test_root = String::formatted("{}/Userland/Libraries/LibJS/Tests", serenity_source_dir);
|
|
||||||
#endif
|
|
||||||
}
|
|
||||||
if (!Core::File::is_directory(test_root)) {
|
|
||||||
warnln("Test root is not a directory: {}", test_root);
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
vm = JS::VM::create();
|
|
||||||
|
|
||||||
Test::Counts result_counts;
|
|
||||||
if (test262_parser_tests)
|
|
||||||
result_counts = Test262ParserTestRunner(test_root, print_times, print_progress).run();
|
|
||||||
else
|
|
||||||
result_counts = TestRunner(test_root, print_times, print_progress).run();
|
|
||||||
|
|
||||||
vm = nullptr;
|
|
||||||
|
|
||||||
return result_counts.tests_failed > 0 ? 1 : 0;
|
|
||||||
}
|
|
||||||
|
|
591
Userland/Libraries/LibTest/JavaScriptTestRunner.h
Normal file
591
Userland/Libraries/LibTest/JavaScriptTestRunner.h
Normal file
|
@ -0,0 +1,591 @@
|
||||||
|
/*
|
||||||
|
* Copyright (c) 2020, Matthew Olsson <mattco@serenityos.org>
|
||||||
|
* Copyright (c) 2020-2021, Linus Groh <linusg@serenityos.org>
|
||||||
|
* Copyright (c) 2021, Ali Mohammad Pur <mpfard@serenityos.org>
|
||||||
|
*
|
||||||
|
* SPDX-License-Identifier: BSD-2-Clause
|
||||||
|
*/
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <AK/ByteBuffer.h>
|
||||||
|
#include <AK/JsonObject.h>
|
||||||
|
#include <AK/JsonValue.h>
|
||||||
|
#include <AK/LexicalPath.h>
|
||||||
|
#include <AK/QuickSort.h>
|
||||||
|
#include <AK/Result.h>
|
||||||
|
#include <AK/Tuple.h>
|
||||||
|
#include <LibCore/ArgsParser.h>
|
||||||
|
#include <LibCore/DirIterator.h>
|
||||||
|
#include <LibCore/File.h>
|
||||||
|
#include <LibJS/Interpreter.h>
|
||||||
|
#include <LibJS/Lexer.h>
|
||||||
|
#include <LibJS/Parser.h>
|
||||||
|
#include <LibJS/Runtime/Array.h>
|
||||||
|
#include <LibJS/Runtime/GlobalObject.h>
|
||||||
|
#include <LibJS/Runtime/JSONObject.h>
|
||||||
|
#include <LibJS/Runtime/TypedArray.h>
|
||||||
|
#include <LibTest/Results.h>
|
||||||
|
#include <sys/time.h>
|
||||||
|
#include <unistd.h>
|
||||||
|
|
||||||
|
#define STRCAT(x, y) __STRCAT(x, y)
|
||||||
|
#define STRSTRCAT(x, y) __STRSTRCAT(x, y)
|
||||||
|
#define __STRCAT(x, y) x #y
|
||||||
|
#define __STRSTRCAT(x, y) x y
|
||||||
|
|
||||||
|
// Note: This is a little weird, so here's an explanation:
|
||||||
|
// If the vararg isn't given, the tuple initializer will simply expand to `fn, ::Test::JS::__testjs_last<1>()`
|
||||||
|
// and if it _is_ given (say as `A`), the tuple initializer will expand to `fn, ::Test::JS::__testjs_last<1, A>()`, which will end up being evaluated as `A`
|
||||||
|
// and if multiple args are given, the static_assert will be sad.
|
||||||
|
#define __TESTJS_REGISTER_GLOBAL_FUNCTION(name, fn, ...) \
|
||||||
|
struct __TestJS_register_##fn { \
|
||||||
|
static_assert( \
|
||||||
|
::Test::JS::__testjs_count(__VA_ARGS__) <= 1, \
|
||||||
|
STRCAT(STRSTRCAT(STRCAT("Expected at most three arguments to TESTJS_GLOBAL_FUNCTION at line", __LINE__), ", in file "), __FILE__)); \
|
||||||
|
__TestJS_register_##fn() noexcept \
|
||||||
|
{ \
|
||||||
|
::Test::JS::s_exposed_global_functions.set( \
|
||||||
|
name, \
|
||||||
|
{ fn, ::Test::JS::__testjs_last<1, ##__VA_ARGS__>() }); \
|
||||||
|
} \
|
||||||
|
} __testjs_register_##fn {};
|
||||||
|
|
||||||
|
#define TESTJS_GLOBAL_FUNCTION(function, exposed_name, ...) \
|
||||||
|
JS_DECLARE_NATIVE_FUNCTION(function); \
|
||||||
|
__TESTJS_REGISTER_GLOBAL_FUNCTION(#exposed_name, function, ##__VA_ARGS__); \
|
||||||
|
JS_DEFINE_NATIVE_FUNCTION(function)
|
||||||
|
|
||||||
|
#define TESTJS_MAIN_HOOK() \
|
||||||
|
struct __TestJS_main_hook { \
|
||||||
|
__TestJS_main_hook() \
|
||||||
|
{ \
|
||||||
|
::Test::JS::g_main_hook = hook; \
|
||||||
|
} \
|
||||||
|
static void hook(); \
|
||||||
|
} __testjs_common_register_##name {}; \
|
||||||
|
void __TestJS_main_hook::hook()
|
||||||
|
|
||||||
|
#define TEST_ROOT(path) \
|
||||||
|
String Test::JS::g_test_root_fragment = path
|
||||||
|
|
||||||
|
namespace Test::JS {
|
||||||
|
|
||||||
|
namespace JS = ::JS;
|
||||||
|
|
||||||
|
template<typename... Args>
|
||||||
|
static consteval size_t __testjs_count(Args...) { return sizeof...(Args); }
|
||||||
|
|
||||||
|
template<auto... Values>
|
||||||
|
static consteval size_t __testjs_last() { return (AK::Detail::IntegralConstant<size_t, Values> {}, ...).value; }
|
||||||
|
|
||||||
|
static constexpr auto TOP_LEVEL_TEST_NAME = "__$$TOP_LEVEL$$__";
|
||||||
|
extern RefPtr<JS::VM> g_vm;
|
||||||
|
extern bool g_collect_on_every_allocation;
|
||||||
|
extern String g_currently_running_test;
|
||||||
|
extern String g_test_glob;
|
||||||
|
struct FunctionWithLength {
|
||||||
|
JS::Value (*function)(JS::VM&, JS::GlobalObject&);
|
||||||
|
size_t length { 0 };
|
||||||
|
};
|
||||||
|
extern HashMap<String, FunctionWithLength> s_exposed_global_functions;
|
||||||
|
extern String g_test_root_fragment;
|
||||||
|
extern String g_test_root;
|
||||||
|
extern int g_test_argc;
|
||||||
|
extern char** g_test_argv;
|
||||||
|
extern Function<void()> g_main_hook;
|
||||||
|
|
||||||
|
struct ParserError {
|
||||||
|
JS::Parser::Error error;
|
||||||
|
String hint;
|
||||||
|
};
|
||||||
|
|
||||||
|
struct JSFileResult {
|
||||||
|
String name;
|
||||||
|
Optional<ParserError> error {};
|
||||||
|
double time_taken { 0 };
|
||||||
|
// A failed test takes precedence over a skipped test, which both have
|
||||||
|
// precedence over a passed test
|
||||||
|
Test::Result most_severe_test_result { Test::Result::Pass };
|
||||||
|
Vector<Test::Suite> suites {};
|
||||||
|
Vector<String> logged_messages {};
|
||||||
|
};
|
||||||
|
|
||||||
|
class TestRunner {
|
||||||
|
public:
|
||||||
|
static TestRunner* the()
|
||||||
|
{
|
||||||
|
return s_the;
|
||||||
|
}
|
||||||
|
|
||||||
|
TestRunner(String test_root, String common_path, bool print_times, bool print_progress)
|
||||||
|
: m_common_path(move(common_path))
|
||||||
|
, m_test_root(move(test_root))
|
||||||
|
, m_print_times(print_times)
|
||||||
|
, m_print_progress(print_progress)
|
||||||
|
{
|
||||||
|
VERIFY(!s_the);
|
||||||
|
s_the = this;
|
||||||
|
g_test_root = m_test_root;
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual ~TestRunner() = default;
|
||||||
|
|
||||||
|
void run();
|
||||||
|
|
||||||
|
const Test::Counts& counts() const { return m_counts; }
|
||||||
|
|
||||||
|
bool is_printing_progress() const { return m_print_progress; }
|
||||||
|
|
||||||
|
protected:
|
||||||
|
static TestRunner* s_the;
|
||||||
|
|
||||||
|
virtual Vector<String> get_test_paths() const;
|
||||||
|
virtual JSFileResult run_file_test(const String& test_path);
|
||||||
|
void print_file_result(const JSFileResult& file_result) const;
|
||||||
|
void print_test_results() const;
|
||||||
|
|
||||||
|
String m_common_path;
|
||||||
|
String m_test_root;
|
||||||
|
bool m_print_times;
|
||||||
|
bool m_print_progress;
|
||||||
|
|
||||||
|
double m_total_elapsed_time_in_ms { 0 };
|
||||||
|
Test::Counts m_counts;
|
||||||
|
|
||||||
|
RefPtr<JS::Program> m_test_program;
|
||||||
|
};
|
||||||
|
|
||||||
|
class TestRunnerGlobalObject final : public JS::GlobalObject {
|
||||||
|
JS_OBJECT(TestRunnerGlobalObject, JS::GlobalObject);
|
||||||
|
|
||||||
|
public:
|
||||||
|
TestRunnerGlobalObject() = default;
|
||||||
|
virtual ~TestRunnerGlobalObject() override = default;
|
||||||
|
|
||||||
|
virtual void initialize_global_object() override;
|
||||||
|
};
|
||||||
|
|
||||||
|
inline void TestRunnerGlobalObject::initialize_global_object()
|
||||||
|
{
|
||||||
|
Base::initialize_global_object();
|
||||||
|
define_property("global", this, JS::Attribute::Enumerable);
|
||||||
|
for (auto& entry : s_exposed_global_functions) {
|
||||||
|
define_native_function(
|
||||||
|
entry.key, [fn = entry.value.function](auto& vm, auto& global_object) {
|
||||||
|
return fn(vm, global_object);
|
||||||
|
},
|
||||||
|
entry.value.length);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
inline void cleanup_and_exit()
|
||||||
|
{
|
||||||
|
// Clear the taskbar progress.
|
||||||
|
if (TestRunner::the() && TestRunner::the()->is_printing_progress())
|
||||||
|
warn("\033]9;-1;\033\\");
|
||||||
|
exit(1);
|
||||||
|
}
|
||||||
|
|
||||||
|
inline double get_time_in_ms()
|
||||||
|
{
|
||||||
|
struct timeval tv1;
|
||||||
|
auto return_code = gettimeofday(&tv1, nullptr);
|
||||||
|
VERIFY(return_code >= 0);
|
||||||
|
return static_cast<double>(tv1.tv_sec) * 1000.0 + static_cast<double>(tv1.tv_usec) / 1000.0;
|
||||||
|
}
|
||||||
|
|
||||||
|
template<typename Callback>
|
||||||
|
inline void iterate_directory_recursively(const String& directory_path, Callback callback)
|
||||||
|
{
|
||||||
|
Core::DirIterator directory_iterator(directory_path, Core::DirIterator::Flags::SkipDots);
|
||||||
|
|
||||||
|
while (directory_iterator.has_next()) {
|
||||||
|
auto file_path = directory_iterator.next_full_path();
|
||||||
|
auto is_directory = Core::File::is_directory(file_path);
|
||||||
|
if (is_directory && !file_path.contains("/Fixtures")) {
|
||||||
|
iterate_directory_recursively(file_path, callback);
|
||||||
|
} else if (!is_directory) {
|
||||||
|
callback(move(file_path));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
inline Vector<String> TestRunner::get_test_paths() const
|
||||||
|
{
|
||||||
|
Vector<String> paths;
|
||||||
|
iterate_directory_recursively(m_test_root, [&](const String& file_path) {
|
||||||
|
if (!file_path.ends_with(".js"))
|
||||||
|
return;
|
||||||
|
if (!file_path.ends_with("test-common.js"))
|
||||||
|
paths.append(file_path);
|
||||||
|
});
|
||||||
|
quick_sort(paths);
|
||||||
|
return paths;
|
||||||
|
}
|
||||||
|
|
||||||
|
inline void TestRunner::run()
|
||||||
|
{
|
||||||
|
size_t progress_counter = 0;
|
||||||
|
auto test_paths = get_test_paths();
|
||||||
|
for (auto& path : test_paths) {
|
||||||
|
if (!path.matches(g_test_glob))
|
||||||
|
continue;
|
||||||
|
++progress_counter;
|
||||||
|
print_file_result(run_file_test(path));
|
||||||
|
if (m_print_progress)
|
||||||
|
warn("\033]9;{};{};\033\\", progress_counter, test_paths.size());
|
||||||
|
}
|
||||||
|
|
||||||
|
if (m_print_progress)
|
||||||
|
warn("\033]9;-1;\033\\");
|
||||||
|
|
||||||
|
print_test_results();
|
||||||
|
}
|
||||||
|
|
||||||
|
inline AK::Result<NonnullRefPtr<JS::Program>, ParserError> parse_file(const String& file_path)
|
||||||
|
{
|
||||||
|
auto file = Core::File::construct(file_path);
|
||||||
|
auto result = file->open(Core::OpenMode::ReadOnly);
|
||||||
|
if (!result) {
|
||||||
|
warnln("Failed to open the following file: \"{}\"", file_path);
|
||||||
|
cleanup_and_exit();
|
||||||
|
}
|
||||||
|
|
||||||
|
auto contents = file->read_all();
|
||||||
|
String test_file_string(reinterpret_cast<const char*>(contents.data()), contents.size());
|
||||||
|
file->close();
|
||||||
|
|
||||||
|
auto parser = JS::Parser(JS::Lexer(test_file_string));
|
||||||
|
auto program = parser.parse_program();
|
||||||
|
|
||||||
|
if (parser.has_errors()) {
|
||||||
|
auto error = parser.errors()[0];
|
||||||
|
return AK::Result<NonnullRefPtr<JS::Program>, ParserError>(ParserError { error, error.source_location_hint(test_file_string) });
|
||||||
|
}
|
||||||
|
|
||||||
|
return AK::Result<NonnullRefPtr<JS::Program>, ParserError>(program);
|
||||||
|
}
|
||||||
|
|
||||||
|
inline Optional<JsonValue> get_test_results(JS::Interpreter& interpreter)
|
||||||
|
{
|
||||||
|
auto result = g_vm->get_variable("__TestResults__", interpreter.global_object());
|
||||||
|
auto json_string = JS::JSONObject::stringify_impl(interpreter.global_object(), result, JS::js_undefined(), JS::js_undefined());
|
||||||
|
|
||||||
|
auto json = JsonValue::from_string(json_string);
|
||||||
|
if (!json.has_value())
|
||||||
|
return {};
|
||||||
|
|
||||||
|
return json.value();
|
||||||
|
}
|
||||||
|
|
||||||
|
inline JSFileResult TestRunner::run_file_test(const String& test_path)
|
||||||
|
{
|
||||||
|
g_currently_running_test = test_path;
|
||||||
|
|
||||||
|
double start_time = get_time_in_ms();
|
||||||
|
auto interpreter = JS::Interpreter::create<TestRunnerGlobalObject>(*g_vm);
|
||||||
|
|
||||||
|
// FIXME: This is a hack while we're refactoring Interpreter/VM stuff.
|
||||||
|
JS::VM::InterpreterExecutionScope scope(*interpreter);
|
||||||
|
|
||||||
|
interpreter->heap().set_should_collect_on_every_allocation(g_collect_on_every_allocation);
|
||||||
|
|
||||||
|
if (!m_test_program) {
|
||||||
|
auto result = parse_file(m_common_path);
|
||||||
|
if (result.is_error()) {
|
||||||
|
warnln("Unable to parse test-common.js");
|
||||||
|
warnln("{}", result.error().error.to_string());
|
||||||
|
warnln("{}", result.error().hint);
|
||||||
|
cleanup_and_exit();
|
||||||
|
}
|
||||||
|
m_test_program = result.value();
|
||||||
|
}
|
||||||
|
|
||||||
|
interpreter->run(interpreter->global_object(), *m_test_program);
|
||||||
|
|
||||||
|
auto file_program = parse_file(test_path);
|
||||||
|
if (file_program.is_error())
|
||||||
|
return { test_path, file_program.error() };
|
||||||
|
interpreter->run(interpreter->global_object(), *file_program.value());
|
||||||
|
|
||||||
|
if (g_vm->exception())
|
||||||
|
g_vm->clear_exception();
|
||||||
|
|
||||||
|
auto test_json = get_test_results(*interpreter);
|
||||||
|
if (!test_json.has_value()) {
|
||||||
|
warnln("Received malformed JSON from test \"{}\"", test_path);
|
||||||
|
cleanup_and_exit();
|
||||||
|
}
|
||||||
|
|
||||||
|
JSFileResult file_result { test_path.substring(m_test_root.length() + 1, test_path.length() - m_test_root.length() - 1) };
|
||||||
|
|
||||||
|
// Collect logged messages
|
||||||
|
auto& arr = interpreter->vm().get_variable("__UserOutput__", interpreter->global_object()).as_array();
|
||||||
|
for (auto& entry : arr.indexed_properties()) {
|
||||||
|
auto message = entry.value_and_attributes(&interpreter->global_object()).value;
|
||||||
|
file_result.logged_messages.append(message.to_string_without_side_effects());
|
||||||
|
}
|
||||||
|
|
||||||
|
test_json.value().as_object().for_each_member([&](const String& suite_name, const JsonValue& suite_value) {
|
||||||
|
Test::Suite suite { suite_name };
|
||||||
|
|
||||||
|
VERIFY(suite_value.is_object());
|
||||||
|
|
||||||
|
suite_value.as_object().for_each_member([&](const String& test_name, const JsonValue& test_value) {
|
||||||
|
Test::Case test { test_name, Test::Result::Fail, "" };
|
||||||
|
|
||||||
|
VERIFY(test_value.is_object());
|
||||||
|
VERIFY(test_value.as_object().has("result"));
|
||||||
|
|
||||||
|
auto result = test_value.as_object().get("result");
|
||||||
|
VERIFY(result.is_string());
|
||||||
|
auto result_string = result.as_string();
|
||||||
|
if (result_string == "pass") {
|
||||||
|
test.result = Test::Result::Pass;
|
||||||
|
m_counts.tests_passed++;
|
||||||
|
} else if (result_string == "fail") {
|
||||||
|
test.result = Test::Result::Fail;
|
||||||
|
m_counts.tests_failed++;
|
||||||
|
suite.most_severe_test_result = Test::Result::Fail;
|
||||||
|
VERIFY(test_value.as_object().has("details"));
|
||||||
|
auto details = test_value.as_object().get("details");
|
||||||
|
VERIFY(result.is_string());
|
||||||
|
test.details = details.as_string();
|
||||||
|
} else {
|
||||||
|
test.result = Test::Result::Skip;
|
||||||
|
if (suite.most_severe_test_result == Test::Result::Pass)
|
||||||
|
suite.most_severe_test_result = Test::Result::Skip;
|
||||||
|
m_counts.tests_skipped++;
|
||||||
|
}
|
||||||
|
|
||||||
|
suite.tests.append(test);
|
||||||
|
});
|
||||||
|
|
||||||
|
if (suite.most_severe_test_result == Test::Result::Fail) {
|
||||||
|
m_counts.suites_failed++;
|
||||||
|
file_result.most_severe_test_result = Test::Result::Fail;
|
||||||
|
} else {
|
||||||
|
if (suite.most_severe_test_result == Test::Result::Skip && file_result.most_severe_test_result == Test::Result::Pass)
|
||||||
|
file_result.most_severe_test_result = Test::Result::Skip;
|
||||||
|
m_counts.suites_passed++;
|
||||||
|
}
|
||||||
|
|
||||||
|
file_result.suites.append(suite);
|
||||||
|
});
|
||||||
|
|
||||||
|
m_counts.files_total++;
|
||||||
|
|
||||||
|
file_result.time_taken = get_time_in_ms() - start_time;
|
||||||
|
m_total_elapsed_time_in_ms += file_result.time_taken;
|
||||||
|
|
||||||
|
return file_result;
|
||||||
|
}
|
||||||
|
|
||||||
|
enum Modifier {
|
||||||
|
BG_RED,
|
||||||
|
BG_GREEN,
|
||||||
|
FG_RED,
|
||||||
|
FG_GREEN,
|
||||||
|
FG_ORANGE,
|
||||||
|
FG_GRAY,
|
||||||
|
FG_BLACK,
|
||||||
|
FG_BOLD,
|
||||||
|
ITALIC,
|
||||||
|
CLEAR,
|
||||||
|
};
|
||||||
|
|
||||||
|
inline void print_modifiers(Vector<Modifier> modifiers)
|
||||||
|
{
|
||||||
|
for (auto& modifier : modifiers) {
|
||||||
|
auto code = [&] {
|
||||||
|
switch (modifier) {
|
||||||
|
case BG_RED:
|
||||||
|
return "\033[48;2;255;0;102m";
|
||||||
|
case BG_GREEN:
|
||||||
|
return "\033[48;2;102;255;0m";
|
||||||
|
case FG_RED:
|
||||||
|
return "\033[38;2;255;0;102m";
|
||||||
|
case FG_GREEN:
|
||||||
|
return "\033[38;2;102;255;0m";
|
||||||
|
case FG_ORANGE:
|
||||||
|
return "\033[38;2;255;102;0m";
|
||||||
|
case FG_GRAY:
|
||||||
|
return "\033[38;2;135;139;148m";
|
||||||
|
case FG_BLACK:
|
||||||
|
return "\033[30m";
|
||||||
|
case FG_BOLD:
|
||||||
|
return "\033[1m";
|
||||||
|
case ITALIC:
|
||||||
|
return "\033[3m";
|
||||||
|
case CLEAR:
|
||||||
|
return "\033[0m";
|
||||||
|
}
|
||||||
|
VERIFY_NOT_REACHED();
|
||||||
|
}();
|
||||||
|
out("{}", code);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
inline void TestRunner::print_file_result(const JSFileResult& file_result) const
|
||||||
|
{
|
||||||
|
if (file_result.most_severe_test_result == Test::Result::Fail || file_result.error.has_value()) {
|
||||||
|
print_modifiers({ BG_RED, FG_BLACK, FG_BOLD });
|
||||||
|
out(" FAIL ");
|
||||||
|
print_modifiers({ CLEAR });
|
||||||
|
} else {
|
||||||
|
if (m_print_times || file_result.most_severe_test_result != Test::Result::Pass) {
|
||||||
|
print_modifiers({ BG_GREEN, FG_BLACK, FG_BOLD });
|
||||||
|
out(" PASS ");
|
||||||
|
print_modifiers({ CLEAR });
|
||||||
|
} else {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
out(" {}", file_result.name);
|
||||||
|
|
||||||
|
if (m_print_times) {
|
||||||
|
print_modifiers({ CLEAR, ITALIC, FG_GRAY });
|
||||||
|
if (file_result.time_taken < 1000) {
|
||||||
|
outln(" ({}ms)", static_cast<int>(file_result.time_taken));
|
||||||
|
} else {
|
||||||
|
outln(" ({:3}s)", file_result.time_taken / 1000.0);
|
||||||
|
}
|
||||||
|
print_modifiers({ CLEAR });
|
||||||
|
} else {
|
||||||
|
outln();
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!file_result.logged_messages.is_empty()) {
|
||||||
|
print_modifiers({ FG_GRAY, FG_BOLD });
|
||||||
|
#ifdef __serenity__
|
||||||
|
outln(" ℹ Console output:");
|
||||||
|
#else
|
||||||
|
// This emoji has a second invisible byte after it. The one above does not
|
||||||
|
outln(" ℹ️ Console output:");
|
||||||
|
#endif
|
||||||
|
print_modifiers({ CLEAR, FG_GRAY });
|
||||||
|
for (auto& message : file_result.logged_messages)
|
||||||
|
outln(" {}", message);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (file_result.error.has_value()) {
|
||||||
|
auto test_error = file_result.error.value();
|
||||||
|
|
||||||
|
print_modifiers({ FG_RED });
|
||||||
|
#ifdef __serenity__
|
||||||
|
outln(" ❌ The file failed to parse");
|
||||||
|
#else
|
||||||
|
// No invisible byte here, but the spacing still needs to be altered on the host
|
||||||
|
outln(" ❌ The file failed to parse");
|
||||||
|
#endif
|
||||||
|
outln();
|
||||||
|
print_modifiers({ FG_GRAY });
|
||||||
|
for (auto& message : test_error.hint.split('\n', true)) {
|
||||||
|
outln(" {}", message);
|
||||||
|
}
|
||||||
|
print_modifiers({ FG_RED });
|
||||||
|
outln(" {}", test_error.error.to_string());
|
||||||
|
outln();
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (file_result.most_severe_test_result != Test::Result::Pass) {
|
||||||
|
for (auto& suite : file_result.suites) {
|
||||||
|
if (suite.most_severe_test_result == Test::Result::Pass)
|
||||||
|
continue;
|
||||||
|
|
||||||
|
bool failed = suite.most_severe_test_result == Test::Result::Fail;
|
||||||
|
|
||||||
|
print_modifiers({ FG_GRAY, FG_BOLD });
|
||||||
|
|
||||||
|
if (failed) {
|
||||||
|
#ifdef __serenity__
|
||||||
|
out(" ❌ Suite: ");
|
||||||
|
#else
|
||||||
|
// No invisible byte here, but the spacing still needs to be altered on the host
|
||||||
|
out(" ❌ Suite: ");
|
||||||
|
#endif
|
||||||
|
} else {
|
||||||
|
#ifdef __serenity__
|
||||||
|
out(" ⚠ Suite: ");
|
||||||
|
#else
|
||||||
|
// This emoji has a second invisible byte after it. The one above does not
|
||||||
|
out(" ⚠️ Suite: ");
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
|
print_modifiers({ CLEAR, FG_GRAY });
|
||||||
|
|
||||||
|
if (suite.name == TOP_LEVEL_TEST_NAME) {
|
||||||
|
outln("<top-level>");
|
||||||
|
} else {
|
||||||
|
outln("{}", suite.name);
|
||||||
|
}
|
||||||
|
print_modifiers({ CLEAR });
|
||||||
|
|
||||||
|
for (auto& test : suite.tests) {
|
||||||
|
if (test.result == Test::Result::Pass)
|
||||||
|
continue;
|
||||||
|
|
||||||
|
print_modifiers({ FG_GRAY, FG_BOLD });
|
||||||
|
out(" Test: ");
|
||||||
|
if (test.result == Test::Result::Fail) {
|
||||||
|
print_modifiers({ CLEAR, FG_RED });
|
||||||
|
outln("{} (failed):", test.name);
|
||||||
|
outln(" {}", test.details);
|
||||||
|
} else {
|
||||||
|
print_modifiers({ CLEAR, FG_ORANGE });
|
||||||
|
outln("{} (skipped)", test.name);
|
||||||
|
}
|
||||||
|
print_modifiers({ CLEAR });
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
inline void TestRunner::print_test_results() const
|
||||||
|
{
|
||||||
|
out("\nTest Suites: ");
|
||||||
|
if (m_counts.suites_failed) {
|
||||||
|
print_modifiers({ FG_RED });
|
||||||
|
out("{} failed, ", m_counts.suites_failed);
|
||||||
|
print_modifiers({ CLEAR });
|
||||||
|
}
|
||||||
|
if (m_counts.suites_passed) {
|
||||||
|
print_modifiers({ FG_GREEN });
|
||||||
|
out("{} passed, ", m_counts.suites_passed);
|
||||||
|
print_modifiers({ CLEAR });
|
||||||
|
}
|
||||||
|
outln("{} total", m_counts.suites_failed + m_counts.suites_passed);
|
||||||
|
|
||||||
|
out("Tests: ");
|
||||||
|
if (m_counts.tests_failed) {
|
||||||
|
print_modifiers({ FG_RED });
|
||||||
|
out("{} failed, ", m_counts.tests_failed);
|
||||||
|
print_modifiers({ CLEAR });
|
||||||
|
}
|
||||||
|
if (m_counts.tests_skipped) {
|
||||||
|
print_modifiers({ FG_ORANGE });
|
||||||
|
out("{} skipped, ", m_counts.tests_skipped);
|
||||||
|
print_modifiers({ CLEAR });
|
||||||
|
}
|
||||||
|
if (m_counts.tests_passed) {
|
||||||
|
print_modifiers({ FG_GREEN });
|
||||||
|
out("{} passed, ", m_counts.tests_passed);
|
||||||
|
print_modifiers({ CLEAR });
|
||||||
|
}
|
||||||
|
outln("{} total", m_counts.tests_failed + m_counts.tests_skipped + m_counts.tests_passed);
|
||||||
|
|
||||||
|
outln("Files: {} total", m_counts.files_total);
|
||||||
|
|
||||||
|
out("Time: ");
|
||||||
|
if (m_total_elapsed_time_in_ms < 1000.0) {
|
||||||
|
outln("{}ms", static_cast<int>(m_total_elapsed_time_in_ms));
|
||||||
|
} else {
|
||||||
|
outln("{:>.3}s", m_total_elapsed_time_in_ms / 1000.0);
|
||||||
|
}
|
||||||
|
outln();
|
||||||
|
}
|
||||||
|
}
|
156
Userland/Libraries/LibTest/JavaScriptTestRunnerMain.cpp
Normal file
156
Userland/Libraries/LibTest/JavaScriptTestRunnerMain.cpp
Normal file
|
@ -0,0 +1,156 @@
|
||||||
|
/*
|
||||||
|
* Copyright (c) 2020, Matthew Olsson <mattco@serenityos.org>
|
||||||
|
* Copyright (c) 2020-2021, Linus Groh <linusg@serenityos.org>
|
||||||
|
* Copyright (c) 2021, Ali Mohammad Pur <mpfard@serenityos.org>
|
||||||
|
*
|
||||||
|
* SPDX-License-Identifier: BSD-2-Clause
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <LibTest/JavaScriptTestRunner.h>
|
||||||
|
#include <signal.h>
|
||||||
|
#include <stdio.h>
|
||||||
|
|
||||||
|
namespace Test::JS {
|
||||||
|
|
||||||
|
RefPtr<::JS::VM> g_vm;
|
||||||
|
bool g_collect_on_every_allocation = false;
|
||||||
|
String g_currently_running_test;
|
||||||
|
String g_test_glob;
|
||||||
|
HashMap<String, FunctionWithLength> s_exposed_global_functions;
|
||||||
|
Function<void()> g_main_hook;
|
||||||
|
TestRunner* TestRunner::s_the = nullptr;
|
||||||
|
String g_test_root;
|
||||||
|
int g_test_argc;
|
||||||
|
char** g_test_argv;
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
using namespace Test::JS;
|
||||||
|
|
||||||
|
static StringView g_program_name { "test-js"sv };
|
||||||
|
|
||||||
|
static void handle_sigabrt(int)
|
||||||
|
{
|
||||||
|
dbgln("{}: SIGABRT received, cleaning up.", g_program_name);
|
||||||
|
cleanup_and_exit();
|
||||||
|
}
|
||||||
|
|
||||||
|
int main(int argc, char** argv)
|
||||||
|
{
|
||||||
|
g_test_argc = argc;
|
||||||
|
g_test_argv = argv;
|
||||||
|
auto program_name = LexicalPath { argv[0] }.basename();
|
||||||
|
g_program_name = program_name;
|
||||||
|
|
||||||
|
struct sigaction act;
|
||||||
|
memset(&act, 0, sizeof(act));
|
||||||
|
act.sa_flags = SA_NOCLDWAIT;
|
||||||
|
act.sa_handler = handle_sigabrt;
|
||||||
|
int rc = sigaction(SIGABRT, &act, nullptr);
|
||||||
|
if (rc < 0) {
|
||||||
|
perror("sigaction");
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
#ifdef SIGINFO
|
||||||
|
signal(SIGINFO, [](int) {
|
||||||
|
static char buffer[4096];
|
||||||
|
auto& counts = TestRunner::the()->counts();
|
||||||
|
int len = snprintf(buffer, sizeof(buffer), "Pass: %d, Fail: %d, Skip: %d\nCurrent test: %s\n", counts.tests_passed, counts.tests_failed, counts.tests_skipped, g_currently_running_test.characters());
|
||||||
|
write(STDOUT_FILENO, buffer, len);
|
||||||
|
});
|
||||||
|
#endif
|
||||||
|
|
||||||
|
bool print_times = false;
|
||||||
|
bool print_progress =
|
||||||
|
#ifdef __serenity__
|
||||||
|
true; // Use OSC 9 to print progress
|
||||||
|
#else
|
||||||
|
false;
|
||||||
|
#endif
|
||||||
|
const char* specified_test_root = nullptr;
|
||||||
|
String common_path;
|
||||||
|
|
||||||
|
Core::ArgsParser args_parser;
|
||||||
|
args_parser.add_option(print_times, "Show duration of each test", "show-time", 't');
|
||||||
|
args_parser.add_option(Core::ArgsParser::Option {
|
||||||
|
.requires_argument = true,
|
||||||
|
.help_string = "Show progress with OSC 9 (true, false)",
|
||||||
|
.long_name = "show-progress",
|
||||||
|
.short_name = 'p',
|
||||||
|
.accept_value = [&](auto* str) {
|
||||||
|
if (StringView { "true" } == str)
|
||||||
|
print_progress = true;
|
||||||
|
else if (StringView { "false" } == str)
|
||||||
|
print_progress = false;
|
||||||
|
else
|
||||||
|
return false;
|
||||||
|
return true;
|
||||||
|
},
|
||||||
|
});
|
||||||
|
args_parser.add_option(g_collect_on_every_allocation, "Collect garbage after every allocation", "collect-often", 'g');
|
||||||
|
args_parser.add_option(g_test_glob, "Only run tests matching the given glob", "filter", 'f', "glob");
|
||||||
|
args_parser.add_positional_argument(specified_test_root, "Tests root directory", "path", Core::ArgsParser::Required::No);
|
||||||
|
args_parser.add_positional_argument(common_path, "Path to tests-common.js", "common-path", Core::ArgsParser::Required::No);
|
||||||
|
args_parser.parse(argc, argv);
|
||||||
|
|
||||||
|
g_test_glob = String::formatted("*{}*", g_test_glob);
|
||||||
|
|
||||||
|
if (getenv("DISABLE_DBG_OUTPUT")) {
|
||||||
|
AK::set_debug_enabled(false);
|
||||||
|
}
|
||||||
|
|
||||||
|
String test_root;
|
||||||
|
|
||||||
|
if (specified_test_root) {
|
||||||
|
test_root = String { specified_test_root };
|
||||||
|
} else {
|
||||||
|
#ifdef __serenity__
|
||||||
|
test_root = LexicalPath::join("/home/anon", String::formatted("{}-tests", program_name.split_view('-').last())).string();
|
||||||
|
#else
|
||||||
|
char* serenity_source_dir = getenv("SERENITY_SOURCE_DIR");
|
||||||
|
if (!serenity_source_dir) {
|
||||||
|
warnln("No test root given, {} requires the SERENITY_SOURCE_DIR environment variable to be set", g_program_name);
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
test_root = String::formatted("{}/{}", serenity_source_dir, g_test_root_fragment);
|
||||||
|
common_path = String::formatted("{}/Userland/Libraries/LibJS/Tests/test-common.js", serenity_source_dir);
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
if (!Core::File::is_directory(test_root)) {
|
||||||
|
warnln("Test root is not a directory: {}", test_root);
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (common_path.is_empty()) {
|
||||||
|
#ifdef __serenity__
|
||||||
|
common_path = "/home/anon/js-tests/test-common.js";
|
||||||
|
#else
|
||||||
|
char* serenity_source_dir = getenv("SERENITY_SOURCE_DIR");
|
||||||
|
if (!serenity_source_dir) {
|
||||||
|
warnln("No test root given, {} requires the SERENITY_SOURCE_DIR environment variable to be set", g_program_name);
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
common_path = String::formatted("{}/Userland/Libraries/LibJS/Tests/test-common.js", serenity_source_dir);
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
|
if (chdir(test_root.characters()) < 0) {
|
||||||
|
auto saved_errno = errno;
|
||||||
|
warnln("chdir failed: {}", strerror(saved_errno));
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (g_main_hook)
|
||||||
|
g_main_hook();
|
||||||
|
|
||||||
|
if (!g_vm)
|
||||||
|
g_vm = JS::VM::create();
|
||||||
|
|
||||||
|
TestRunner test_runner(test_root, common_path, print_times, print_progress);
|
||||||
|
test_runner.run();
|
||||||
|
|
||||||
|
g_vm = nullptr;
|
||||||
|
|
||||||
|
return test_runner.counts().tests_failed > 0 ? 1 : 0;
|
||||||
|
}
|
Loading…
Reference in a new issue