summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--Meta/CMake/utils.cmake23
-rw-r--r--Meta/Lagom/CMakeLists.txt4
-rw-r--r--Tests/LibJS/CMakeLists.txt3
-rw-r--r--Tests/LibJS/test-js.cpp732
-rw-r--r--Userland/Libraries/LibTest/JavaScriptTestRunner.h591
-rw-r--r--Userland/Libraries/LibTest/JavaScriptTestRunnerMain.cpp156
6 files changed, 776 insertions, 733 deletions
diff --git a/Meta/CMake/utils.cmake b/Meta/CMake/utils.cmake
index a0c170e5c7..6cadb1879e 100644
--- a/Meta/CMake/utils.cmake
+++ b/Meta/CMake/utils.cmake
@@ -70,10 +70,14 @@ function(serenity_bin target_name)
endfunction()
function(serenity_test test_src sub_dir)
- cmake_parse_arguments(SERENITY_TEST "CUSTOM_MAIN" "" "LIBS" ${ARGN})
+ cmake_parse_arguments(SERENITY_TEST "MAIN_ALREADY_DEFINED" "CUSTOM_MAIN" "LIBS" ${ARGN})
set(TEST_SOURCES ${test_src})
- if (NOT ${SERENITY_TEST_CUSTOM_MAIN})
- list(APPEND TEST_SOURCES "${CMAKE_SOURCE_DIR}/Userland/Libraries/LibTest/TestMain.cpp")
+ if ("${SERENITY_TEST_CUSTOM_MAIN}" STREQUAL "")
+ set(SERENITY_TEST_CUSTOM_MAIN
+ "${CMAKE_SOURCE_DIR}/Userland/Libraries/LibTest/TestMain.cpp")
+ endif()
+ if (NOT ${SERENITY_TEST_MAIN_ALREADY_DEFINED})
+ list(PREPEND TEST_SOURCES "${SERENITY_TEST_CUSTOM_MAIN}")
endif()
get_filename_component(test_name ${test_src} NAME_WE)
add_executable(${test_name} ${TEST_SOURCES})
@@ -84,6 +88,19 @@ function(serenity_test test_src sub_dir)
install(TARGETS ${test_name} RUNTIME DESTINATION usr/Tests/${sub_dir})
endfunction()
+
+function(serenity_testjs_test test_src sub_dir)
+ cmake_parse_arguments(SERENITY_TEST "" "CUSTOM_MAIN" "LIBS" ${ARGN})
+ if ("${SERENITY_TEST_CUSTOM_MAIN}" STREQUAL "")
+ set(SERENITY_TEST_CUSTOM_MAIN
+ "${CMAKE_SOURCE_DIR}/Userland/Libraries/LibTest/JavaScriptTestRunnerMain.cpp")
+ endif()
+ list(APPEND SERENITY_TEST_LIBS LibJS LibCore)
+ serenity_test(${test_src} ${sub_dir}
+ CUSTOM_MAIN "${SERENITY_TEST_CUSTOM_MAIN}"
+ LIBS ${SERENITY_TEST_LIBS})
+endfunction()
+
function(serenity_app target_name)
cmake_parse_arguments(SERENITY_APP "" "ICON" "" ${ARGN})
diff --git a/Meta/Lagom/CMakeLists.txt b/Meta/Lagom/CMakeLists.txt
index 52fb4346aa..b071468b2f 100644
--- a/Meta/Lagom/CMakeLists.txt
+++ b/Meta/Lagom/CMakeLists.txt
@@ -132,7 +132,9 @@ if (BUILD_LAGOM)
set_target_properties(ntpquery_lagom PROPERTIES OUTPUT_NAME ntpquery)
target_link_libraries(ntpquery_lagom Lagom)
- add_executable(test-js_lagom ../../Tests/LibJS/test-js.cpp)
+ add_executable(test-js_lagom
+ ../../Tests/LibJS/test-js.cpp
+ ../../Userland/Libraries/LibTest/JavaScriptTestRunnerMain.cpp)
set_target_properties(test-js_lagom PROPERTIES OUTPUT_NAME test-js)
target_link_libraries(test-js_lagom Lagom)
target_link_libraries(test-js_lagom stdc++)
diff --git a/Tests/LibJS/CMakeLists.txt b/Tests/LibJS/CMakeLists.txt
index 14a5c6d031..8d425f8650 100644
--- a/Tests/LibJS/CMakeLists.txt
+++ b/Tests/LibJS/CMakeLists.txt
@@ -1,3 +1,2 @@
-add_executable(test-js test-js.cpp)
-target_link_libraries(test-js LibJS LibLine LibCore)
+serenity_testjs_test(test-js.cpp test-js)
install(TARGETS test-js RUNTIME DESTINATION bin)
diff --git a/Tests/LibJS/test-js.cpp b/Tests/LibJS/test-js.cpp
index 095e927584..129c2149e4 100644
--- a/Tests/LibJS/test-js.cpp
+++ b/Tests/LibJS/test-js.cpp
@@ -5,135 +5,16 @@
* SPDX-License-Identifier: BSD-2-Clause
*/
-#include <AK/ByteBuffer.h>
-#include <AK/JsonObject.h>
-#include <AK/JsonValue.h>
-#include <AK/LexicalPath.h>
-#include <AK/QuickSort.h>
-#include <LibCore/ArgsParser.h>
-#include <LibCore/DirIterator.h>
-#include <LibCore/File.h>
-#include <LibJS/Interpreter.h>
-#include <LibJS/Lexer.h>
-#include <LibJS/Parser.h>
-#include <LibJS/Runtime/Array.h>
-#include <LibJS/Runtime/GlobalObject.h>
-#include <LibJS/Runtime/JSONObject.h>
-#include <LibTest/Results.h>
-#include <signal.h>
-#include <stdlib.h>
-#include <sys/time.h>
-#include <unistd.h>
+#include <LibTest/JavaScriptTestRunner.h>
-#define TOP_LEVEL_TEST_NAME "__$$TOP_LEVEL$$__"
+TEST_ROOT("Userland/Libraries/LibJS/Tests");
-RefPtr<JS::VM> vm;
-
-static bool collect_on_every_allocation = false;
-static String currently_running_test;
-
-struct ParserError {
- JS::Parser::Error error;
- String hint;
-};
-
-struct JSFileResult {
- String name;
- Optional<ParserError> error {};
- double time_taken { 0 };
- // A failed test takes precedence over a skipped test, which both have
- // precedence over a passed test
- Test::Result most_severe_test_result { Test::Result::Pass };
- Vector<Test::Suite> suites {};
- Vector<String> logged_messages {};
-};
-
-class TestRunnerGlobalObject final : public JS::GlobalObject {
- JS_OBJECT(TestRunnerGlobalObject, JS::GlobalObject);
-
-public:
- TestRunnerGlobalObject();
- virtual ~TestRunnerGlobalObject() override;
-
- virtual void initialize_global_object() override;
-
-private:
- JS_DECLARE_NATIVE_FUNCTION(is_strict_mode);
- JS_DECLARE_NATIVE_FUNCTION(can_parse_source);
- JS_DECLARE_NATIVE_FUNCTION(run_queued_promise_jobs);
-};
-
-class TestRunner {
-public:
- static TestRunner* the()
- {
- return s_the;
- }
-
- TestRunner(String test_root, bool print_times, bool print_progress)
- : m_test_root(move(test_root))
- , m_print_times(print_times)
- , m_print_progress(print_progress)
- {
- VERIFY(!s_the);
- s_the = this;
- }
-
- virtual ~TestRunner() { s_the = nullptr; };
-
- Test::Counts run();
-
- const Test::Counts& counts() const { return m_counts; }
-
- bool is_printing_progress() const { return m_print_progress; }
-
-protected:
- static TestRunner* s_the;
-
- virtual Vector<String> get_test_paths() const;
- virtual JSFileResult run_file_test(const String& test_path);
- void print_file_result(const JSFileResult& file_result) const;
- void print_test_results() const;
-
- String m_test_root;
- bool m_print_times;
- bool m_print_progress;
-
- double m_total_elapsed_time_in_ms { 0 };
- Test::Counts m_counts;
-
- RefPtr<JS::Program> m_test_program;
-};
-
-TestRunner* TestRunner::s_the = nullptr;
-
-TestRunnerGlobalObject::TestRunnerGlobalObject()
-{
-}
-
-TestRunnerGlobalObject::~TestRunnerGlobalObject()
-{
-}
-
-void TestRunnerGlobalObject::initialize_global_object()
-{
- Base::initialize_global_object();
- static FlyString global_property_name { "global" };
- static FlyString is_strict_mode_property_name { "isStrictMode" };
- static FlyString can_parse_source_property_name { "canParseSource" };
- static FlyString run_queued_promise_jobs_property_name { "runQueuedPromiseJobs" };
- define_property(global_property_name, this, JS::Attribute::Enumerable);
- define_native_function(is_strict_mode_property_name, is_strict_mode);
- define_native_function(can_parse_source_property_name, can_parse_source);
- define_native_function(run_queued_promise_jobs_property_name, run_queued_promise_jobs);
-}
-
-JS_DEFINE_NATIVE_FUNCTION(TestRunnerGlobalObject::is_strict_mode)
+TESTJS_GLOBAL_FUNCTION(is_strict_mode, isStrictMode, 0)
{
return JS::Value(vm.in_strict_mode());
}
-JS_DEFINE_NATIVE_FUNCTION(TestRunnerGlobalObject::can_parse_source)
+TESTJS_GLOBAL_FUNCTION(can_parse_source, canParseSource)
{
auto source = vm.argument(0).to_string(global_object);
if (vm.exception())
@@ -143,611 +24,8 @@ JS_DEFINE_NATIVE_FUNCTION(TestRunnerGlobalObject::can_parse_source)
return JS::Value(!parser.has_errors());
}
-JS_DEFINE_NATIVE_FUNCTION(TestRunnerGlobalObject::run_queued_promise_jobs)
+TESTJS_GLOBAL_FUNCTION(run_queued_promise_jobs, runQueuedPromiseJobs)
{
vm.run_queued_promise_jobs();
return JS::js_undefined();
}
-
-static void cleanup_and_exit()
-{
- // Clear the taskbar progress.
- if (TestRunner::the() && TestRunner::the()->is_printing_progress())
- warn("\033]9;-1;\033\\");
- exit(1);
-}
-
-static void handle_sigabrt(int)
-{
- dbgln("test-js: SIGABRT received, cleaning up.");
- cleanup_and_exit();
-}
-
-static double get_time_in_ms()
-{
- struct timeval tv1;
- auto return_code = gettimeofday(&tv1, nullptr);
- VERIFY(return_code >= 0);
- return static_cast<double>(tv1.tv_sec) * 1000.0 + static_cast<double>(tv1.tv_usec) / 1000.0;
-}
-
-template<typename Callback>
-static void iterate_directory_recursively(const String& directory_path, Callback callback)
-{
- Core::DirIterator directory_iterator(directory_path, Core::DirIterator::Flags::SkipDots);
-
- while (directory_iterator.has_next()) {
- auto file_path = directory_iterator.next_full_path();
- if (Core::File::is_directory(file_path)) {
- iterate_directory_recursively(file_path, callback);
- } else {
- callback(move(file_path));
- }
- }
-}
-
-Vector<String> TestRunner::get_test_paths() const
-{
- Vector<String> paths;
- iterate_directory_recursively(m_test_root, [&](const String& file_path) {
- if (!file_path.ends_with("test-common.js"))
- paths.append(file_path);
- });
- quick_sort(paths);
- return paths;
-}
-
-Test::Counts TestRunner::run()
-{
- size_t progress_counter = 0;
- auto test_paths = get_test_paths();
- for (auto& path : test_paths) {
- ++progress_counter;
- print_file_result(run_file_test(path));
- if (m_print_progress)
- warn("\033]9;{};{};\033\\", progress_counter, test_paths.size());
- }
-
- if (m_print_progress)
- warn("\033]9;-1;\033\\");
-
- print_test_results();
-
- return m_counts;
-}
-
-static Result<NonnullRefPtr<JS::Program>, ParserError> parse_file(const String& file_path)
-{
- auto file = Core::File::construct(file_path);
- auto result = file->open(Core::OpenMode::ReadOnly);
- if (!result) {
- warnln("Failed to open the following file: \"{}\"", file_path);
- cleanup_and_exit();
- }
-
- auto contents = file->read_all();
- String test_file_string(reinterpret_cast<const char*>(contents.data()), contents.size());
- file->close();
-
- auto parser = JS::Parser(JS::Lexer(test_file_string));
- auto program = parser.parse_program();
-
- if (parser.has_errors()) {
- auto error = parser.errors()[0];
- return Result<NonnullRefPtr<JS::Program>, ParserError>(ParserError { error, error.source_location_hint(test_file_string) });
- }
-
- return Result<NonnullRefPtr<JS::Program>, ParserError>(program);
-}
-
-static Optional<JsonValue> get_test_results(JS::Interpreter& interpreter)
-{
- auto result = vm->get_variable("__TestResults__", interpreter.global_object());
- auto json_string = JS::JSONObject::stringify_impl(interpreter.global_object(), result, JS::js_undefined(), JS::js_undefined());
-
- auto json = JsonValue::from_string(json_string);
- if (!json.has_value())
- return {};
-
- return json.value();
-}
-
-JSFileResult TestRunner::run_file_test(const String& test_path)
-{
- currently_running_test = test_path;
-
- double start_time = get_time_in_ms();
- auto interpreter = JS::Interpreter::create<TestRunnerGlobalObject>(*vm);
-
- // FIXME: This is a hack while we're refactoring Interpreter/VM stuff.
- JS::VM::InterpreterExecutionScope scope(*interpreter);
-
- interpreter->heap().set_should_collect_on_every_allocation(collect_on_every_allocation);
-
- if (!m_test_program) {
- auto result = parse_file(String::formatted("{}/test-common.js", m_test_root));
- if (result.is_error()) {
- warnln("Unable to parse test-common.js");
- warnln("{}", result.error().error.to_string());
- warnln("{}", result.error().hint);
- cleanup_and_exit();
- }
- m_test_program = result.value();
- }
-
- interpreter->run(interpreter->global_object(), *m_test_program);
-
- auto file_program = parse_file(test_path);
- if (file_program.is_error())
- return { test_path, file_program.error() };
- interpreter->run(interpreter->global_object(), *file_program.value());
-
- auto test_json = get_test_results(*interpreter);
- if (!test_json.has_value()) {
- warnln("Received malformed JSON from test \"{}\"", test_path);
- cleanup_and_exit();
- }
-
- JSFileResult file_result { test_path.substring(m_test_root.length() + 1, test_path.length() - m_test_root.length() - 1) };
-
- // Collect logged messages
- auto& arr = interpreter->vm().get_variable("__UserOutput__", interpreter->global_object()).as_array();
- for (auto& entry : arr.indexed_properties()) {
- auto message = entry.value_and_attributes(&interpreter->global_object()).value;
- file_result.logged_messages.append(message.to_string_without_side_effects());
- }
-
- test_json.value().as_object().for_each_member([&](const String& suite_name, const JsonValue& suite_value) {
- Test::Suite suite { suite_name };
-
- VERIFY(suite_value.is_object());
-
- suite_value.as_object().for_each_member([&](const String& test_name, const JsonValue& test_value) {
- Test::Case test { test_name, Test::Result::Fail, "" };
-
- VERIFY(test_value.is_object());
- VERIFY(test_value.as_object().has("result"));
-
- auto result = test_value.as_object().get("result");
- VERIFY(result.is_string());
- auto result_string = result.as_string();
- if (result_string == "pass") {
- test.result = Test::Result::Pass;
- m_counts.tests_passed++;
- } else if (result_string == "fail") {
- test.result = Test::Result::Fail;
- m_counts.tests_failed++;
- suite.most_severe_test_result = Test::Result::Fail;
- VERIFY(test_value.as_object().has("details"));
- auto details = test_value.as_object().get("details");
- VERIFY(result.is_string());
- test.details = details.as_string();
- } else {
- test.result = Test::Result::Skip;
- if (suite.most_severe_test_result == Test::Result::Pass)
- suite.most_severe_test_result = Test::Result::Skip;
- m_counts.tests_skipped++;
- }
-
- suite.tests.append(test);
- });
-
- if (suite.most_severe_test_result == Test::Result::Fail) {
- m_counts.suites_failed++;
- file_result.most_severe_test_result = Test::Result::Fail;
- } else {
- if (suite.most_severe_test_result == Test::Result::Skip && file_result.most_severe_test_result == Test::Result::Pass)
- file_result.most_severe_test_result = Test::Result::Skip;
- m_counts.suites_passed++;
- }
-
- file_result.suites.append(suite);
- });
-
- m_counts.files_total++;
-
- file_result.time_taken = get_time_in_ms() - start_time;
- m_total_elapsed_time_in_ms += file_result.time_taken;
-
- return file_result;
-}
-
-enum Modifier {
- BG_RED,
- BG_GREEN,
- FG_RED,
- FG_GREEN,
- FG_ORANGE,
- FG_GRAY,
- FG_BLACK,
- FG_BOLD,
- ITALIC,
- CLEAR,
-};
-
-static void print_modifiers(Vector<Modifier> modifiers)
-{
- for (auto& modifier : modifiers) {
- auto code = [&] {
- switch (modifier) {
- case BG_RED:
- return "\033[48;2;255;0;102m";
- case BG_GREEN:
- return "\033[48;2;102;255;0m";
- case FG_RED:
- return "\033[38;2;255;0;102m";
- case FG_GREEN:
- return "\033[38;2;102;255;0m";
- case FG_ORANGE:
- return "\033[38;2;255;102;0m";
- case FG_GRAY:
- return "\033[38;2;135;139;148m";
- case FG_BLACK:
- return "\033[30m";
- case FG_BOLD:
- return "\033[1m";
- case ITALIC:
- return "\033[3m";
- case CLEAR:
- return "\033[0m";
- }
- VERIFY_NOT_REACHED();
- }();
- out("{}", code);
- }
-}
-
-void TestRunner::print_file_result(const JSFileResult& file_result) const
-{
- if (file_result.most_severe_test_result == Test::Result::Fail || file_result.error.has_value()) {
- print_modifiers({ BG_RED, FG_BLACK, FG_BOLD });
- out(" FAIL ");
- print_modifiers({ CLEAR });
- } else {
- if (m_print_times || file_result.most_severe_test_result != Test::Result::Pass) {
- print_modifiers({ BG_GREEN, FG_BLACK, FG_BOLD });
- out(" PASS ");
- print_modifiers({ CLEAR });
- } else {
- return;
- }
- }
-
- out(" {}", file_result.name);
-
- if (m_print_times) {
- print_modifiers({ CLEAR, ITALIC, FG_GRAY });
- if (file_result.time_taken < 1000) {
- outln(" ({}ms)", static_cast<int>(file_result.time_taken));
- } else {
- outln(" ({:3}s)", file_result.time_taken / 1000.0);
- }
- print_modifiers({ CLEAR });
- } else {
- outln();
- }
-
- if (!file_result.logged_messages.is_empty()) {
- print_modifiers({ FG_GRAY, FG_BOLD });
-#ifdef __serenity__
- outln(" ℹ Console output:");
-#else
- // This emoji has a second invisible byte after it. The one above does not
- outln(" ℹī¸ Console output:");
-#endif
- print_modifiers({ CLEAR, FG_GRAY });
- for (auto& message : file_result.logged_messages)
- outln(" {}", message);
- }
-
- if (file_result.error.has_value()) {
- auto test_error = file_result.error.value();
-
- print_modifiers({ FG_RED });
-#ifdef __serenity__
- outln(" ❌ The file failed to parse");
-#else
- // No invisible byte here, but the spacing still needs to be altered on the host
- outln(" ❌ The file failed to parse");
-#endif
- outln();
- print_modifiers({ FG_GRAY });
- for (auto& message : test_error.hint.split('\n', true)) {
- outln(" {}", message);
- }
- print_modifiers({ FG_RED });
- outln(" {}", test_error.error.to_string());
- outln();
- return;
- }
-
- if (file_result.most_severe_test_result != Test::Result::Pass) {
- for (auto& suite : file_result.suites) {
- if (suite.most_severe_test_result == Test::Result::Pass)
- continue;
-
- bool failed = suite.most_severe_test_result == Test::Result::Fail;
-
- print_modifiers({ FG_GRAY, FG_BOLD });
-
- if (failed) {
-#ifdef __serenity__
- out(" ❌ Suite: ");
-#else
- // No invisible byte here, but the spacing still needs to be altered on the host
- out(" ❌ Suite: ");
-#endif
- } else {
-#ifdef __serenity__
- out(" ⚠ Suite: ");
-#else
- // This emoji has a second invisible byte after it. The one above does not
- out(" ⚠ī¸ Suite: ");
-#endif
- }
-
- print_modifiers({ CLEAR, FG_GRAY });
-
- if (suite.name == TOP_LEVEL_TEST_NAME) {
- outln("<top-level>");
- } else {
- outln("{}", suite.name);
- }
- print_modifiers({ CLEAR });
-
- for (auto& test : suite.tests) {
- if (test.result == Test::Result::Pass)
- continue;
-
- print_modifiers({ FG_GRAY, FG_BOLD });
- out(" Test: ");
- if (test.result == Test::Result::Fail) {
- print_modifiers({ CLEAR, FG_RED });
- outln("{} (failed):", test.name);
- outln(" {}", test.details);
- } else {
- print_modifiers({ CLEAR, FG_ORANGE });
- outln("{} (skipped)", test.name);
- }
- print_modifiers({ CLEAR });
- }
- }
- }
-}
-
-void TestRunner::print_test_results() const
-{
- out("\nTest Suites: ");
- if (m_counts.suites_failed) {
- print_modifiers({ FG_RED });
- out("{} failed, ", m_counts.suites_failed);
- print_modifiers({ CLEAR });
- }
- if (m_counts.suites_passed) {
- print_modifiers({ FG_GREEN });
- out("{} passed, ", m_counts.suites_passed);
- print_modifiers({ CLEAR });
- }
- outln("{} total", m_counts.suites_failed + m_counts.suites_passed);
-
- out("Tests: ");
- if (m_counts.tests_failed) {
- print_modifiers({ FG_RED });
- out("{} failed, ", m_counts.tests_failed);
- print_modifiers({ CLEAR });
- }
- if (m_counts.tests_skipped) {
- print_modifiers({ FG_ORANGE });
- out("{} skipped, ", m_counts.tests_skipped);
- print_modifiers({ CLEAR });
- }
- if (m_counts.tests_passed) {
- print_modifiers({ FG_GREEN });
- out("{} passed, ", m_counts.tests_passed);
- print_modifiers({ CLEAR });
- }
- outln("{} total", m_counts.tests_failed + m_counts.tests_skipped + m_counts.tests_passed);
-
- outln("Files: {} total", m_counts.files_total);
-
- out("Time: ");
- if (m_total_elapsed_time_in_ms < 1000.0) {
- outln("{}ms", static_cast<int>(m_total_elapsed_time_in_ms));
- } else {
- outln("{:>.3}s", m_total_elapsed_time_in_ms / 1000.0);
- }
- outln();
-}
-
-class Test262ParserTestRunner final : public TestRunner {
-public:
- using TestRunner::TestRunner;
-
-private:
- virtual Vector<String> get_test_paths() const override;
- virtual JSFileResult run_file_test(const String& test_path) override;
-};
-
-Vector<String> Test262ParserTestRunner::get_test_paths() const
-{
- Vector<String> paths;
- iterate_directory_recursively(m_test_root, [&](const String& file_path) {
- auto dirname = LexicalPath(file_path).dirname();
- if (dirname.ends_with("early") || dirname.ends_with("fail") || dirname.ends_with("pass") || dirname.ends_with("pass-explicit"))
- paths.append(file_path);
- });
- quick_sort(paths);
- return paths;
-}
-
-JSFileResult Test262ParserTestRunner::run_file_test(const String& test_path)
-{
- currently_running_test = test_path;
-
- auto dirname = LexicalPath(test_path).dirname();
- bool expecting_file_to_parse;
- if (dirname.ends_with("early") || dirname.ends_with("fail")) {
- expecting_file_to_parse = false;
- } else if (dirname.ends_with("pass") || dirname.ends_with("pass-explicit")) {
- expecting_file_to_parse = true;
- } else {
- VERIFY_NOT_REACHED();
- }
-
- auto start_time = get_time_in_ms();
- String details = "";
- Test::Result test_result;
- if (test_path.ends_with(".module.js")) {
- test_result = Test::Result::Skip;
- m_counts.tests_skipped++;
- m_counts.suites_passed++;
- } else {
- auto parse_result = parse_file(test_path);
- if (expecting_file_to_parse) {
- if (!parse_result.is_error()) {
- test_result = Test::Result::Pass;
- } else {
- test_result = Test::Result::Fail;
- details = parse_result.error().error.to_string();
- }
- } else {
- if (parse_result.is_error()) {
- test_result = Test::Result::Pass;
- } else {
- test_result = Test::Result::Fail;
- details = "File was expected to produce a parser error but didn't";
- }
- }
- }
-
- // test262-parser-tests doesn't have "suites" and "tests" in the usual sense, it just has files
- // and an expectation whether they should parse or not. We add one suite with one test nonetheless:
- //
- // - This makes interpreting skipped test easier as their file is shown as "PASS"
- // - That way we can show additional information such as "file parsed but shouldn't have" or
- // parser errors for files that should parse respectively
-
- Test::Case test { expecting_file_to_parse ? "file should parse" : "file should not parse", test_result, details };
- Test::Suite suite { "Parse file", test_result, { test } };
- JSFileResult file_result {
- test_path.substring(m_test_root.length() + 1, test_path.length() - m_test_root.length() - 1),
- {},
- get_time_in_ms() - start_time,
- test_result,
- { suite }
- };
-
- if (test_result == Test::Result::Fail) {
- m_counts.tests_failed++;
- m_counts.suites_failed++;
- } else {
- m_counts.tests_passed++;
- m_counts.suites_passed++;
- }
- m_counts.files_total++;
- m_total_elapsed_time_in_ms += file_result.time_taken;
-
- return file_result;
-}
-
-int main(int argc, char** argv)
-{
- struct sigaction act;
- memset(&act, 0, sizeof(act));
- act.sa_flags = SA_NOCLDWAIT;
- act.sa_handler = handle_sigabrt;
- int rc = sigaction(SIGABRT, &act, nullptr);
- if (rc < 0) {
- perror("sigaction");
- return 1;
- }
-
-#ifdef SIGINFO
- signal(SIGINFO, [](int) {
- static char buffer[4096];
- auto& counts = TestRunner::the()->counts();
- int len = snprintf(buffer, sizeof(buffer), "Pass: %d, Fail: %d, Skip: %d\nCurrent test: %s\n", counts.tests_passed, counts.tests_failed, counts.tests_skipped, currently_running_test.characters());
- write(STDOUT_FILENO, buffer, len);
- });
-#endif
-
- bool print_times = false;
- bool print_progress =
-#ifdef __serenity__
- true; // Use OSC 9 to print progress
-#else
- false;
-#endif
- bool test262_parser_tests = false;
- const char* specified_test_root = nullptr;
-
- Core::ArgsParser args_parser;
- args_parser.add_option(print_times, "Show duration of each test", "show-time", 't');
- args_parser.add_option(Core::ArgsParser::Option {
- .requires_argument = true,
- .help_string = "Show progress with OSC 9 (true, false)",
- .long_name = "show-progress",
- .short_name = 'p',
- .accept_value = [&](auto* str) {
- if (StringView { "true" } == str)
- print_progress = true;
- else if (StringView { "false" } == str)
- print_progress = false;
- else
- return false;
- return true;
- },
- });
- args_parser.add_option(collect_on_every_allocation, "Collect garbage after every allocation", "collect-often", 'g');
- args_parser.add_option(test262_parser_tests, "Run test262 parser tests", "test262-parser-tests", 0);
- args_parser.add_positional_argument(specified_test_root, "Tests root directory", "path", Core::ArgsParser::Required::No);
- args_parser.parse(argc, argv);
-
- if (test262_parser_tests) {
- if (collect_on_every_allocation) {
- warnln("--collect-often and --test262-parser-tests options must not be used together");
- return 1;
- }
- if (!specified_test_root) {
- warnln("Test root is required with --test262-parser-tests");
- return 1;
- }
- }
-
- if (getenv("DISABLE_DBG_OUTPUT")) {
- AK::set_debug_enabled(false);
- }
-
- String test_root;
-
- if (specified_test_root) {
- test_root = String { specified_test_root };
- } else {
-#ifdef __serenity__
- test_root = "/home/anon/js-tests";
-#else
- char* serenity_source_dir = getenv("SERENITY_SOURCE_DIR");
- if (!serenity_source_dir) {
- warnln("No test root given, test-js requires the SERENITY_SOURCE_DIR environment variable to be set");
- return 1;
- }
- test_root = String::formatted("{}/Userland/Libraries/LibJS/Tests", serenity_source_dir);
-#endif
- }
- if (!Core::File::is_directory(test_root)) {
- warnln("Test root is not a directory: {}", test_root);
- return 1;
- }
-
- vm = JS::VM::create();
-
- Test::Counts result_counts;
- if (test262_parser_tests)
- result_counts = Test262ParserTestRunner(test_root, print_times, print_progress).run();
- else
- result_counts = TestRunner(test_root, print_times, print_progress).run();
-
- vm = nullptr;
-
- return result_counts.tests_failed > 0 ? 1 : 0;
-}
diff --git a/Userland/Libraries/LibTest/JavaScriptTestRunner.h b/Userland/Libraries/LibTest/JavaScriptTestRunner.h
new file mode 100644
index 0000000000..1aa788badf
--- /dev/null
+++ b/Userland/Libraries/LibTest/JavaScriptTestRunner.h
@@ -0,0 +1,591 @@
+/*
+ * Copyright (c) 2020, Matthew Olsson <mattco@serenityos.org>
+ * Copyright (c) 2020-2021, Linus Groh <linusg@serenityos.org>
+ * Copyright (c) 2021, Ali Mohammad Pur <mpfard@serenityos.org>
+ *
+ * SPDX-License-Identifier: BSD-2-Clause
+ */
+
+#pragma once
+
+#include <AK/ByteBuffer.h>
+#include <AK/JsonObject.h>
+#include <AK/JsonValue.h>
+#include <AK/LexicalPath.h>
+#include <AK/QuickSort.h>
+#include <AK/Result.h>
+#include <AK/Tuple.h>
+#include <LibCore/ArgsParser.h>
+#include <LibCore/DirIterator.h>
+#include <LibCore/File.h>
+#include <LibJS/Interpreter.h>
+#include <LibJS/Lexer.h>
+#include <LibJS/Parser.h>
+#include <LibJS/Runtime/Array.h>
+#include <LibJS/Runtime/GlobalObject.h>
+#include <LibJS/Runtime/JSONObject.h>
+#include <LibJS/Runtime/TypedArray.h>
+#include <LibTest/Results.h>
+#include <sys/time.h>
+#include <unistd.h>
+
+#define STRCAT(x, y) __STRCAT(x, y)
+#define STRSTRCAT(x, y) __STRSTRCAT(x, y)
+#define __STRCAT(x, y) x #y
+#define __STRSTRCAT(x, y) x y
+
+// Note: This is a little weird, so here's an explanation:
+// If the vararg isn't given, the tuple initializer will simply expand to `fn, ::Test::JS::__testjs_last<1>()`
+// and if it _is_ given (say as `A`), the tuple initializer will expand to `fn, ::Test::JS::__testjs_last<1, A>()`, which will end up being evaluated as `A`
+// and if multiple args are given, the static_assert will be sad.
+#define __TESTJS_REGISTER_GLOBAL_FUNCTION(name, fn, ...) \
+ struct __TestJS_register_##fn { \
+ static_assert( \
+ ::Test::JS::__testjs_count(__VA_ARGS__) <= 1, \
+ STRCAT(STRSTRCAT(STRCAT("Expected at most three arguments to TESTJS_GLOBAL_FUNCTION at line", __LINE__), ", in file "), __FILE__)); \
+ __TestJS_register_##fn() noexcept \
+ { \
+ ::Test::JS::s_exposed_global_functions.set( \
+ name, \
+ { fn, ::Test::JS::__testjs_last<1, ##__VA_ARGS__>() }); \
+ } \
+ } __testjs_register_##fn {};
+
+#define TESTJS_GLOBAL_FUNCTION(function, exposed_name, ...) \
+ JS_DECLARE_NATIVE_FUNCTION(function); \
+ __TESTJS_REGISTER_GLOBAL_FUNCTION(#exposed_name, function, ##__VA_ARGS__); \
+ JS_DEFINE_NATIVE_FUNCTION(function)
+
+#define TESTJS_MAIN_HOOK() \
+ struct __TestJS_main_hook { \
+ __TestJS_main_hook() \
+ { \
+ ::Test::JS::g_main_hook = hook; \
+ } \
+ static void hook(); \
+ } __testjs_common_register_##name {}; \
+ void __TestJS_main_hook::hook()
+
+#define TEST_ROOT(path) \
+ String Test::JS::g_test_root_fragment = path
+
+namespace Test::JS {
+
+namespace JS = ::JS;
+
+template<typename... Args>
+static consteval size_t __testjs_count(Args...) { return sizeof...(Args); }
+
+template<auto... Values>
+static consteval size_t __testjs_last() { return (AK::Detail::IntegralConstant<size_t, Values> {}, ...).value; }
+
+static constexpr auto TOP_LEVEL_TEST_NAME = "__$$TOP_LEVEL$$__";
+extern RefPtr<JS::VM> g_vm;
+extern bool g_collect_on_every_allocation;
+extern String g_currently_running_test;
+extern String g_test_glob;
+struct FunctionWithLength {
+ JS::Value (*function)(JS::VM&, JS::GlobalObject&);
+ size_t length { 0 };
+};
+extern HashMap<String, FunctionWithLength> s_exposed_global_functions;
+extern String g_test_root_fragment;
+extern String g_test_root;
+extern int g_test_argc;
+extern char** g_test_argv;
+extern Function<void()> g_main_hook;
+
+struct ParserError {
+ JS::Parser::Error error;
+ String hint;
+};
+
+struct JSFileResult {
+ String name;
+ Optional<ParserError> error {};
+ double time_taken { 0 };
+ // A failed test takes precedence over a skipped test, which both have
+ // precedence over a passed test
+ Test::Result most_severe_test_result { Test::Result::Pass };
+ Vector<Test::Suite> suites {};
+ Vector<String> logged_messages {};
+};
+
+class TestRunner {
+public:
+ static TestRunner* the()
+ {
+ return s_the;
+ }
+
+ TestRunner(String test_root, String common_path, bool print_times, bool print_progress)
+ : m_common_path(move(common_path))
+ , m_test_root(move(test_root))
+ , m_print_times(print_times)
+ , m_print_progress(print_progress)
+ {
+ VERIFY(!s_the);
+ s_the = this;
+ g_test_root = m_test_root;
+ }
+
+ virtual ~TestRunner() = default;
+
+ void run();
+
+ const Test::Counts& counts() const { return m_counts; }
+
+ bool is_printing_progress() const { return m_print_progress; }
+
+protected:
+ static TestRunner* s_the;
+
+ virtual Vector<String> get_test_paths() const;
+ virtual JSFileResult run_file_test(const String& test_path);
+ void print_file_result(const JSFileResult& file_result) const;
+ void print_test_results() const;
+
+ String m_common_path;
+ String m_test_root;
+ bool m_print_times;
+ bool m_print_progress;
+
+ double m_total_elapsed_time_in_ms { 0 };
+ Test::Counts m_counts;
+
+ RefPtr<JS::Program> m_test_program;
+};
+
+class TestRunnerGlobalObject final : public JS::GlobalObject {
+ JS_OBJECT(TestRunnerGlobalObject, JS::GlobalObject);
+
+public:
+ TestRunnerGlobalObject() = default;
+ virtual ~TestRunnerGlobalObject() override = default;
+
+ virtual void initialize_global_object() override;
+};
+
+inline void TestRunnerGlobalObject::initialize_global_object()
+{
+ Base::initialize_global_object();
+ define_property("global", this, JS::Attribute::Enumerable);
+ for (auto& entry : s_exposed_global_functions) {
+ define_native_function(
+ entry.key, [fn = entry.value.function](auto& vm, auto& global_object) {
+ return fn(vm, global_object);
+ },
+ entry.value.length);
+ }
+}
+
+inline void cleanup_and_exit()
+{
+ // Clear the taskbar progress.
+ if (TestRunner::the() && TestRunner::the()->is_printing_progress())
+ warn("\033]9;-1;\033\\");
+ exit(1);
+}
+
+inline double get_time_in_ms()
+{
+ struct timeval tv1;
+ auto return_code = gettimeofday(&tv1, nullptr);
+ VERIFY(return_code >= 0);
+ return static_cast<double>(tv1.tv_sec) * 1000.0 + static_cast<double>(tv1.tv_usec) / 1000.0;
+}
+
+template<typename Callback>
+inline void iterate_directory_recursively(const String& directory_path, Callback callback)
+{
+ Core::DirIterator directory_iterator(directory_path, Core::DirIterator::Flags::SkipDots);
+
+ while (directory_iterator.has_next()) {
+ auto file_path = directory_iterator.next_full_path();
+ auto is_directory = Core::File::is_directory(file_path);
+ if (is_directory && !file_path.contains("/Fixtures")) {
+ iterate_directory_recursively(file_path, callback);
+ } else if (!is_directory) {
+ callback(move(file_path));
+ }
+ }
+}
+
+inline Vector<String> TestRunner::get_test_paths() const
+{
+ Vector<String> paths;
+ iterate_directory_recursively(m_test_root, [&](const String& file_path) {
+ if (!file_path.ends_with(".js"))
+ return;
+ if (!file_path.ends_with("test-common.js"))
+ paths.append(file_path);
+ });
+ quick_sort(paths);
+ return paths;
+}
+
+inline void TestRunner::run()
+{
+ size_t progress_counter = 0;
+ auto test_paths = get_test_paths();
+ for (auto& path : test_paths) {
+ if (!path.matches(g_test_glob))
+ continue;
+ ++progress_counter;
+ print_file_result(run_file_test(path));
+ if (m_print_progress)
+ warn("\033]9;{};{};\033\\", progress_counter, test_paths.size());
+ }
+
+ if (m_print_progress)
+ warn("\033]9;-1;\033\\");
+
+ print_test_results();
+}
+
+inline AK::Result<NonnullRefPtr<JS::Program>, ParserError> parse_file(const String& file_path)
+{
+ auto file = Core::File::construct(file_path);
+ auto result = file->open(Core::OpenMode::ReadOnly);
+ if (!result) {
+ warnln("Failed to open the following file: \"{}\"", file_path);
+ cleanup_and_exit();
+ }
+
+ auto contents = file->read_all();
+ String test_file_string(reinterpret_cast<const char*>(contents.data()), contents.size());
+ file->close();
+
+ auto parser = JS::Parser(JS::Lexer(test_file_string));
+ auto program = parser.parse_program();
+
+ if (parser.has_errors()) {
+ auto error = parser.errors()[0];
+ return AK::Result<NonnullRefPtr<JS::Program>, ParserError>(ParserError { error, error.source_location_hint(test_file_string) });
+ }
+
+ return AK::Result<NonnullRefPtr<JS::Program>, ParserError>(program);
+}
+
+inline Optional<JsonValue> get_test_results(JS::Interpreter& interpreter)
+{
+ auto result = g_vm->get_variable("__TestResults__", interpreter.global_object());
+ auto json_string = JS::JSONObject::stringify_impl(interpreter.global_object(), result, JS::js_undefined(), JS::js_undefined());
+
+ auto json = JsonValue::from_string(json_string);
+ if (!json.has_value())
+ return {};
+
+ return json.value();
+}
+
+inline JSFileResult TestRunner::run_file_test(const String& test_path)
+{
+ g_currently_running_test = test_path;
+
+ double start_time = get_time_in_ms();
+ auto interpreter = JS::Interpreter::create<TestRunnerGlobalObject>(*g_vm);
+
+ // FIXME: This is a hack while we're refactoring Interpreter/VM stuff.
+ JS::VM::InterpreterExecutionScope scope(*interpreter);
+
+ interpreter->heap().set_should_collect_on_every_allocation(g_collect_on_every_allocation);
+
+ if (!m_test_program) {
+ auto result = parse_file(m_common_path);
+ if (result.is_error()) {
+ warnln("Unable to parse test-common.js");
+ warnln("{}", result.error().error.to_string());
+ warnln("{}", result.error().hint);
+ cleanup_and_exit();
+ }
+ m_test_program = result.value();
+ }
+
+ interpreter->run(interpreter->global_object(), *m_test_program);
+
+ auto file_program = parse_file(test_path);
+ if (file_program.is_error())
+ return { test_path, file_program.error() };
+ interpreter->run(interpreter->global_object(), *file_program.value());
+
+ if (g_vm->exception())
+ g_vm->clear_exception();
+
+ auto test_json = get_test_results(*interpreter);
+ if (!test_json.has_value()) {
+ warnln("Received malformed JSON from test \"{}\"", test_path);
+ cleanup_and_exit();
+ }
+
+ JSFileResult file_result { test_path.substring(m_test_root.length() + 1, test_path.length() - m_test_root.length() - 1) };
+
+ // Collect logged messages
+ auto& arr = interpreter->vm().get_variable("__UserOutput__", interpreter->global_object()).as_array();
+ for (auto& entry : arr.indexed_properties()) {
+ auto message = entry.value_and_attributes(&interpreter->global_object()).value;
+ file_result.logged_messages.append(message.to_string_without_side_effects());
+ }
+
+ test_json.value().as_object().for_each_member([&](const String& suite_name, const JsonValue& suite_value) {
+ Test::Suite suite { suite_name };
+
+ VERIFY(suite_value.is_object());
+
+ suite_value.as_object().for_each_member([&](const String& test_name, const JsonValue& test_value) {
+ Test::Case test { test_name, Test::Result::Fail, "" };
+
+ VERIFY(test_value.is_object());
+ VERIFY(test_value.as_object().has("result"));
+
+ auto result = test_value.as_object().get("result");
+ VERIFY(result.is_string());
+ auto result_string = result.as_string();
+ if (result_string == "pass") {
+ test.result = Test::Result::Pass;
+ m_counts.tests_passed++;
+ } else if (result_string == "fail") {
+ test.result = Test::Result::Fail;
+ m_counts.tests_failed++;
+ suite.most_severe_test_result = Test::Result::Fail;
+ VERIFY(test_value.as_object().has("details"));
+ auto details = test_value.as_object().get("details");
+ VERIFY(result.is_string());
+ test.details = details.as_string();
+ } else {
+ test.result = Test::Result::Skip;
+ if (suite.most_severe_test_result == Test::Result::Pass)
+ suite.most_severe_test_result = Test::Result::Skip;
+ m_counts.tests_skipped++;
+ }
+
+ suite.tests.append(test);
+ });
+
+ if (suite.most_severe_test_result == Test::Result::Fail) {
+ m_counts.suites_failed++;
+ file_result.most_severe_test_result = Test::Result::Fail;
+ } else {
+ if (suite.most_severe_test_result == Test::Result::Skip && file_result.most_severe_test_result == Test::Result::Pass)
+ file_result.most_severe_test_result = Test::Result::Skip;
+ m_counts.suites_passed++;
+ }
+
+ file_result.suites.append(suite);
+ });
+
+ m_counts.files_total++;
+
+ file_result.time_taken = get_time_in_ms() - start_time;
+ m_total_elapsed_time_in_ms += file_result.time_taken;
+
+ return file_result;
+}
+
+enum Modifier {
+ BG_RED,
+ BG_GREEN,
+ FG_RED,
+ FG_GREEN,
+ FG_ORANGE,
+ FG_GRAY,
+ FG_BLACK,
+ FG_BOLD,
+ ITALIC,
+ CLEAR,
+};
+
+inline void print_modifiers(Vector<Modifier> modifiers)
+{
+ for (auto& modifier : modifiers) {
+ auto code = [&] {
+ switch (modifier) {
+ case BG_RED:
+ return "\033[48;2;255;0;102m";
+ case BG_GREEN:
+ return "\033[48;2;102;255;0m";
+ case FG_RED:
+ return "\033[38;2;255;0;102m";
+ case FG_GREEN:
+ return "\033[38;2;102;255;0m";
+ case FG_ORANGE:
+ return "\033[38;2;255;102;0m";
+ case FG_GRAY:
+ return "\033[38;2;135;139;148m";
+ case FG_BLACK:
+ return "\033[30m";
+ case FG_BOLD:
+ return "\033[1m";
+ case ITALIC:
+ return "\033[3m";
+ case CLEAR:
+ return "\033[0m";
+ }
+ VERIFY_NOT_REACHED();
+ }();
+ out("{}", code);
+ }
+}
+
+inline void TestRunner::print_file_result(const JSFileResult& file_result) const
+{
+ if (file_result.most_severe_test_result == Test::Result::Fail || file_result.error.has_value()) {
+ print_modifiers({ BG_RED, FG_BLACK, FG_BOLD });
+ out(" FAIL ");
+ print_modifiers({ CLEAR });
+ } else {
+ if (m_print_times || file_result.most_severe_test_result != Test::Result::Pass) {
+ print_modifiers({ BG_GREEN, FG_BLACK, FG_BOLD });
+ out(" PASS ");
+ print_modifiers({ CLEAR });
+ } else {
+ return;
+ }
+ }
+
+ out(" {}", file_result.name);
+
+ if (m_print_times) {
+ print_modifiers({ CLEAR, ITALIC, FG_GRAY });
+ if (file_result.time_taken < 1000) {
+ outln(" ({}ms)", static_cast<int>(file_result.time_taken));
+ } else {
+ outln(" ({:3}s)", file_result.time_taken / 1000.0);
+ }
+ print_modifiers({ CLEAR });
+ } else {
+ outln();
+ }
+
+ if (!file_result.logged_messages.is_empty()) {
+ print_modifiers({ FG_GRAY, FG_BOLD });
+#ifdef __serenity__
+ outln(" ℹ Console output:");
+#else
+ // This emoji has a second invisible byte after it. The one above does not
+ outln(" ℹī¸ Console output:");
+#endif
+ print_modifiers({ CLEAR, FG_GRAY });
+ for (auto& message : file_result.logged_messages)
+ outln(" {}", message);
+ }
+
+ if (file_result.error.has_value()) {
+ auto test_error = file_result.error.value();
+
+ print_modifiers({ FG_RED });
+#ifdef __serenity__
+ outln(" ❌ The file failed to parse");
+#else
+ // No invisible byte here, but the spacing still needs to be altered on the host
+ outln(" ❌ The file failed to parse");
+#endif
+ outln();
+ print_modifiers({ FG_GRAY });
+ for (auto& message : test_error.hint.split('\n', true)) {
+ outln(" {}", message);
+ }
+ print_modifiers({ FG_RED });
+ outln(" {}", test_error.error.to_string());
+ outln();
+ return;
+ }
+
+ if (file_result.most_severe_test_result != Test::Result::Pass) {
+ for (auto& suite : file_result.suites) {
+ if (suite.most_severe_test_result == Test::Result::Pass)
+ continue;
+
+ bool failed = suite.most_severe_test_result == Test::Result::Fail;
+
+ print_modifiers({ FG_GRAY, FG_BOLD });
+
+ if (failed) {
+#ifdef __serenity__
+ out(" ❌ Suite: ");
+#else
+ // No invisible byte here, but the spacing still needs to be altered on the host
+ out(" ❌ Suite: ");
+#endif
+ } else {
+#ifdef __serenity__
+ out(" ⚠ Suite: ");
+#else
+ // This emoji has a second invisible byte after it. The one above does not
+ out(" ⚠ī¸ Suite: ");
+#endif
+ }
+
+ print_modifiers({ CLEAR, FG_GRAY });
+
+ if (suite.name == TOP_LEVEL_TEST_NAME) {
+ outln("<top-level>");
+ } else {
+ outln("{}", suite.name);
+ }
+ print_modifiers({ CLEAR });
+
+ for (auto& test : suite.tests) {
+ if (test.result == Test::Result::Pass)
+ continue;
+
+ print_modifiers({ FG_GRAY, FG_BOLD });
+ out(" Test: ");
+ if (test.result == Test::Result::Fail) {
+ print_modifiers({ CLEAR, FG_RED });
+ outln("{} (failed):", test.name);
+ outln(" {}", test.details);
+ } else {
+ print_modifiers({ CLEAR, FG_ORANGE });
+ outln("{} (skipped)", test.name);
+ }
+ print_modifiers({ CLEAR });
+ }
+ }
+ }
+}
+
+inline void TestRunner::print_test_results() const
+{
+ out("\nTest Suites: ");
+ if (m_counts.suites_failed) {
+ print_modifiers({ FG_RED });
+ out("{} failed, ", m_counts.suites_failed);
+ print_modifiers({ CLEAR });
+ }
+ if (m_counts.suites_passed) {
+ print_modifiers({ FG_GREEN });
+ out("{} passed, ", m_counts.suites_passed);
+ print_modifiers({ CLEAR });
+ }
+ outln("{} total", m_counts.suites_failed + m_counts.suites_passed);
+
+ out("Tests: ");
+ if (m_counts.tests_failed) {
+ print_modifiers({ FG_RED });
+ out("{} failed, ", m_counts.tests_failed);
+ print_modifiers({ CLEAR });
+ }
+ if (m_counts.tests_skipped) {
+ print_modifiers({ FG_ORANGE });
+ out("{} skipped, ", m_counts.tests_skipped);
+ print_modifiers({ CLEAR });
+ }
+ if (m_counts.tests_passed) {
+ print_modifiers({ FG_GREEN });
+ out("{} passed, ", m_counts.tests_passed);
+ print_modifiers({ CLEAR });
+ }
+ outln("{} total", m_counts.tests_failed + m_counts.tests_skipped + m_counts.tests_passed);
+
+ outln("Files: {} total", m_counts.files_total);
+
+ out("Time: ");
+ if (m_total_elapsed_time_in_ms < 1000.0) {
+ outln("{}ms", static_cast<int>(m_total_elapsed_time_in_ms));
+ } else {
+ outln("{:>.3}s", m_total_elapsed_time_in_ms / 1000.0);
+ }
+ outln();
+}
+}
diff --git a/Userland/Libraries/LibTest/JavaScriptTestRunnerMain.cpp b/Userland/Libraries/LibTest/JavaScriptTestRunnerMain.cpp
new file mode 100644
index 0000000000..5e9889d4ca
--- /dev/null
+++ b/Userland/Libraries/LibTest/JavaScriptTestRunnerMain.cpp
@@ -0,0 +1,156 @@
+/*
+ * Copyright (c) 2020, Matthew Olsson <mattco@serenityos.org>
+ * Copyright (c) 2020-2021, Linus Groh <linusg@serenityos.org>
+ * Copyright (c) 2021, Ali Mohammad Pur <mpfard@serenityos.org>
+ *
+ * SPDX-License-Identifier: BSD-2-Clause
+ */
+
+#include <LibTest/JavaScriptTestRunner.h>
+#include <signal.h>
+#include <stdio.h>
+
+namespace Test::JS {
+
+RefPtr<::JS::VM> g_vm;
+bool g_collect_on_every_allocation = false;
+String g_currently_running_test;
+String g_test_glob;
+HashMap<String, FunctionWithLength> s_exposed_global_functions;
+Function<void()> g_main_hook;
+TestRunner* TestRunner::s_the = nullptr;
+String g_test_root;
+int g_test_argc;
+char** g_test_argv;
+
+}
+
+using namespace Test::JS;
+
+static StringView g_program_name { "test-js"sv };
+
+static void handle_sigabrt(int)
+{
+ dbgln("{}: SIGABRT received, cleaning up.", g_program_name);
+ cleanup_and_exit();
+}
+
+int main(int argc, char** argv)
+{
+ g_test_argc = argc;
+ g_test_argv = argv;
+ auto program_name = LexicalPath { argv[0] }.basename();
+ g_program_name = program_name;
+
+ struct sigaction act;
+ memset(&act, 0, sizeof(act));
+ act.sa_flags = SA_NOCLDWAIT;
+ act.sa_handler = handle_sigabrt;
+ int rc = sigaction(SIGABRT, &act, nullptr);
+ if (rc < 0) {
+ perror("sigaction");
+ return 1;
+ }
+
+#ifdef SIGINFO
+ signal(SIGINFO, [](int) {
+ static char buffer[4096];
+ auto& counts = TestRunner::the()->counts();
+ int len = snprintf(buffer, sizeof(buffer), "Pass: %d, Fail: %d, Skip: %d\nCurrent test: %s\n", counts.tests_passed, counts.tests_failed, counts.tests_skipped, g_currently_running_test.characters());
+ write(STDOUT_FILENO, buffer, len);
+ });
+#endif
+
+ bool print_times = false;
+ bool print_progress =
+#ifdef __serenity__
+ true; // Use OSC 9 to print progress
+#else
+ false;
+#endif
+ const char* specified_test_root = nullptr;
+ String common_path;
+
+ Core::ArgsParser args_parser;
+ args_parser.add_option(print_times, "Show duration of each test", "show-time", 't');
+ args_parser.add_option(Core::ArgsParser::Option {
+ .requires_argument = true,
+ .help_string = "Show progress with OSC 9 (true, false)",
+ .long_name = "show-progress",
+ .short_name = 'p',
+ .accept_value = [&](auto* str) {
+ if (StringView { "true" } == str)
+ print_progress = true;
+ else if (StringView { "false" } == str)
+ print_progress = false;
+ else
+ return false;
+ return true;
+ },
+ });
+ args_parser.add_option(g_collect_on_every_allocation, "Collect garbage after every allocation", "collect-often", 'g');
+ args_parser.add_option(g_test_glob, "Only run tests matching the given glob", "filter", 'f', "glob");
+ args_parser.add_positional_argument(specified_test_root, "Tests root directory", "path", Core::ArgsParser::Required::No);
+ args_parser.add_positional_argument(common_path, "Path to tests-common.js", "common-path", Core::ArgsParser::Required::No);
+ args_parser.parse(argc, argv);
+
+ g_test_glob = String::formatted("*{}*", g_test_glob);
+
+ if (getenv("DISABLE_DBG_OUTPUT")) {
+ AK::set_debug_enabled(false);
+ }
+
+ String test_root;
+
+ if (specified_test_root) {
+ test_root = String { specified_test_root };
+ } else {
+#ifdef __serenity__
+ test_root = LexicalPath::join("/home/anon", String::formatted("{}-tests", program_name.split_view('-').last())).string();
+#else
+ char* serenity_source_dir = getenv("SERENITY_SOURCE_DIR");
+ if (!serenity_source_dir) {
+ warnln("No test root given, {} requires the SERENITY_SOURCE_DIR environment variable to be set", g_program_name);
+ return 1;
+ }
+ test_root = String::formatted("{}/{}", serenity_source_dir, g_test_root_fragment);
+ common_path = String::formatted("{}/Userland/Libraries/LibJS/Tests/test-common.js", serenity_source_dir);
+#endif
+ }
+ if (!Core::File::is_directory(test_root)) {
+ warnln("Test root is not a directory: {}", test_root);
+ return 1;
+ }
+
+ if (common_path.is_empty()) {
+#ifdef __serenity__
+ common_path = "/home/anon/js-tests/test-common.js";
+#else
+ char* serenity_source_dir = getenv("SERENITY_SOURCE_DIR");
+ if (!serenity_source_dir) {
+ warnln("No test root given, {} requires the SERENITY_SOURCE_DIR environment variable to be set", g_program_name);
+ return 1;
+ }
+ common_path = String::formatted("{}/Userland/Libraries/LibJS/Tests/test-common.js", serenity_source_dir);
+#endif
+ }
+
+ if (chdir(test_root.characters()) < 0) {
+ auto saved_errno = errno;
+ warnln("chdir failed: {}", strerror(saved_errno));
+ return 1;
+ }
+
+ if (g_main_hook)
+ g_main_hook();
+
+ if (!g_vm)
+ g_vm = JS::VM::create();
+
+ TestRunner test_runner(test_root, common_path, print_times, print_progress);
+ test_runner.run();
+
+ g_vm = nullptr;
+
+ return test_runner.counts().tests_failed > 0 ? 1 : 0;
+}