summaryrefslogtreecommitdiff
path: root/Userland/Utilities/test-js.cpp
diff options
context:
space:
mode:
authorAndreas Kling <kling@serenityos.org>2021-01-12 11:57:58 +0100
committerAndreas Kling <kling@serenityos.org>2021-01-12 12:04:09 +0100
commitececac65c23ef06243814725c49553ca94f676a2 (patch)
treea3f477d840bde6858d243bf57c4bcd956c97698c /Userland/Utilities/test-js.cpp
parentc4e2fd8123b83a6b890b410741f1d416a1162fdb (diff)
downloadserenity-ececac65c23ef06243814725c49553ca94f676a2.zip
Userland: Move command-line utilities to Userland/Utilities/
Diffstat (limited to 'Userland/Utilities/test-js.cpp')
-rw-r--r--Userland/Utilities/test-js.cpp765
1 files changed, 765 insertions, 0 deletions
diff --git a/Userland/Utilities/test-js.cpp b/Userland/Utilities/test-js.cpp
new file mode 100644
index 0000000000..9245d06492
--- /dev/null
+++ b/Userland/Utilities/test-js.cpp
@@ -0,0 +1,765 @@
+/*
+ * Copyright (c) 2020, Matthew Olsson <matthewcolsson@gmail.com>
+ * Copyright (c) 2020, Linus Groh <mail@linusgroh.de>
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions are met:
+ *
+ * 1. Redistributions of source code must retain the above copyright notice, this
+ * list of conditions and the following disclaimer.
+ *
+ * 2. Redistributions in binary form must reproduce the above copyright notice,
+ * this list of conditions and the following disclaimer in the documentation
+ * and/or other materials provided with the distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
+ * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
+ * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
+ * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
+ * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
+ * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+ * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+ */
+
+#include <AK/ByteBuffer.h>
+#include <AK/JsonObject.h>
+#include <AK/JsonValue.h>
+#include <AK/LexicalPath.h>
+#include <AK/LogStream.h>
+#include <AK/QuickSort.h>
+#include <LibCore/ArgsParser.h>
+#include <LibCore/DirIterator.h>
+#include <LibCore/File.h>
+#include <LibJS/Interpreter.h>
+#include <LibJS/Lexer.h>
+#include <LibJS/Parser.h>
+#include <LibJS/Runtime/Array.h>
+#include <LibJS/Runtime/GlobalObject.h>
+#include <LibJS/Runtime/JSONObject.h>
+#include <signal.h>
+#include <stdlib.h>
+#include <sys/time.h>
+
+#define TOP_LEVEL_TEST_NAME "__$$TOP_LEVEL$$__"
+
+RefPtr<JS::VM> vm;
+
+static bool collect_on_every_allocation = false;
+static String currently_running_test;
+
+enum class TestResult {
+ Pass,
+ Fail,
+ Skip,
+};
+
+struct JSTest {
+ String name;
+ TestResult result;
+ String details;
+};
+
+struct JSSuite {
+ String name;
+ // A failed test takes precedence over a skipped test, which both have
+ // precedence over a passed test
+ TestResult most_severe_test_result { TestResult::Pass };
+ Vector<JSTest> tests {};
+};
+
+struct ParserError {
+ JS::Parser::Error error;
+ String hint;
+};
+
+struct JSFileResult {
+ String name;
+ Optional<ParserError> error {};
+ double time_taken { 0 };
+ // A failed test takes precedence over a skipped test, which both have
+ // precedence over a passed test
+ TestResult most_severe_test_result { TestResult::Pass };
+ Vector<JSSuite> suites {};
+ Vector<String> logged_messages {};
+};
+
+struct JSTestRunnerCounts {
+ int tests_failed { 0 };
+ int tests_passed { 0 };
+ int tests_skipped { 0 };
+ int suites_failed { 0 };
+ int suites_passed { 0 };
+ int files_total { 0 };
+};
+
+class TestRunnerGlobalObject : public JS::GlobalObject {
+public:
+ TestRunnerGlobalObject();
+ virtual ~TestRunnerGlobalObject() override;
+
+ virtual void initialize() override;
+
+private:
+ virtual const char* class_name() const override { return "TestRunnerGlobalObject"; }
+
+ JS_DECLARE_NATIVE_FUNCTION(is_strict_mode);
+ JS_DECLARE_NATIVE_FUNCTION(can_parse_source);
+};
+
+class TestRunner {
+public:
+ static TestRunner* the()
+ {
+ return s_the;
+ }
+
+ TestRunner(String test_root, bool print_times)
+ : m_test_root(move(test_root))
+ , m_print_times(print_times)
+ {
+ ASSERT(!s_the);
+ s_the = this;
+ }
+
+ void run();
+
+ const JSTestRunnerCounts& counts() const { return m_counts; }
+
+protected:
+ static TestRunner* s_the;
+
+ virtual Vector<String> get_test_paths() const;
+ virtual JSFileResult run_file_test(const String& test_path);
+ void print_file_result(const JSFileResult& file_result) const;
+ void print_test_results() const;
+
+ String m_test_root;
+ bool m_print_times;
+
+ double m_total_elapsed_time_in_ms { 0 };
+ JSTestRunnerCounts m_counts;
+
+ RefPtr<JS::Program> m_test_program;
+};
+
+TestRunner* TestRunner::s_the = nullptr;
+
+TestRunnerGlobalObject::TestRunnerGlobalObject()
+{
+}
+
+TestRunnerGlobalObject::~TestRunnerGlobalObject()
+{
+}
+
+void TestRunnerGlobalObject::initialize()
+{
+ JS::GlobalObject::initialize();
+ static FlyString global_property_name { "global" };
+ static FlyString is_strict_mode_property_name { "isStrictMode" };
+ static FlyString can_parse_source_property_name { "canParseSource" };
+ define_property(global_property_name, this, JS::Attribute::Enumerable);
+ define_native_function(is_strict_mode_property_name, is_strict_mode);
+ define_native_function(can_parse_source_property_name, can_parse_source);
+}
+
+JS_DEFINE_NATIVE_FUNCTION(TestRunnerGlobalObject::is_strict_mode)
+{
+ return JS::Value(vm.in_strict_mode());
+}
+
+JS_DEFINE_NATIVE_FUNCTION(TestRunnerGlobalObject::can_parse_source)
+{
+ auto source = vm.argument(0).to_string(global_object);
+ if (vm.exception())
+ return {};
+ auto parser = JS::Parser(JS::Lexer(source));
+ parser.parse_program();
+ return JS::Value(!parser.has_errors());
+}
+
+static void cleanup_and_exit()
+{
+ // Clear the taskbar progress.
+#ifdef __serenity__
+ warn("\033]9;-1;\033\\");
+#endif
+ exit(1);
+}
+
+static void handle_sigabrt(int)
+{
+ dbgln("test-js: SIGABRT received, cleaning up.");
+ cleanup_and_exit();
+}
+
+static double get_time_in_ms()
+{
+ struct timeval tv1;
+ auto return_code = gettimeofday(&tv1, nullptr);
+ ASSERT(return_code >= 0);
+ return static_cast<double>(tv1.tv_sec) * 1000.0 + static_cast<double>(tv1.tv_usec) / 1000.0;
+}
+
+template<typename Callback>
+static void iterate_directory_recursively(const String& directory_path, Callback callback)
+{
+ Core::DirIterator directory_iterator(directory_path, Core::DirIterator::Flags::SkipDots);
+
+ while (directory_iterator.has_next()) {
+ auto file_path = String::formatted("{}/{}", directory_path, directory_iterator.next_path());
+ if (Core::File::is_directory(file_path)) {
+ iterate_directory_recursively(file_path, callback);
+ } else {
+ callback(move(file_path));
+ }
+ }
+}
+
+Vector<String> TestRunner::get_test_paths() const
+{
+ Vector<String> paths;
+ iterate_directory_recursively(m_test_root, [&](const String& file_path) {
+ if (!file_path.ends_with("test-common.js"))
+ paths.append(file_path);
+ });
+ quick_sort(paths);
+ return paths;
+}
+
+void TestRunner::run()
+{
+ size_t progress_counter = 0;
+ auto test_paths = get_test_paths();
+ for (auto& path : test_paths) {
+ ++progress_counter;
+ print_file_result(run_file_test(path));
+#ifdef __serenity__
+ warn("\033]9;{};{};\033\\", progress_counter, test_paths.size());
+#endif
+ }
+
+#ifdef __serenity__
+ warn("\033]9;-1;\033\\");
+#endif
+
+ print_test_results();
+}
+
+static Result<NonnullRefPtr<JS::Program>, ParserError> parse_file(const String& file_path)
+{
+ auto file = Core::File::construct(file_path);
+ auto result = file->open(Core::IODevice::ReadOnly);
+ if (!result) {
+ warnln("Failed to open the following file: \"{}\"", file_path);
+ cleanup_and_exit();
+ }
+
+ auto contents = file->read_all();
+ String test_file_string(reinterpret_cast<const char*>(contents.data()), contents.size());
+ file->close();
+
+ auto parser = JS::Parser(JS::Lexer(test_file_string));
+ auto program = parser.parse_program();
+
+ if (parser.has_errors()) {
+ auto error = parser.errors()[0];
+ return Result<NonnullRefPtr<JS::Program>, ParserError>(ParserError { error, error.source_location_hint(test_file_string) });
+ }
+
+ return Result<NonnullRefPtr<JS::Program>, ParserError>(program);
+}
+
+static Optional<JsonValue> get_test_results(JS::Interpreter& interpreter)
+{
+ auto result = vm->get_variable("__TestResults__", interpreter.global_object());
+ auto json_string = JS::JSONObject::stringify_impl(interpreter.global_object(), result, JS::js_undefined(), JS::js_undefined());
+
+ auto json = JsonValue::from_string(json_string);
+ if (!json.has_value())
+ return {};
+
+ return json.value();
+}
+
+JSFileResult TestRunner::run_file_test(const String& test_path)
+{
+ currently_running_test = test_path;
+
+ double start_time = get_time_in_ms();
+ auto interpreter = JS::Interpreter::create<TestRunnerGlobalObject>(*vm);
+
+ // FIXME: This is a hack while we're refactoring Interpreter/VM stuff.
+ JS::VM::InterpreterExecutionScope scope(*interpreter);
+
+ interpreter->heap().set_should_collect_on_every_allocation(collect_on_every_allocation);
+
+ if (!m_test_program) {
+ auto result = parse_file(String::formatted("{}/test-common.js", m_test_root));
+ if (result.is_error()) {
+ warnln("Unable to parse test-common.js");
+ warnln("{}", result.error().error.to_string());
+ warnln("{}", result.error().hint);
+ cleanup_and_exit();
+ }
+ m_test_program = result.value();
+ }
+
+ interpreter->run(interpreter->global_object(), *m_test_program);
+
+ auto file_program = parse_file(test_path);
+ if (file_program.is_error())
+ return { test_path, file_program.error() };
+ interpreter->run(interpreter->global_object(), *file_program.value());
+
+ auto test_json = get_test_results(*interpreter);
+ if (!test_json.has_value()) {
+ warnln("Received malformed JSON from test \"{}\"", test_path);
+ cleanup_and_exit();
+ }
+
+ JSFileResult file_result { test_path.substring(m_test_root.length() + 1, test_path.length() - m_test_root.length() - 1) };
+
+ // Collect logged messages
+ auto& arr = interpreter->vm().get_variable("__UserOutput__", interpreter->global_object()).as_array();
+ for (auto& entry : arr.indexed_properties()) {
+ auto message = entry.value_and_attributes(&interpreter->global_object()).value;
+ file_result.logged_messages.append(message.to_string_without_side_effects());
+ }
+
+ test_json.value().as_object().for_each_member([&](const String& suite_name, const JsonValue& suite_value) {
+ JSSuite suite { suite_name };
+
+ ASSERT(suite_value.is_object());
+
+ suite_value.as_object().for_each_member([&](const String& test_name, const JsonValue& test_value) {
+ JSTest test { test_name, TestResult::Fail, "" };
+
+ ASSERT(test_value.is_object());
+ ASSERT(test_value.as_object().has("result"));
+
+ auto result = test_value.as_object().get("result");
+ ASSERT(result.is_string());
+ auto result_string = result.as_string();
+ if (result_string == "pass") {
+ test.result = TestResult::Pass;
+ m_counts.tests_passed++;
+ } else if (result_string == "fail") {
+ test.result = TestResult::Fail;
+ m_counts.tests_failed++;
+ suite.most_severe_test_result = TestResult::Fail;
+ ASSERT(test_value.as_object().has("details"));
+ auto details = test_value.as_object().get("details");
+ ASSERT(result.is_string());
+ test.details = details.as_string();
+ } else {
+ test.result = TestResult::Skip;
+ if (suite.most_severe_test_result == TestResult::Pass)
+ suite.most_severe_test_result = TestResult::Skip;
+ m_counts.tests_skipped++;
+ }
+
+ suite.tests.append(test);
+ });
+
+ if (suite.most_severe_test_result == TestResult::Fail) {
+ m_counts.suites_failed++;
+ file_result.most_severe_test_result = TestResult::Fail;
+ } else {
+ if (suite.most_severe_test_result == TestResult::Skip && file_result.most_severe_test_result == TestResult::Pass)
+ file_result.most_severe_test_result = TestResult::Skip;
+ m_counts.suites_passed++;
+ }
+
+ file_result.suites.append(suite);
+ });
+
+ m_counts.files_total++;
+
+ file_result.time_taken = get_time_in_ms() - start_time;
+ m_total_elapsed_time_in_ms += file_result.time_taken;
+
+ return file_result;
+}
+
+enum Modifier {
+ BG_RED,
+ BG_GREEN,
+ FG_RED,
+ FG_GREEN,
+ FG_ORANGE,
+ FG_GRAY,
+ FG_BLACK,
+ FG_BOLD,
+ ITALIC,
+ CLEAR,
+};
+
+static void print_modifiers(Vector<Modifier> modifiers)
+{
+ for (auto& modifier : modifiers) {
+ auto code = [&] {
+ switch (modifier) {
+ case BG_RED:
+ return "\033[48;2;255;0;102m";
+ case BG_GREEN:
+ return "\033[48;2;102;255;0m";
+ case FG_RED:
+ return "\033[38;2;255;0;102m";
+ case FG_GREEN:
+ return "\033[38;2;102;255;0m";
+ case FG_ORANGE:
+ return "\033[38;2;255;102;0m";
+ case FG_GRAY:
+ return "\033[38;2;135;139;148m";
+ case FG_BLACK:
+ return "\033[30m";
+ case FG_BOLD:
+ return "\033[1m";
+ case ITALIC:
+ return "\033[3m";
+ case CLEAR:
+ return "\033[0m";
+ }
+ ASSERT_NOT_REACHED();
+ }();
+ out("{}", code);
+ }
+}
+
+void TestRunner::print_file_result(const JSFileResult& file_result) const
+{
+ if (file_result.most_severe_test_result == TestResult::Fail || file_result.error.has_value()) {
+ print_modifiers({ BG_RED, FG_BLACK, FG_BOLD });
+ out(" FAIL ");
+ print_modifiers({ CLEAR });
+ } else {
+ if (m_print_times || file_result.most_severe_test_result != TestResult::Pass) {
+ print_modifiers({ BG_GREEN, FG_BLACK, FG_BOLD });
+ out(" PASS ");
+ print_modifiers({ CLEAR });
+ } else {
+ return;
+ }
+ }
+
+ out(" {}", file_result.name);
+
+ if (m_print_times) {
+ print_modifiers({ CLEAR, ITALIC, FG_GRAY });
+ if (file_result.time_taken < 1000) {
+ outln(" ({}ms)", static_cast<int>(file_result.time_taken));
+ } else {
+ outln(" ({:3}s)", file_result.time_taken / 1000.0);
+ }
+ print_modifiers({ CLEAR });
+ } else {
+ outln();
+ }
+
+ if (!file_result.logged_messages.is_empty()) {
+ print_modifiers({ FG_GRAY, FG_BOLD });
+#ifdef __serenity__
+ outln(" ℹ Console output:");
+#else
+ // This emoji has a second invisible byte after it. The one above does not
+ outln(" ℹī¸ Console output:");
+#endif
+ print_modifiers({ CLEAR, FG_GRAY });
+ for (auto& message : file_result.logged_messages)
+ outln(" {}", message);
+ }
+
+ if (file_result.error.has_value()) {
+ auto test_error = file_result.error.value();
+
+ print_modifiers({ FG_RED });
+#ifdef __serenity__
+ outln(" ❌ The file failed to parse");
+#else
+ // No invisible byte here, but the spacing still needs to be altered on the host
+ outln(" ❌ The file failed to parse");
+#endif
+ outln();
+ print_modifiers({ FG_GRAY });
+ for (auto& message : test_error.hint.split('\n', true)) {
+ outln(" {}", message);
+ }
+ print_modifiers({ FG_RED });
+ outln(" {}", test_error.error.to_string());
+ outln();
+ return;
+ }
+
+ if (file_result.most_severe_test_result != TestResult::Pass) {
+ for (auto& suite : file_result.suites) {
+ if (suite.most_severe_test_result == TestResult::Pass)
+ continue;
+
+ bool failed = suite.most_severe_test_result == TestResult::Fail;
+
+ print_modifiers({ FG_GRAY, FG_BOLD });
+
+ if (failed) {
+#ifdef __serenity__
+ out(" ❌ Suite: ");
+#else
+ // No invisible byte here, but the spacing still needs to be altered on the host
+ out(" ❌ Suite: ");
+#endif
+ } else {
+#ifdef __serenity__
+ out(" ⚠ Suite: ");
+#else
+ // This emoji has a second invisible byte after it. The one above does not
+ out(" ⚠ī¸ Suite: ");
+#endif
+ }
+
+ print_modifiers({ CLEAR, FG_GRAY });
+
+ if (suite.name == TOP_LEVEL_TEST_NAME) {
+ outln("<top-level>");
+ } else {
+ outln("{}", suite.name);
+ }
+ print_modifiers({ CLEAR });
+
+ for (auto& test : suite.tests) {
+ if (test.result == TestResult::Pass)
+ continue;
+
+ print_modifiers({ FG_GRAY, FG_BOLD });
+ out(" Test: ");
+ if (test.result == TestResult::Fail) {
+ print_modifiers({ CLEAR, FG_RED });
+ outln("{} (failed):", test.name);
+ outln(" {}", test.details);
+ } else {
+ print_modifiers({ CLEAR, FG_ORANGE });
+ outln("{} (skipped)", test.name);
+ }
+ print_modifiers({ CLEAR });
+ }
+ }
+ }
+}
+
+void TestRunner::print_test_results() const
+{
+ out("\nTest Suites: ");
+ if (m_counts.suites_failed) {
+ print_modifiers({ FG_RED });
+ out("{} failed, ", m_counts.suites_failed);
+ print_modifiers({ CLEAR });
+ }
+ if (m_counts.suites_passed) {
+ print_modifiers({ FG_GREEN });
+ out("{} passed, ", m_counts.suites_passed);
+ print_modifiers({ CLEAR });
+ }
+ outln("{} total", m_counts.suites_failed + m_counts.suites_passed);
+
+ out("Tests: ");
+ if (m_counts.tests_failed) {
+ print_modifiers({ FG_RED });
+ out("{} failed, ", m_counts.tests_failed);
+ print_modifiers({ CLEAR });
+ }
+ if (m_counts.tests_skipped) {
+ print_modifiers({ FG_ORANGE });
+ out("{} skipped, ", m_counts.tests_skipped);
+ print_modifiers({ CLEAR });
+ }
+ if (m_counts.tests_passed) {
+ print_modifiers({ FG_GREEN });
+ out("{} passed, ", m_counts.tests_passed);
+ print_modifiers({ CLEAR });
+ }
+ outln("{} total", m_counts.tests_failed + m_counts.tests_skipped + m_counts.tests_passed);
+
+ outln("Files: {} total", m_counts.files_total);
+
+ out("Time: ");
+ if (m_total_elapsed_time_in_ms < 1000.0) {
+ outln("{}ms", static_cast<int>(m_total_elapsed_time_in_ms));
+ } else {
+ outln("{:>.3}s", m_total_elapsed_time_in_ms / 1000.0);
+ }
+ outln();
+}
+
+class Test262ParserTestRunner final : public TestRunner {
+public:
+ using TestRunner::TestRunner;
+
+private:
+ virtual Vector<String> get_test_paths() const override;
+ virtual JSFileResult run_file_test(const String& test_path) override;
+};
+
+Vector<String> Test262ParserTestRunner::get_test_paths() const
+{
+ Vector<String> paths;
+ iterate_directory_recursively(m_test_root, [&](const String& file_path) {
+ auto dirname = LexicalPath(file_path).dirname();
+ if (dirname.ends_with("early") || dirname.ends_with("fail") || dirname.ends_with("pass") || dirname.ends_with("pass-explicit"))
+ paths.append(file_path);
+ });
+ quick_sort(paths);
+ return paths;
+}
+
+JSFileResult Test262ParserTestRunner::run_file_test(const String& test_path)
+{
+ currently_running_test = test_path;
+
+ auto dirname = LexicalPath(test_path).dirname();
+ bool expecting_file_to_parse;
+ if (dirname.ends_with("early") || dirname.ends_with("fail")) {
+ expecting_file_to_parse = false;
+ } else if (dirname.ends_with("pass") || dirname.ends_with("pass-explicit")) {
+ expecting_file_to_parse = true;
+ } else {
+ ASSERT_NOT_REACHED();
+ }
+
+ auto start_time = get_time_in_ms();
+ String details = "";
+ TestResult test_result;
+ if (test_path.ends_with(".module.js")) {
+ test_result = TestResult::Skip;
+ m_counts.tests_skipped++;
+ m_counts.suites_passed++;
+ } else {
+ auto parse_result = parse_file(test_path);
+ if (expecting_file_to_parse) {
+ if (!parse_result.is_error()) {
+ test_result = TestResult::Pass;
+ } else {
+ test_result = TestResult::Fail;
+ details = parse_result.error().error.to_string();
+ }
+ } else {
+ if (parse_result.is_error()) {
+ test_result = TestResult::Pass;
+ } else {
+ test_result = TestResult::Fail;
+ details = "File was expected to produce a parser error but didn't";
+ }
+ }
+ }
+
+ // test262-parser-tests doesn't have "suites" and "tests" in the usual sense, it just has files
+ // and an expectation whether they should parse or not. We add one suite with one test nonetheless:
+ //
+ // - This makes interpreting skipped test easier as their file is shown as "PASS"
+ // - That way we can show additional information such as "file parsed but shouldn't have" or
+ // parser errors for files that should parse respectively
+
+ JSTest test { expecting_file_to_parse ? "file should parse" : "file should not parse", test_result, details };
+ JSSuite suite { "Parse file", test_result, { test } };
+ JSFileResult file_result {
+ test_path.substring(m_test_root.length() + 1, test_path.length() - m_test_root.length() - 1),
+ {},
+ get_time_in_ms() - start_time,
+ test_result,
+ { suite }
+ };
+
+ if (test_result == TestResult::Fail) {
+ m_counts.tests_failed++;
+ m_counts.suites_failed++;
+ } else {
+ m_counts.tests_passed++;
+ m_counts.suites_passed++;
+ }
+ m_counts.files_total++;
+ m_total_elapsed_time_in_ms += file_result.time_taken;
+
+ return file_result;
+}
+
+int main(int argc, char** argv)
+{
+ struct sigaction act;
+ memset(&act, 0, sizeof(act));
+ act.sa_flags = SA_NOCLDWAIT;
+ act.sa_handler = handle_sigabrt;
+ int rc = sigaction(SIGABRT, &act, nullptr);
+ if (rc < 0) {
+ perror("sigaction");
+ return 1;
+ }
+
+#ifdef SIGINFO
+ signal(SIGINFO, [](int) {
+ static char buffer[4096];
+ auto& counts = TestRunner::the()->counts();
+ int len = snprintf(buffer, sizeof(buffer), "Pass: %d, Fail: %d, Skip: %d\nCurrent test: %s\n", counts.tests_passed, counts.tests_failed, counts.tests_skipped, currently_running_test.characters());
+ write(STDOUT_FILENO, buffer, len);
+ });
+#endif
+
+ bool print_times = false;
+ bool test262_parser_tests = false;
+ const char* specified_test_root = nullptr;
+
+ Core::ArgsParser args_parser;
+ args_parser.add_option(print_times, "Show duration of each test", "show-time", 't');
+ args_parser.add_option(collect_on_every_allocation, "Collect garbage after every allocation", "collect-often", 'g');
+ args_parser.add_option(test262_parser_tests, "Run test262 parser tests", "test262-parser-tests", 0);
+ args_parser.add_positional_argument(specified_test_root, "Tests root directory", "path", Core::ArgsParser::Required::No);
+ args_parser.parse(argc, argv);
+
+ if (test262_parser_tests) {
+ if (collect_on_every_allocation) {
+ warnln("--collect-often and --test262-parser-tests options must not be used together");
+ return 1;
+ }
+ if (!specified_test_root) {
+ warnln("Test root is required with --test262-parser-tests");
+ return 1;
+ }
+ }
+
+ if (getenv("DISABLE_DBG_OUTPUT")) {
+ DebugLogStream::set_enabled(false);
+ }
+
+ String test_root;
+
+ if (specified_test_root) {
+ test_root = String { specified_test_root };
+ } else {
+#ifdef __serenity__
+ test_root = "/home/anon/js-tests";
+#else
+ char* serenity_root = getenv("SERENITY_ROOT");
+ if (!serenity_root) {
+ warnln("No test root given, test-js requires the SERENITY_ROOT environment variable to be set");
+ return 1;
+ }
+ test_root = String::formatted("{}/Libraries/LibJS/Tests", serenity_root);
+#endif
+ }
+ if (!Core::File::is_directory(test_root)) {
+ warnln("Test root is not a directory: {}", test_root);
+ return 1;
+ }
+
+ vm = JS::VM::create();
+
+ if (test262_parser_tests)
+ Test262ParserTestRunner(test_root, print_times).run();
+ else
+ TestRunner(test_root, print_times).run();
+
+ vm = nullptr;
+
+ return TestRunner::the()->counts().tests_failed > 0 ? 1 : 0;
+}