diff options
author | Timothy Flynn <trflynn89@pm.me> | 2021-04-18 17:35:40 -0400 |
---|---|---|
committer | Andreas Kling <kling@serenityos.org> | 2021-04-20 18:28:34 +0200 |
commit | 90517da9ca6d3cb3b44a0c251a926905cb779744 (patch) | |
tree | fea5fb94390aee6c3f6a56c56eac817a3ccbf133 | |
parent | 55b7b8e93080aa7e9c52db505da32dfa00b4b4bc (diff) | |
download | serenity-90517da9ca6d3cb3b44a0c251a926905cb779744.zip |
LibSQL: Introduce a SQL library with a tokenizer
LibSQL aims to be a SQLite clone for SerenityOS. Step 1 is creating a
tokenizer to lex SQL tokens. This lexer is heavily influenced by the
LibJS lexer.
-rw-r--r-- | Userland/Libraries/CMakeLists.txt | 1 | ||||
-rw-r--r-- | Userland/Libraries/LibSQL/CMakeLists.txt | 7 | ||||
-rw-r--r-- | Userland/Libraries/LibSQL/Lexer.cpp | 306 | ||||
-rw-r--r-- | Userland/Libraries/LibSQL/Lexer.h | 71 | ||||
-rw-r--r-- | Userland/Libraries/LibSQL/Token.cpp | 73 | ||||
-rw-r--r-- | Userland/Libraries/LibSQL/Token.h | 258 |
6 files changed, 716 insertions, 0 deletions
diff --git a/Userland/Libraries/CMakeLists.txt b/Userland/Libraries/CMakeLists.txt index c0fa49770b..1c495ddd6d 100644 --- a/Userland/Libraries/CMakeLists.txt +++ b/Userland/Libraries/CMakeLists.txt @@ -27,6 +27,7 @@ add_subdirectory(LibPCIDB) add_subdirectory(LibProtocol) add_subdirectory(LibPthread) add_subdirectory(LibRegex) +add_subdirectory(LibSQL) add_subdirectory(LibSymbolClient) add_subdirectory(LibSyntax) add_subdirectory(LibSystem) diff --git a/Userland/Libraries/LibSQL/CMakeLists.txt b/Userland/Libraries/LibSQL/CMakeLists.txt new file mode 100644 index 0000000000..05631d71a2 --- /dev/null +++ b/Userland/Libraries/LibSQL/CMakeLists.txt @@ -0,0 +1,7 @@ +set(SOURCES + Lexer.cpp + Token.cpp +) + +serenity_lib(LibSQL sql) +target_link_libraries(LibSQL LibCore) diff --git a/Userland/Libraries/LibSQL/Lexer.cpp b/Userland/Libraries/LibSQL/Lexer.cpp new file mode 100644 index 0000000000..7802befe54 --- /dev/null +++ b/Userland/Libraries/LibSQL/Lexer.cpp @@ -0,0 +1,306 @@ +/* + * Copyright (c) 2021, Tim Flynn <trflynn89@pm.me> + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are met: + * + * 1. Redistributions of source code must retain the above copyright notice, this + * list of conditions and the following disclaimer. + * + * 2. Redistributions in binary form must reproduce the above copyright notice, + * this list of conditions and the following disclaimer in the documentation + * and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE + * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL + * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR + * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER + * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, + * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +#include "Lexer.h" +#include <AK/Debug.h> +#include <ctype.h> + +namespace SQL { + +HashMap<String, TokenType> Lexer::s_keywords; +HashMap<char, TokenType> Lexer::s_one_char_tokens; +HashMap<String, TokenType> Lexer::s_two_char_tokens; + +Lexer::Lexer(StringView source) + : m_source(source) +{ + if (s_keywords.is_empty()) { +#define __ENUMERATE_SQL_TOKEN(value, type, category) \ + if (TokenCategory::category == TokenCategory::Keyword) \ + s_keywords.set(value, TokenType::type); + ENUMERATE_SQL_TOKENS +#undef __ENUMERATE_SQL_TOKEN + } + + if (s_one_char_tokens.is_empty()) { +#define __ENUMERATE_SQL_TOKEN(value, type, category) \ + if (TokenCategory::category != TokenCategory::Keyword && StringView(value).length() == 1) \ + s_one_char_tokens.set(value[0], TokenType::type); + ENUMERATE_SQL_TOKENS +#undef __ENUMERATE_SQL_TOKEN + } + + if (s_two_char_tokens.is_empty()) { +#define __ENUMERATE_SQL_TOKEN(value, type, category) \ + if (TokenCategory::category != TokenCategory::Keyword && StringView(value).length() == 2) \ + s_two_char_tokens.set(value, TokenType::type); + ENUMERATE_SQL_TOKENS +#undef __ENUMERATE_SQL_TOKEN + } + + consume(); +} + +Token Lexer::next() +{ + bool found_invalid_comment = consume_whitespace_and_comments(); + + size_t value_start = m_position; + size_t value_start_line_number = m_line_number; + size_t value_start_column_number = m_line_column; + auto token_type = TokenType::Invalid; + + if (is_eof()) { + token_type = found_invalid_comment ? TokenType::Invalid : TokenType::Eof; + } else if (is_identifier_start()) { + do { + consume(); + } while (is_identifier_middle()); + + if (auto it = s_keywords.find(m_source.substring_view(value_start - 1, m_position - value_start)); it != s_keywords.end()) { + token_type = it->value; + } else { + token_type = TokenType::Identifier; + } + } else if (is_numeric_literal_start()) { + token_type = TokenType::NumericLiteral; + if (!consume_numeric_literal()) + token_type = TokenType::Invalid; + } else { + bool found_two_char_token = false; + if (m_position < m_source.length()) { + if (auto it = s_two_char_tokens.find(m_source.substring_view(m_position - 1, 2)); it != s_two_char_tokens.end()) { + found_two_char_token = true; + token_type = it->value; + consume(); + consume(); + } + } + + bool found_one_char_token = false; + if (!found_two_char_token) { + if (auto it = s_one_char_tokens.find(m_current_char); it != s_one_char_tokens.end()) { + found_one_char_token = true; + token_type = it->value; + consume(); + } + } + + if (!found_two_char_token && !found_one_char_token) { + token_type = TokenType::Invalid; + consume(); + } + } + + Token token(token_type, m_source.substring_view(value_start - 1, m_position - value_start), value_start_line_number, value_start_column_number); + + if constexpr (SQL_DEBUG) { + dbgln("------------------------------"); + dbgln("Token: {}", token.name()); + dbgln("Value: {}", token.value()); + dbgln("Line: {}, Column: {}", token.line_number(), token.line_column()); + dbgln("------------------------------"); + } + + return token; +} + +void Lexer::consume() +{ + auto did_reach_eof = [this] { + if (m_position != m_source.length()) + return false; + m_current_char = EOF; + ++m_line_column; + ++m_position; + return true; + }; + + if (m_position > m_source.length()) + return; + + if (did_reach_eof()) + return; + + if (is_line_break()) { + ++m_line_number; + m_line_column = 1; + } else { + ++m_line_column; + } + + m_current_char = m_source[m_position++]; +} + +bool Lexer::consume_whitespace_and_comments() +{ + bool found_invalid_comment = false; + + while (true) { + if (isspace(m_current_char)) { + do { + consume(); + } while (isspace(m_current_char)); + } else if (is_line_comment_start()) { + consume(); + do { + consume(); + } while (!is_eof() && !is_line_break()); + } else if (is_block_comment_start()) { + consume(); + do { + consume(); + } while (!is_eof() && !is_block_comment_end()); + if (is_eof()) + found_invalid_comment = true; + consume(); // consume * + if (is_eof()) + found_invalid_comment = true; + consume(); // consume / + } else { + break; + } + } + + return found_invalid_comment; +} + +bool Lexer::consume_numeric_literal() +{ + // https://www.sqlite.org/draft/syntax/numeric-literal.html + bool is_valid_numeric_literal = true; + + if (m_current_char == '0') { + consume(); + if (m_current_char == '.') { + consume(); + while (isdigit(m_current_char)) + consume(); + if (m_current_char == 'e' || m_current_char == 'E') + is_valid_numeric_literal = consume_exponent(); + } else if (m_current_char == 'e' || m_current_char == 'E') { + is_valid_numeric_literal = consume_exponent(); + } else if (m_current_char == 'x' || m_current_char == 'X') { + is_valid_numeric_literal = consume_hexadecimal_number(); + } else if (isdigit(m_current_char)) { + do { + consume(); + } while (isdigit(m_current_char)); + } + } else { + do { + consume(); + } while (isdigit(m_current_char)); + + if (m_current_char == '.') { + consume(); + while (isdigit(m_current_char)) + consume(); + } + if (m_current_char == 'e' || m_current_char == 'E') + is_valid_numeric_literal = consume_exponent(); + } + + return is_valid_numeric_literal; +} + +bool Lexer::consume_exponent() +{ + consume(); + if (m_current_char == '-' || m_current_char == '+') + consume(); + + if (!isdigit(m_current_char)) + return false; + + while (isdigit(m_current_char)) { + consume(); + } + return true; +} + +bool Lexer::consume_hexadecimal_number() +{ + consume(); + if (!isxdigit(m_current_char)) + return false; + + while (isxdigit(m_current_char)) + consume(); + + return true; +} + +bool Lexer::match(char a, char b) const +{ + if (m_position >= m_source.length()) + return false; + + return m_current_char == a + && m_source[m_position] == b; +} + +bool Lexer::is_identifier_start() const +{ + return isalpha(m_current_char) || m_current_char == '_'; +} + +bool Lexer::is_identifier_middle() const +{ + return is_identifier_start() || isdigit(m_current_char); +} + +bool Lexer::is_numeric_literal_start() const +{ + return isdigit(m_current_char) || (m_current_char == '.' && m_position < m_source.length() && isdigit(m_source[m_position])); +} + +bool Lexer::is_line_comment_start() const +{ + return match('-', '-'); +} + +bool Lexer::is_block_comment_start() const +{ + return match('/', '*'); +} + +bool Lexer::is_block_comment_end() const +{ + return match('*', '/'); +} + +bool Lexer::is_line_break() const +{ + return m_current_char == '\n'; +} + +bool Lexer::is_eof() const +{ + return m_current_char == EOF; +} + +} diff --git a/Userland/Libraries/LibSQL/Lexer.h b/Userland/Libraries/LibSQL/Lexer.h new file mode 100644 index 0000000000..502bc0668c --- /dev/null +++ b/Userland/Libraries/LibSQL/Lexer.h @@ -0,0 +1,71 @@ +/* + * Copyright (c) 2021, Tim Flynn <trflynn89@pm.me> + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are met: + * + * 1. Redistributions of source code must retain the above copyright notice, this + * list of conditions and the following disclaimer. + * + * 2. Redistributions in binary form must reproduce the above copyright notice, + * this list of conditions and the following disclaimer in the documentation + * and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE + * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL + * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR + * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER + * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, + * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +#pragma once + +#include "Token.h" +#include <AK/HashMap.h> +#include <AK/String.h> +#include <AK/StringView.h> + +namespace SQL { + +class Lexer { +public: + explicit Lexer(StringView source); + + Token next(); + +private: + void consume(); + + bool consume_whitespace_and_comments(); + bool consume_numeric_literal(); + bool consume_exponent(); + bool consume_hexadecimal_number(); + + bool match(char a, char b) const; + bool is_identifier_start() const; + bool is_identifier_middle() const; + bool is_numeric_literal_start() const; + bool is_line_comment_start() const; + bool is_block_comment_start() const; + bool is_block_comment_end() const; + bool is_line_break() const; + bool is_eof() const; + + static HashMap<String, TokenType> s_keywords; + static HashMap<char, TokenType> s_one_char_tokens; + static HashMap<String, TokenType> s_two_char_tokens; + + StringView m_source; + size_t m_line_number { 1 }; + size_t m_line_column { 0 }; + char m_current_char { 0 }; + size_t m_position { 0 }; +}; + +} diff --git a/Userland/Libraries/LibSQL/Token.cpp b/Userland/Libraries/LibSQL/Token.cpp new file mode 100644 index 0000000000..672855d0c2 --- /dev/null +++ b/Userland/Libraries/LibSQL/Token.cpp @@ -0,0 +1,73 @@ +/* + * Copyright (c) 2021, Tim Flynn <trflynn89@pm.me> + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are met: + * + * 1. Redistributions of source code must retain the above copyright notice, this + * list of conditions and the following disclaimer. + * + * 2. Redistributions in binary form must reproduce the above copyright notice, + * this list of conditions and the following disclaimer in the documentation + * and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE + * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL + * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR + * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER + * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, + * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +#include "Token.h" +#include <AK/Assertions.h> +#include <AK/String.h> +#include <stdlib.h> + +namespace SQL { + +StringView Token::name(TokenType type) +{ + switch (type) { +#define __ENUMERATE_SQL_TOKEN(value, type, category) \ + case TokenType::type: \ + return #type; + ENUMERATE_SQL_TOKENS +#undef __ENUMERATE_SQL_TOKEN + default: + VERIFY_NOT_REACHED(); + } +} + +TokenCategory Token::category(TokenType type) +{ + switch (type) { +#define __ENUMERATE_SQL_TOKEN(value, type, category) \ + case TokenType::type: \ + return TokenCategory::category; + ENUMERATE_SQL_TOKENS +#undef __ENUMERATE_SQL_TOKEN + default: + VERIFY_NOT_REACHED(); + } +} + +double Token::double_value() const +{ + VERIFY(type() == TokenType::NumericLiteral); + String value(m_value); + + if (value[0] == '0' && value.length() >= 2) { + if (value[1] == 'x' || value[1] == 'X') + return static_cast<double>(strtoul(value.characters() + 2, nullptr, 16)); + } + + return strtod(value.characters(), nullptr); +} + +} diff --git a/Userland/Libraries/LibSQL/Token.h b/Userland/Libraries/LibSQL/Token.h new file mode 100644 index 0000000000..a882b600b5 --- /dev/null +++ b/Userland/Libraries/LibSQL/Token.h @@ -0,0 +1,258 @@ +/* + * Copyright (c) 2021, Tim Flynn <trflynn89@pm.me> + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are met: + * + * 1. Redistributions of source code must retain the above copyright notice, this + * list of conditions and the following disclaimer. + * + * 2. Redistributions in binary form must reproduce the above copyright notice, + * this list of conditions and the following disclaimer in the documentation + * and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE + * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL + * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR + * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER + * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, + * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +#pragma once + +#include <AK/HashMap.h> +#include <AK/StringView.h> + +namespace SQL { + +// https://sqlite.org/lang_keywords.html +#define ENUMERATE_SQL_TOKENS \ + __ENUMERATE_SQL_TOKEN("ABORT", Abort, Keyword) \ + __ENUMERATE_SQL_TOKEN("ACTION", Action, Keyword) \ + __ENUMERATE_SQL_TOKEN("ADD", Add, Keyword) \ + __ENUMERATE_SQL_TOKEN("AFTER", After, Keyword) \ + __ENUMERATE_SQL_TOKEN("ALL", All, Keyword) \ + __ENUMERATE_SQL_TOKEN("ALTER", Alter, Keyword) \ + __ENUMERATE_SQL_TOKEN("ALWAYS", Always, Keyword) \ + __ENUMERATE_SQL_TOKEN("ANALYZE", Analyze, Keyword) \ + __ENUMERATE_SQL_TOKEN("AND", And, Keyword) \ + __ENUMERATE_SQL_TOKEN("AS", As, Keyword) \ + __ENUMERATE_SQL_TOKEN("ASC", Asc, Keyword) \ + __ENUMERATE_SQL_TOKEN("ATTACH", Attach, Keyword) \ + __ENUMERATE_SQL_TOKEN("AUTOINCREMENT", Autoincrement, Keyword) \ + __ENUMERATE_SQL_TOKEN("BEFORE", Before, Keyword) \ + __ENUMERATE_SQL_TOKEN("BEGIN", Begin, Keyword) \ + __ENUMERATE_SQL_TOKEN("BETWEEN", Between, Keyword) \ + __ENUMERATE_SQL_TOKEN("BY", By, Keyword) \ + __ENUMERATE_SQL_TOKEN("CASCADE", Cascade, Keyword) \ + __ENUMERATE_SQL_TOKEN("CASE", Case, Keyword) \ + __ENUMERATE_SQL_TOKEN("CAST", Cast, Keyword) \ + __ENUMERATE_SQL_TOKEN("CHECK", Check, Keyword) \ + __ENUMERATE_SQL_TOKEN("COLLATE", Collate, Keyword) \ + __ENUMERATE_SQL_TOKEN("COLUMN", Column, Keyword) \ + __ENUMERATE_SQL_TOKEN("COMMIT", Commit, Keyword) \ + __ENUMERATE_SQL_TOKEN("CONFLICT", Conflict, Keyword) \ + __ENUMERATE_SQL_TOKEN("CONSTRAINT", Constraint, Keyword) \ + __ENUMERATE_SQL_TOKEN("CREATE", Create, Keyword) \ + __ENUMERATE_SQL_TOKEN("CROSS", Cross, Keyword) \ + __ENUMERATE_SQL_TOKEN("CURRENT", Current, Keyword) \ + __ENUMERATE_SQL_TOKEN("CURRENT_DATE", CurrentDate, Keyword) \ + __ENUMERATE_SQL_TOKEN("CURRENT_TIME", CurrentTime, Keyword) \ + __ENUMERATE_SQL_TOKEN("CURRENT_TIMESTAMP", CurrentTimestamp, Keyword) \ + __ENUMERATE_SQL_TOKEN("DATABASE", Database, Keyword) \ + __ENUMERATE_SQL_TOKEN("DEFAULT", Default, Keyword) \ + __ENUMERATE_SQL_TOKEN("DEFERRABLE", Deferrable, Keyword) \ + __ENUMERATE_SQL_TOKEN("DEFERRED", Deferred, Keyword) \ + __ENUMERATE_SQL_TOKEN("DELETE", Delete, Keyword) \ + __ENUMERATE_SQL_TOKEN("DESC", Desc, Keyword) \ + __ENUMERATE_SQL_TOKEN("DETACH", Detach, Keyword) \ + __ENUMERATE_SQL_TOKEN("DISTINCT", Distinct, Keyword) \ + __ENUMERATE_SQL_TOKEN("DO", Do, Keyword) \ + __ENUMERATE_SQL_TOKEN("DROP", Drop, Keyword) \ + __ENUMERATE_SQL_TOKEN("EACH", Each, Keyword) \ + __ENUMERATE_SQL_TOKEN("ELSE", Else, Keyword) \ + __ENUMERATE_SQL_TOKEN("END", End, Keyword) \ + __ENUMERATE_SQL_TOKEN("ESCAPE", Escape, Keyword) \ + __ENUMERATE_SQL_TOKEN("EXCEPT", Except, Keyword) \ + __ENUMERATE_SQL_TOKEN("EXCLUDE", Exclude, Keyword) \ + __ENUMERATE_SQL_TOKEN("EXCLUSIVE", Exclusive, Keyword) \ + __ENUMERATE_SQL_TOKEN("EXISTS", Exists, Keyword) \ + __ENUMERATE_SQL_TOKEN("EXPLAIN", Explain, Keyword) \ + __ENUMERATE_SQL_TOKEN("FAIL", Fail, Keyword) \ + __ENUMERATE_SQL_TOKEN("FILTER", Filter, Keyword) \ + __ENUMERATE_SQL_TOKEN("FIRST", First, Keyword) \ + __ENUMERATE_SQL_TOKEN("FOLLOWING", Following, Keyword) \ + __ENUMERATE_SQL_TOKEN("FOR", For, Keyword) \ + __ENUMERATE_SQL_TOKEN("FOREIGN", Foreign, Keyword) \ + __ENUMERATE_SQL_TOKEN("FROM", From, Keyword) \ + __ENUMERATE_SQL_TOKEN("FULL", Full, Keyword) \ + __ENUMERATE_SQL_TOKEN("GENERATED", Generated, Keyword) \ + __ENUMERATE_SQL_TOKEN("GLOB", Glob, Keyword) \ + __ENUMERATE_SQL_TOKEN("GROUP", Group, Keyword) \ + __ENUMERATE_SQL_TOKEN("GROUPS", Groups, Keyword) \ + __ENUMERATE_SQL_TOKEN("HAVING", Having, Keyword) \ + __ENUMERATE_SQL_TOKEN("IF", If, Keyword) \ + __ENUMERATE_SQL_TOKEN("IGNORE", Ignore, Keyword) \ + __ENUMERATE_SQL_TOKEN("IMMEDIATE", Immediate, Keyword) \ + __ENUMERATE_SQL_TOKEN("IN", In, Keyword) \ + __ENUMERATE_SQL_TOKEN("INDEX", Index, Keyword) \ + __ENUMERATE_SQL_TOKEN("INDEXED", Indexed, Keyword) \ + __ENUMERATE_SQL_TOKEN("INITIALLY", Initially, Keyword) \ + __ENUMERATE_SQL_TOKEN("INNER", Inner, Keyword) \ + __ENUMERATE_SQL_TOKEN("INSERT", Insert, Keyword) \ + __ENUMERATE_SQL_TOKEN("INSTEAD", Instead, Keyword) \ + __ENUMERATE_SQL_TOKEN("INTERSECT", Intersect, Keyword) \ + __ENUMERATE_SQL_TOKEN("INTO", Into, Keyword) \ + __ENUMERATE_SQL_TOKEN("IS", Is, Keyword) \ + __ENUMERATE_SQL_TOKEN("ISNULL", Isnull, Keyword) \ + __ENUMERATE_SQL_TOKEN("JOIN", Join, Keyword) \ + __ENUMERATE_SQL_TOKEN("KEY", Key, Keyword) \ + __ENUMERATE_SQL_TOKEN("LAST", Last, Keyword) \ + __ENUMERATE_SQL_TOKEN("LEFT", Left, Keyword) \ + __ENUMERATE_SQL_TOKEN("LIKE", Like, Keyword) \ + __ENUMERATE_SQL_TOKEN("LIMIT", Limit, Keyword) \ + __ENUMERATE_SQL_TOKEN("MATCH", Match, Keyword) \ + __ENUMERATE_SQL_TOKEN("MATERIALIZED", Materialized, Keyword) \ + __ENUMERATE_SQL_TOKEN("NATURAL", Natural, Keyword) \ + __ENUMERATE_SQL_TOKEN("NO", No, Keyword) \ + __ENUMERATE_SQL_TOKEN("NOT", Not, Keyword) \ + __ENUMERATE_SQL_TOKEN("NOTHING", Nothing, Keyword) \ + __ENUMERATE_SQL_TOKEN("NOTNULL", Notnull, Keyword) \ + __ENUMERATE_SQL_TOKEN("NULL", Null, Keyword) \ + __ENUMERATE_SQL_TOKEN("NULLS", Nulls, Keyword) \ + __ENUMERATE_SQL_TOKEN("OF", Of, Keyword) \ + __ENUMERATE_SQL_TOKEN("OFFSET", Offset, Keyword) \ + __ENUMERATE_SQL_TOKEN("ON", On, Keyword) \ + __ENUMERATE_SQL_TOKEN("OR", Or, Keyword) \ + __ENUMERATE_SQL_TOKEN("ORDER", Order, Keyword) \ + __ENUMERATE_SQL_TOKEN("OTHERS", Others, Keyword) \ + __ENUMERATE_SQL_TOKEN("OUTER", Outer, Keyword) \ + __ENUMERATE_SQL_TOKEN("OVER", Over, Keyword) \ + __ENUMERATE_SQL_TOKEN("PARTITION", Partition, Keyword) \ + __ENUMERATE_SQL_TOKEN("PLAN", Plan, Keyword) \ + __ENUMERATE_SQL_TOKEN("PRAGMA", Pragma, Keyword) \ + __ENUMERATE_SQL_TOKEN("PRECEDING", Preceding, Keyword) \ + __ENUMERATE_SQL_TOKEN("PRIMARY", Primary, Keyword) \ + __ENUMERATE_SQL_TOKEN("QUERY", Query, Keyword) \ + __ENUMERATE_SQL_TOKEN("RAISE", Raise, Keyword) \ + __ENUMERATE_SQL_TOKEN("RANGE", Range, Keyword) \ + __ENUMERATE_SQL_TOKEN("RECURSIVE", Recursive, Keyword) \ + __ENUMERATE_SQL_TOKEN("REFERENCES", References, Keyword) \ + __ENUMERATE_SQL_TOKEN("REGEXP", Regexp, Keyword) \ + __ENUMERATE_SQL_TOKEN("REINDEX", Reindex, Keyword) \ + __ENUMERATE_SQL_TOKEN("RELEASE", Release, Keyword) \ + __ENUMERATE_SQL_TOKEN("RENAME", Rename, Keyword) \ + __ENUMERATE_SQL_TOKEN("REPLACE", Replace, Keyword) \ + __ENUMERATE_SQL_TOKEN("RESTRICT", Restrict, Keyword) \ + __ENUMERATE_SQL_TOKEN("RETURNING", Returning, Keyword) \ + __ENUMERATE_SQL_TOKEN("RIGHT", Right, Keyword) \ + __ENUMERATE_SQL_TOKEN("ROLLBACK", Rollback, Keyword) \ + __ENUMERATE_SQL_TOKEN("ROW", Row, Keyword) \ + __ENUMERATE_SQL_TOKEN("ROWS", Rows, Keyword) \ + __ENUMERATE_SQL_TOKEN("SAVEPOINT", Savepoint, Keyword) \ + __ENUMERATE_SQL_TOKEN("SELECT", Select, Keyword) \ + __ENUMERATE_SQL_TOKEN("SET", Set, Keyword) \ + __ENUMERATE_SQL_TOKEN("TABLE", Table, Keyword) \ + __ENUMERATE_SQL_TOKEN("TEMP", Temp, Keyword) \ + __ENUMERATE_SQL_TOKEN("TEMPORARY", Temporary, Keyword) \ + __ENUMERATE_SQL_TOKEN("THEN", Then, Keyword) \ + __ENUMERATE_SQL_TOKEN("TIES", Ties, Keyword) \ + __ENUMERATE_SQL_TOKEN("TO", To, Keyword) \ + __ENUMERATE_SQL_TOKEN("TRANSACTION", Transaction, Keyword) \ + __ENUMERATE_SQL_TOKEN("TRIGGER", Trigger, Keyword) \ + __ENUMERATE_SQL_TOKEN("UNBOUNDED", Unbounded, Keyword) \ + __ENUMERATE_SQL_TOKEN("UNION", Union, Keyword) \ + __ENUMERATE_SQL_TOKEN("UNIQUE", Unique, Keyword) \ + __ENUMERATE_SQL_TOKEN("UPDATE", Update, Keyword) \ + __ENUMERATE_SQL_TOKEN("USING", Using, Keyword) \ + __ENUMERATE_SQL_TOKEN("VACUUM", Vacuum, Keyword) \ + __ENUMERATE_SQL_TOKEN("VALUES", Values, Keyword) \ + __ENUMERATE_SQL_TOKEN("VIEW", View, Keyword) \ + __ENUMERATE_SQL_TOKEN("VIRTUAL", Virtual, Keyword) \ + __ENUMERATE_SQL_TOKEN("WHEN", When, Keyword) \ + __ENUMERATE_SQL_TOKEN("WHERE", Where, Keyword) \ + __ENUMERATE_SQL_TOKEN("WINDOW", Window, Keyword) \ + __ENUMERATE_SQL_TOKEN("WITH", With, Keyword) \ + __ENUMERATE_SQL_TOKEN("WITHOUT", Without, Keyword) \ + __ENUMERATE_SQL_TOKEN("_identifier_", Identifier, Identifier) \ + __ENUMERATE_SQL_TOKEN("_numeric_", NumericLiteral, Number) \ + __ENUMERATE_SQL_TOKEN("_eof_", Eof, Invalid) \ + __ENUMERATE_SQL_TOKEN("_invalid_", Invalid, Invalid) \ + __ENUMERATE_SQL_TOKEN("&", Ampersand, Operator) \ + __ENUMERATE_SQL_TOKEN("*", Asterisk, Operator) \ + __ENUMERATE_SQL_TOKEN(",", Comma, Punctuation) \ + __ENUMERATE_SQL_TOKEN("/", Divide, Operator) \ + __ENUMERATE_SQL_TOKEN("||", DoublePipe, Operator) \ + __ENUMERATE_SQL_TOKEN("=", Equals, Operator) \ + __ENUMERATE_SQL_TOKEN("==", EqualsEquals, Operator) \ + __ENUMERATE_SQL_TOKEN(">", GreaterThan, Operator) \ + __ENUMERATE_SQL_TOKEN(">=", GreaterThanEquals, Operator) \ + __ENUMERATE_SQL_TOKEN("<", LessThan, Operator) \ + __ENUMERATE_SQL_TOKEN("<=", LessThanEquals, Operator) \ + __ENUMERATE_SQL_TOKEN("-", Minus, Operator) \ + __ENUMERATE_SQL_TOKEN("%", Modulus, Operator) \ + __ENUMERATE_SQL_TOKEN("!=", NotEquals1, Operator) \ + __ENUMERATE_SQL_TOKEN("<>", NotEquals2, Operator) \ + __ENUMERATE_SQL_TOKEN(")", ParenClose, Punctuation) \ + __ENUMERATE_SQL_TOKEN("(", ParenOpen, Punctuation) \ + __ENUMERATE_SQL_TOKEN(".", Period, Operator) \ + __ENUMERATE_SQL_TOKEN("|", Pipe, Operator) \ + __ENUMERATE_SQL_TOKEN("+", Plus, Operator) \ + __ENUMERATE_SQL_TOKEN(";", SemiColon, Punctuation) \ + __ENUMERATE_SQL_TOKEN("<<", ShiftLeft, Operator) \ + __ENUMERATE_SQL_TOKEN(">>", ShiftRight, Operator) \ + __ENUMERATE_SQL_TOKEN("~", Tilde, Operator) + +enum class TokenType { +#define __ENUMERATE_SQL_TOKEN(value, type, category) type, + ENUMERATE_SQL_TOKENS +#undef __ENUMERATE_SQL_TOKEN + _COUNT_OF_TOKENS, +}; + +enum class TokenCategory { + Invalid, + Keyword, + Identifier, + Number, + Operator, + Punctuation, +}; + +class Token { +public: + Token(TokenType type, StringView value, size_t line_number, size_t line_column) + : m_type(type) + , m_value(value) + , m_line_number(line_number) + , m_line_column(line_column) + { + } + + static StringView name(TokenType); + static TokenCategory category(TokenType); + + StringView name() const { return name(m_type); } + TokenType type() const { return m_type; } + TokenCategory category() const { return category(m_type); } + + StringView value() const { return m_value; } + double double_value() const; + + size_t line_number() const { return m_line_number; } + size_t line_column() const { return m_line_column; } + +private: + TokenType m_type; + StringView m_value; + size_t m_line_number; + size_t m_line_column; +}; + +} |