summaryrefslogtreecommitdiff
path: root/Userland
diff options
context:
space:
mode:
authorItamar <itamar8910@gmail.com>2021-03-12 13:11:41 +0200
committerAndreas Kling <kling@serenityos.org>2021-03-13 10:17:02 +0100
commit26d9485562f90dd6ea750ba1206d0793395fb4d2 (patch)
treec6c3d9b34eb3aef04ad81085f8e46b01004bb07a /Userland
parent5cd1c69b968293ac95f51c68f2a78b16c671cf58 (diff)
downloadserenity-26d9485562f90dd6ea750ba1206d0793395fb4d2.zip
LibCpp: Store the text of a token as a StringView member
Diffstat (limited to 'Userland')
-rw-r--r--Userland/Libraries/LibCpp/Lexer.cpp30
-rw-r--r--Userland/Libraries/LibCpp/Parser.cpp8
-rw-r--r--Userland/Libraries/LibCpp/Token.h10
3 files changed, 25 insertions, 23 deletions
diff --git a/Userland/Libraries/LibCpp/Lexer.cpp b/Userland/Libraries/LibCpp/Lexer.cpp
index 1ed8b782d7..cac7c09298 100644
--- a/Userland/Libraries/LibCpp/Lexer.cpp
+++ b/Userland/Libraries/LibCpp/Lexer.cpp
@@ -231,8 +231,8 @@ Vector<Token> Lexer::lex()
size_t token_start_index = 0;
Position token_start_position;
- auto emit_token = [&](auto type) {
- tokens.empend(type, m_position, m_position);
+ auto emit_single_char_token = [&](auto type) {
+ tokens.empend(type, m_position, m_position, m_input.substring_view(m_index, 1));
consume();
};
@@ -241,7 +241,7 @@ Vector<Token> Lexer::lex()
token_start_position = m_position;
};
auto commit_token = [&](auto type) {
- tokens.empend(type, token_start_position, m_previous_position);
+ tokens.empend(type, token_start_position, m_previous_position, m_input.substring_view(token_start_index, m_index - token_start_index));
};
auto emit_token_equals = [&](auto type, auto equals_type) {
@@ -252,7 +252,7 @@ Vector<Token> Lexer::lex()
commit_token(equals_type);
return;
}
- emit_token(type);
+ emit_single_char_token(type);
};
auto match_escape_sequence = [&]() -> size_t {
@@ -335,27 +335,27 @@ Vector<Token> Lexer::lex()
continue;
}
if (ch == '(') {
- emit_token(Token::Type::LeftParen);
+ emit_single_char_token(Token::Type::LeftParen);
continue;
}
if (ch == ')') {
- emit_token(Token::Type::RightParen);
+ emit_single_char_token(Token::Type::RightParen);
continue;
}
if (ch == '{') {
- emit_token(Token::Type::LeftCurly);
+ emit_single_char_token(Token::Type::LeftCurly);
continue;
}
if (ch == '}') {
- emit_token(Token::Type::RightCurly);
+ emit_single_char_token(Token::Type::RightCurly);
continue;
}
if (ch == '[') {
- emit_token(Token::Type::LeftBracket);
+ emit_single_char_token(Token::Type::LeftBracket);
continue;
}
if (ch == ']') {
- emit_token(Token::Type::RightBracket);
+ emit_single_char_token(Token::Type::RightBracket);
continue;
}
if (ch == '<') {
@@ -406,7 +406,7 @@ Vector<Token> Lexer::lex()
continue;
}
if (ch == ',') {
- emit_token(Token::Type::Comma);
+ emit_single_char_token(Token::Type::Comma);
continue;
}
if (ch == '+') {
@@ -504,11 +504,11 @@ Vector<Token> Lexer::lex()
continue;
}
if (ch == '~') {
- emit_token(Token::Type::Tilde);
+ emit_single_char_token(Token::Type::Tilde);
continue;
}
if (ch == '?') {
- emit_token(Token::Type::QuestionMark);
+ emit_single_char_token(Token::Type::QuestionMark);
continue;
}
if (ch == ':') {
@@ -528,7 +528,7 @@ Vector<Token> Lexer::lex()
continue;
}
if (ch == ';') {
- emit_token(Token::Type::Semicolon);
+ emit_single_char_token(Token::Type::Semicolon);
continue;
}
if (ch == '.') {
@@ -778,7 +778,7 @@ Vector<Token> Lexer::lex()
continue;
}
dbgln("Unimplemented token character: {}", ch);
- emit_token(Token::Type::Unknown);
+ emit_single_char_token(Token::Type::Unknown);
}
return tokens;
}
diff --git a/Userland/Libraries/LibCpp/Parser.cpp b/Userland/Libraries/LibCpp/Parser.cpp
index 0e84c7f139..d4c5a3c540 100644
--- a/Userland/Libraries/LibCpp/Parser.cpp
+++ b/Userland/Libraries/LibCpp/Parser.cpp
@@ -657,7 +657,7 @@ Token Parser::consume()
{
if (eof()) {
error("C++ Parser: out of tokens");
- return { Token::Type::EOF_TOKEN, position(), position() };
+ return { Token::Type::EOF_TOKEN, position(), position(), {} };
}
return m_tokens[m_state.token_index++];
}
@@ -665,7 +665,7 @@ Token Parser::consume()
Token Parser::peek(size_t offset) const
{
if (m_state.token_index + offset >= m_tokens.size())
- return { Token::Type::EOF_TOKEN, position(), position() };
+ return { Token::Type::EOF_TOKEN, position(), position(), {} };
return m_tokens[m_state.token_index + offset];
}
@@ -699,9 +699,7 @@ bool Parser::done()
StringView Parser::text_of_token(const Cpp::Token& token) const
{
- VERIFY(token.start().line == token.end().line);
- VERIFY(token.start().column <= token.end().column);
- return m_lines[token.start().line].substring_view(token.start().column, token.end().column - token.start().column + 1);
+ return token.text();
}
StringView Parser::text_of_node(const ASTNode& node) const
diff --git a/Userland/Libraries/LibCpp/Token.h b/Userland/Libraries/LibCpp/Token.h
index e32c309b5a..73c23e6c9b 100644
--- a/Userland/Libraries/LibCpp/Token.h
+++ b/Userland/Libraries/LibCpp/Token.h
@@ -26,6 +26,7 @@
#pragma once
+#include <AK/StringView.h>
#include <AK/Types.h>
namespace Cpp {
@@ -114,10 +115,11 @@ struct Token {
#undef __TOKEN
};
- Token(Type type, const Position& start, const Position& end)
+ Token(Type type, const Position& start, const Position& end, const StringView& text)
: m_type(type)
, m_start(start)
, m_end(end)
+ , m_text(text)
{
}
@@ -140,14 +142,16 @@ struct Token {
const Position& start() const { return m_start; }
const Position& end() const { return m_end; }
- void set_start(const Position& other) {m_start = other;}
- void set_end(const Position& other) {m_end = other;}
+ void set_start(const Position& other) { m_start = other; }
+ void set_end(const Position& other) { m_end = other; }
Type type() const { return m_type; }
+ const StringView& text() const { return m_text; }
private:
Type m_type { Type::Unknown };
Position m_start;
Position m_end;
+ StringView m_text;
};
}