mirror of
https://github.com/godotengine/godot.git
synced 2024-11-27 09:16:35 +08:00
3e78963bb9
Using codespell 1.17.1. Method: ``` $ cat > ../godot-word-whitelist.txt << EOF ang curvelinear dof doubleclick fave findn leapyear lod merchantibility nd numer ois ony que seeked synching te uint unselect webp EOF $ codespell -w -q 3 -I ../godot-word-whitelist.txt --skip="./thirdparty,*.po" $ git diff // undo unwanted changes ```
1414 lines
37 KiB
C++
1414 lines
37 KiB
C++
/*************************************************************************/
|
|
/* gdscript_tokenizer.cpp */
|
|
/*************************************************************************/
|
|
/* This file is part of: */
|
|
/* GODOT ENGINE */
|
|
/* https://godotengine.org */
|
|
/*************************************************************************/
|
|
/* Copyright (c) 2007-2020 Juan Linietsky, Ariel Manzur. */
|
|
/* Copyright (c) 2014-2020 Godot Engine contributors (cf. AUTHORS.md). */
|
|
/* */
|
|
/* Permission is hereby granted, free of charge, to any person obtaining */
|
|
/* a copy of this software and associated documentation files (the */
|
|
/* "Software"), to deal in the Software without restriction, including */
|
|
/* without limitation the rights to use, copy, modify, merge, publish, */
|
|
/* distribute, sublicense, and/or sell copies of the Software, and to */
|
|
/* permit persons to whom the Software is furnished to do so, subject to */
|
|
/* the following conditions: */
|
|
/* */
|
|
/* The above copyright notice and this permission notice shall be */
|
|
/* included in all copies or substantial portions of the Software. */
|
|
/* */
|
|
/* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */
|
|
/* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */
|
|
/* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.*/
|
|
/* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */
|
|
/* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */
|
|
/* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */
|
|
/* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */
|
|
/*************************************************************************/
|
|
|
|
#include "gdscript_tokenizer.h"
|
|
|
|
#include "core/error_macros.h"
|
|
|
|
#ifdef TOOLS_ENABLED
|
|
#include "editor/editor_settings.h"
|
|
#endif
|
|
|
|
static const char *token_names[] = {
|
|
"Empty", // EMPTY,
|
|
// Basic
|
|
"Annotation", // ANNOTATION
|
|
"Identifier", // IDENTIFIER,
|
|
"Literal", // LITERAL,
|
|
// Comparison
|
|
"<", // LESS,
|
|
"<=", // LESS_EQUAL,
|
|
">", // GREATER,
|
|
">=", // GREATER_EQUAL,
|
|
"==", // EQUAL_EQUAL,
|
|
"!=", // BANG_EQUAL,
|
|
// Logical
|
|
"and", // AND,
|
|
"or", // OR,
|
|
"not", // NOT,
|
|
"&&", // AMPERSAND_AMPERSAND,
|
|
"||", // PIPE_PIPE,
|
|
"!", // BANG,
|
|
// Bitwise
|
|
"&", // AMPERSAND,
|
|
"|", // PIPE,
|
|
"~", // TILDE,
|
|
"^", // CARET,
|
|
"<<", // LESS_LESS,
|
|
">>", // GREATER_GREATER,
|
|
// Math
|
|
"+", // PLUS,
|
|
"-", // MINUS,
|
|
"*", // STAR,
|
|
"/", // SLASH,
|
|
"%", // PERCENT,
|
|
// Assignment
|
|
"=", // EQUAL,
|
|
"+=", // PLUS_EQUAL,
|
|
"-=", // MINUS_EQUAL,
|
|
"*=", // STAR_EQUAL,
|
|
"/=", // SLASH_EQUAL,
|
|
"%=", // PERCENT_EQUAL,
|
|
"<<=", // LESS_LESS_EQUAL,
|
|
">>=", // GREATER_GREATER_EQUAL,
|
|
"&=", // AMPERSAND_EQUAL,
|
|
"|=", // PIPE_EQUAL,
|
|
"^=", // CARET_EQUAL,
|
|
// Control flow
|
|
"if", // IF,
|
|
"elif", // ELIF,
|
|
"else", // ELSE,
|
|
"for", // FOR,
|
|
"while", // WHILE,
|
|
"break", // BREAK,
|
|
"continue", // CONTINUE,
|
|
"pass", // PASS,
|
|
"return", // RETURN,
|
|
"match", // MATCH,
|
|
// Keywords
|
|
"as", // AS,
|
|
"assert", // ASSERT,
|
|
"await", // AWAIT,
|
|
"breakpoint", // BREAKPOINT,
|
|
"class", // CLASS,
|
|
"class_name", // CLASS_NAME,
|
|
"const", // CONST,
|
|
"enum", // ENUM,
|
|
"extends", // EXTENDS,
|
|
"func", // FUNC,
|
|
"in", // IN,
|
|
"is", // IS,
|
|
"namespace", // NAMESPACE
|
|
"preload", // PRELOAD,
|
|
"self", // SELF,
|
|
"signal", // SIGNAL,
|
|
"static", // STATIC,
|
|
"super", // SUPER,
|
|
"trait", // TRAIT,
|
|
"var", // VAR,
|
|
"void", // VOID,
|
|
"yield", // YIELD,
|
|
// Punctuation
|
|
"[", // BRACKET_OPEN,
|
|
"]", // BRACKET_CLOSE,
|
|
"{", // BRACE_OPEN,
|
|
"}", // BRACE_CLOSE,
|
|
"(", // PARENTHESIS_OPEN,
|
|
")", // PARENTHESIS_CLOSE,
|
|
",", // COMMA,
|
|
";", // SEMICOLON,
|
|
".", // PERIOD,
|
|
"..", // PERIOD_PERIOD,
|
|
":", // COLON,
|
|
"$", // DOLLAR,
|
|
"->", // FORWARD_ARROW,
|
|
"_", // UNDERSCORE,
|
|
// Whitespace
|
|
"Newline", // NEWLINE,
|
|
"Indent", // INDENT,
|
|
"Dedent", // DEDENT,
|
|
// Constants
|
|
"PI", // CONST_PI,
|
|
"TAU", // CONST_TAU,
|
|
"INF", // CONST_INF,
|
|
"NaN", // CONST_NAN,
|
|
// Error message improvement
|
|
"VCS conflict marker", // VCS_CONFLICT_MARKER,
|
|
"`", // BACKTICK,
|
|
"?", // QUESTION_MARK,
|
|
// Special
|
|
"Error", // ERROR,
|
|
"End of file", // EOF,
|
|
};
|
|
|
|
// Avoid desync.
|
|
static_assert(sizeof(token_names) / sizeof(token_names[0]) == GDScriptTokenizer::Token::TK_MAX, "Amount of token names don't match the amount of token types.");
|
|
|
|
const char *GDScriptTokenizer::Token::get_name() const {
|
|
ERR_FAIL_INDEX_V_MSG(type, TK_MAX, "<error>", "Using token type out of the enum.");
|
|
return token_names[type];
|
|
}
|
|
|
|
bool GDScriptTokenizer::Token::is_identifier() const {
|
|
// Note: Most keywords should not be recognized as identifiers.
|
|
// These are only exceptions for stuff that already is on the engine's API.
|
|
switch (type) {
|
|
case IDENTIFIER:
|
|
case MATCH: // Used in String.match().
|
|
return true;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
bool GDScriptTokenizer::Token::is_node_name() const {
|
|
// This is meant to allow keywords with the $ notation, but not as general identifiers.
|
|
switch (type) {
|
|
case IDENTIFIER:
|
|
case AND:
|
|
case AS:
|
|
case ASSERT:
|
|
case AWAIT:
|
|
case BREAK:
|
|
case BREAKPOINT:
|
|
case CLASS_NAME:
|
|
case CLASS:
|
|
case CONST:
|
|
case CONTINUE:
|
|
case ELIF:
|
|
case ELSE:
|
|
case ENUM:
|
|
case EXTENDS:
|
|
case FOR:
|
|
case FUNC:
|
|
case IF:
|
|
case IN:
|
|
case IS:
|
|
case MATCH:
|
|
case NAMESPACE:
|
|
case NOT:
|
|
case OR:
|
|
case PASS:
|
|
case PRELOAD:
|
|
case RETURN:
|
|
case SELF:
|
|
case SIGNAL:
|
|
case STATIC:
|
|
case SUPER:
|
|
case TRAIT:
|
|
case UNDERSCORE:
|
|
case VAR:
|
|
case VOID:
|
|
case WHILE:
|
|
case YIELD:
|
|
return true;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
String GDScriptTokenizer::get_token_name(Token::Type p_token_type) {
|
|
ERR_FAIL_INDEX_V_MSG(p_token_type, Token::TK_MAX, "<error>", "Using token type out of the enum.");
|
|
return token_names[p_token_type];
|
|
}
|
|
|
|
void GDScriptTokenizer::set_source_code(const String &p_source_code) {
|
|
source = p_source_code;
|
|
if (source.empty()) {
|
|
_source = U"";
|
|
} else {
|
|
_source = source.ptr();
|
|
}
|
|
_current = _source;
|
|
line = 1;
|
|
column = 1;
|
|
length = p_source_code.length();
|
|
position = 0;
|
|
}
|
|
|
|
void GDScriptTokenizer::set_cursor_position(int p_line, int p_column) {
|
|
cursor_line = p_line;
|
|
cursor_column = p_column;
|
|
}
|
|
|
|
void GDScriptTokenizer::set_multiline_mode(bool p_state) {
|
|
multiline_mode = p_state;
|
|
}
|
|
|
|
int GDScriptTokenizer::get_cursor_line() const {
|
|
return cursor_line;
|
|
}
|
|
|
|
int GDScriptTokenizer::get_cursor_column() const {
|
|
return cursor_column;
|
|
}
|
|
|
|
bool GDScriptTokenizer::is_past_cursor() const {
|
|
if (line < cursor_line) {
|
|
return false;
|
|
}
|
|
if (line > cursor_line) {
|
|
return true;
|
|
}
|
|
if (column < cursor_column) {
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
char32_t GDScriptTokenizer::_advance() {
|
|
if (unlikely(_is_at_end())) {
|
|
return '\0';
|
|
}
|
|
_current++;
|
|
column++;
|
|
position++;
|
|
if (column > rightmost_column) {
|
|
rightmost_column = column;
|
|
}
|
|
if (unlikely(_is_at_end())) {
|
|
// Add extra newline even if it's not there, to satisfy the parser.
|
|
newline(true);
|
|
// Also add needed unindent.
|
|
check_indent();
|
|
}
|
|
return _peek(-1);
|
|
}
|
|
|
|
void GDScriptTokenizer::push_paren(char32_t p_char) {
|
|
paren_stack.push_back(p_char);
|
|
}
|
|
|
|
bool GDScriptTokenizer::pop_paren(char32_t p_expected) {
|
|
if (paren_stack.empty()) {
|
|
return false;
|
|
}
|
|
char32_t actual = paren_stack.back()->get();
|
|
paren_stack.pop_back();
|
|
|
|
return actual == p_expected;
|
|
}
|
|
|
|
GDScriptTokenizer::Token GDScriptTokenizer::pop_error() {
|
|
Token error = error_stack.back()->get();
|
|
error_stack.pop_back();
|
|
return error;
|
|
}
|
|
|
|
static bool _is_alphanumeric(char32_t c) {
|
|
return (c >= 'a' && c <= 'z') || (c >= 'A' && c <= 'Z') || (c >= '0' && c <= '9') || c == '_';
|
|
}
|
|
|
|
static bool _is_digit(char32_t c) {
|
|
return (c >= '0' && c <= '9');
|
|
}
|
|
|
|
static bool _is_hex_digit(char32_t c) {
|
|
return (c >= '0' && c <= '9') || (c >= 'a' && c <= 'f') || (c >= 'A' && c <= 'F');
|
|
}
|
|
|
|
static bool _is_binary_digit(char32_t c) {
|
|
return (c == '0' || c == '1');
|
|
}
|
|
|
|
GDScriptTokenizer::Token GDScriptTokenizer::make_token(Token::Type p_type) {
|
|
Token token(p_type);
|
|
token.start_line = start_line;
|
|
token.end_line = line;
|
|
token.start_column = start_column;
|
|
token.end_column = column;
|
|
token.leftmost_column = leftmost_column;
|
|
token.rightmost_column = rightmost_column;
|
|
token.source = String(_start, _current - _start);
|
|
|
|
if (p_type != Token::ERROR && cursor_line > -1) {
|
|
// Also count whitespace after token.
|
|
int offset = 0;
|
|
while (_peek(offset) == ' ' || _peek(offset) == '\t') {
|
|
offset++;
|
|
}
|
|
int last_column = column + offset;
|
|
// Check cursor position in token.
|
|
if (start_line == line) {
|
|
// Single line token.
|
|
if (cursor_line == start_line && cursor_column >= start_column && cursor_column <= last_column) {
|
|
token.cursor_position = cursor_column - start_column;
|
|
if (cursor_column == start_column) {
|
|
token.cursor_place = CURSOR_BEGINNING;
|
|
} else if (cursor_column < column) {
|
|
token.cursor_place = CURSOR_MIDDLE;
|
|
} else {
|
|
token.cursor_place = CURSOR_END;
|
|
}
|
|
}
|
|
} else {
|
|
// Multi line token.
|
|
if (cursor_line == start_line && cursor_column >= start_column) {
|
|
// Is in first line.
|
|
token.cursor_position = cursor_column - start_column;
|
|
if (cursor_column == start_column) {
|
|
token.cursor_place = CURSOR_BEGINNING;
|
|
} else {
|
|
token.cursor_place = CURSOR_MIDDLE;
|
|
}
|
|
} else if (cursor_line == line && cursor_column <= last_column) {
|
|
// Is in last line.
|
|
token.cursor_position = cursor_column - start_column;
|
|
if (cursor_column < column) {
|
|
token.cursor_place = CURSOR_MIDDLE;
|
|
} else {
|
|
token.cursor_place = CURSOR_END;
|
|
}
|
|
} else if (cursor_line > start_line && cursor_line < line) {
|
|
// Is in middle line.
|
|
token.cursor_position = CURSOR_MIDDLE;
|
|
}
|
|
}
|
|
}
|
|
|
|
return token;
|
|
}
|
|
|
|
GDScriptTokenizer::Token GDScriptTokenizer::make_literal(const Variant &p_literal) {
|
|
Token token = make_token(Token::LITERAL);
|
|
token.literal = p_literal;
|
|
return token;
|
|
}
|
|
|
|
GDScriptTokenizer::Token GDScriptTokenizer::make_identifier(const StringName &p_identifier) {
|
|
Token identifier = make_token(Token::IDENTIFIER);
|
|
identifier.literal = p_identifier;
|
|
return identifier;
|
|
}
|
|
|
|
GDScriptTokenizer::Token GDScriptTokenizer::make_error(const String &p_message) {
|
|
Token error = make_token(Token::ERROR);
|
|
error.literal = p_message;
|
|
|
|
return error;
|
|
}
|
|
|
|
void GDScriptTokenizer::push_error(const String &p_message) {
|
|
Token error = make_error(p_message);
|
|
error_stack.push_back(error);
|
|
}
|
|
|
|
void GDScriptTokenizer::push_error(const Token &p_error) {
|
|
error_stack.push_back(p_error);
|
|
}
|
|
|
|
GDScriptTokenizer::Token GDScriptTokenizer::make_paren_error(char32_t p_paren) {
|
|
if (paren_stack.empty()) {
|
|
return make_error(vformat("Closing \"%c\" doesn't have an opening counterpart.", p_paren));
|
|
}
|
|
Token error = make_error(vformat("Closing \"%c\" doesn't match the opening \"%c\".", p_paren, paren_stack.back()->get()));
|
|
paren_stack.pop_back(); // Remove opening one anyway.
|
|
return error;
|
|
}
|
|
|
|
GDScriptTokenizer::Token GDScriptTokenizer::check_vcs_marker(char32_t p_test, Token::Type p_double_type) {
|
|
const char32_t *next = _current + 1;
|
|
int chars = 2; // Two already matched.
|
|
|
|
// Test before consuming characters, since we don't want to consume more than needed.
|
|
while (*next == p_test) {
|
|
chars++;
|
|
next++;
|
|
}
|
|
if (chars >= 7) {
|
|
// It is a VCS conflict marker.
|
|
while (chars > 1) {
|
|
// Consume all characters (first was already consumed by scan()).
|
|
_advance();
|
|
chars--;
|
|
}
|
|
return make_token(Token::VCS_CONFLICT_MARKER);
|
|
} else {
|
|
// It is only a regular double character token, so we consume the second character.
|
|
_advance();
|
|
return make_token(p_double_type);
|
|
}
|
|
}
|
|
|
|
GDScriptTokenizer::Token GDScriptTokenizer::annotation() {
|
|
if (!_is_alphanumeric(_peek())) {
|
|
push_error("Expected annotation identifier after \"@\".");
|
|
}
|
|
while (_is_alphanumeric(_peek())) {
|
|
// Consume all identifier characters.
|
|
_advance();
|
|
}
|
|
Token annotation = make_token(Token::ANNOTATION);
|
|
annotation.literal = StringName(annotation.source);
|
|
return annotation;
|
|
}
|
|
|
|
GDScriptTokenizer::Token GDScriptTokenizer::potential_identifier() {
|
|
#define KEYWORDS(KEYWORD_GROUP, KEYWORD) \
|
|
KEYWORD_GROUP('a') \
|
|
KEYWORD("as", Token::AS) \
|
|
KEYWORD("and", Token::AND) \
|
|
KEYWORD("assert", Token::ASSERT) \
|
|
KEYWORD("await", Token::AWAIT) \
|
|
KEYWORD_GROUP('b') \
|
|
KEYWORD("break", Token::BREAK) \
|
|
KEYWORD("breakpoint", Token::BREAKPOINT) \
|
|
KEYWORD_GROUP('c') \
|
|
KEYWORD("class", Token::CLASS) \
|
|
KEYWORD("class_name", Token::CLASS_NAME) \
|
|
KEYWORD("const", Token::CONST) \
|
|
KEYWORD("continue", Token::CONTINUE) \
|
|
KEYWORD_GROUP('e') \
|
|
KEYWORD("elif", Token::ELIF) \
|
|
KEYWORD("else", Token::ELSE) \
|
|
KEYWORD("enum", Token::ENUM) \
|
|
KEYWORD("extends", Token::EXTENDS) \
|
|
KEYWORD_GROUP('f') \
|
|
KEYWORD("for", Token::FOR) \
|
|
KEYWORD("func", Token::FUNC) \
|
|
KEYWORD_GROUP('i') \
|
|
KEYWORD("if", Token::IF) \
|
|
KEYWORD("in", Token::IN) \
|
|
KEYWORD("is", Token::IS) \
|
|
KEYWORD_GROUP('m') \
|
|
KEYWORD("match", Token::MATCH) \
|
|
KEYWORD_GROUP('n') \
|
|
KEYWORD("namespace", Token::NAMESPACE) \
|
|
KEYWORD("not", Token::NOT) \
|
|
KEYWORD_GROUP('o') \
|
|
KEYWORD("or", Token::OR) \
|
|
KEYWORD_GROUP('p') \
|
|
KEYWORD("pass", Token::PASS) \
|
|
KEYWORD("preload", Token::PRELOAD) \
|
|
KEYWORD_GROUP('r') \
|
|
KEYWORD("return", Token::RETURN) \
|
|
KEYWORD_GROUP('s') \
|
|
KEYWORD("self", Token::SELF) \
|
|
KEYWORD("signal", Token::SIGNAL) \
|
|
KEYWORD("static", Token::STATIC) \
|
|
KEYWORD("super", Token::SUPER) \
|
|
KEYWORD_GROUP('t') \
|
|
KEYWORD("trait", Token::TRAIT) \
|
|
KEYWORD_GROUP('v') \
|
|
KEYWORD("var", Token::VAR) \
|
|
KEYWORD("void", Token::VOID) \
|
|
KEYWORD_GROUP('w') \
|
|
KEYWORD("while", Token::WHILE) \
|
|
KEYWORD_GROUP('y') \
|
|
KEYWORD("yield", Token::YIELD) \
|
|
KEYWORD_GROUP('I') \
|
|
KEYWORD("INF", Token::CONST_INF) \
|
|
KEYWORD_GROUP('N') \
|
|
KEYWORD("NAN", Token::CONST_NAN) \
|
|
KEYWORD_GROUP('P') \
|
|
KEYWORD("PI", Token::CONST_PI) \
|
|
KEYWORD_GROUP('T') \
|
|
KEYWORD("TAU", Token::CONST_TAU)
|
|
|
|
#define MIN_KEYWORD_LENGTH 2
|
|
#define MAX_KEYWORD_LENGTH 10
|
|
|
|
// Consume all alphanumeric characters.
|
|
while (_is_alphanumeric(_peek())) {
|
|
_advance();
|
|
}
|
|
|
|
int length = _current - _start;
|
|
|
|
if (length == 1 && _peek(-1) == '_') {
|
|
// Lone underscore.
|
|
return make_token(Token::UNDERSCORE);
|
|
}
|
|
|
|
String name(_start, length);
|
|
if (length < MIN_KEYWORD_LENGTH || length > MAX_KEYWORD_LENGTH) {
|
|
// Cannot be a keyword, as the length doesn't match any.
|
|
return make_identifier(name);
|
|
}
|
|
|
|
// Define some helper macros for the switch case.
|
|
#define KEYWORD_GROUP_CASE(char) \
|
|
break; \
|
|
case char:
|
|
#define KEYWORD(keyword, token_type) \
|
|
{ \
|
|
const int keyword_length = sizeof(keyword) - 1; \
|
|
static_assert(keyword_length <= MAX_KEYWORD_LENGTH, "There's a keyword longer than the defined maximum length"); \
|
|
static_assert(keyword_length >= MIN_KEYWORD_LENGTH, "There's a keyword shorter than the defined minimum length"); \
|
|
if (keyword_length == length && name == keyword) { \
|
|
return make_token(token_type); \
|
|
} \
|
|
}
|
|
|
|
// Find if it's a keyword.
|
|
switch (_start[0]) {
|
|
default:
|
|
KEYWORDS(KEYWORD_GROUP_CASE, KEYWORD)
|
|
break;
|
|
}
|
|
|
|
// Check if it's a special literal
|
|
if (length == 4) {
|
|
if (name == "true") {
|
|
return make_literal(true);
|
|
} else if (name == "null") {
|
|
return make_literal(Variant());
|
|
}
|
|
} else if (length == 5) {
|
|
if (name == "false") {
|
|
return make_literal(false);
|
|
}
|
|
}
|
|
|
|
// Not a keyword, so must be an identifier.
|
|
return make_identifier(name);
|
|
|
|
#undef KEYWORDS
|
|
#undef MIN_KEYWORD_LENGTH
|
|
#undef MAX_KEYWORD_LENGTH
|
|
#undef KEYWORD_GROUP_CASE
|
|
#undef KEYWORD
|
|
}
|
|
|
|
void GDScriptTokenizer::newline(bool p_make_token) {
|
|
// Don't overwrite previous newline, nor create if we want a line continuation.
|
|
if (p_make_token && !pending_newline && !line_continuation) {
|
|
Token newline(Token::NEWLINE);
|
|
newline.start_line = line;
|
|
newline.end_line = line;
|
|
newline.start_column = column - 1;
|
|
newline.end_column = column;
|
|
newline.leftmost_column = newline.start_column;
|
|
newline.rightmost_column = newline.end_column;
|
|
pending_newline = true;
|
|
last_newline = newline;
|
|
}
|
|
|
|
// Increment line/column counters.
|
|
line++;
|
|
column = 1;
|
|
leftmost_column = 1;
|
|
}
|
|
|
|
GDScriptTokenizer::Token GDScriptTokenizer::number() {
|
|
int base = 10;
|
|
bool has_decimal = false;
|
|
bool has_exponent = false;
|
|
bool has_error = false;
|
|
bool (*digit_check_func)(char32_t) = _is_digit;
|
|
|
|
if (_peek(-1) == '.') {
|
|
has_decimal = true;
|
|
} else if (_peek(-1) == '0') {
|
|
if (_peek() == 'x') {
|
|
// Hexadecimal.
|
|
base = 16;
|
|
digit_check_func = _is_hex_digit;
|
|
_advance();
|
|
} else if (_peek() == 'b') {
|
|
// Binary.
|
|
base = 2;
|
|
digit_check_func = _is_binary_digit;
|
|
_advance();
|
|
}
|
|
}
|
|
|
|
// Allow '_' to be used in a number, for readability.
|
|
bool previous_was_underscore = false;
|
|
while (digit_check_func(_peek()) || _peek() == '_') {
|
|
if (_peek() == '_') {
|
|
if (previous_was_underscore) {
|
|
Token error = make_error(R"(Only one underscore can be used as a numeric separator.)");
|
|
error.start_column = column;
|
|
error.leftmost_column = column;
|
|
error.end_column = column + 1;
|
|
error.rightmost_column = column + 1;
|
|
push_error(error);
|
|
}
|
|
previous_was_underscore = true;
|
|
}
|
|
_advance();
|
|
}
|
|
|
|
// It might be a ".." token (instead of decimal point) so we check if it's not.
|
|
if (_peek() == '.' && _peek(1) != '.') {
|
|
if (base == 10 && !has_decimal) {
|
|
has_decimal = true;
|
|
} else if (base == 10) {
|
|
Token error = make_error("Cannot use a decimal point twice in a number.");
|
|
error.start_column = column;
|
|
error.leftmost_column = column;
|
|
error.end_column = column + 1;
|
|
error.rightmost_column = column + 1;
|
|
push_error(error);
|
|
has_error = true;
|
|
} else if (base == 16) {
|
|
Token error = make_error("Cannot use a decimal point in a hexadecimal number.");
|
|
error.start_column = column;
|
|
error.leftmost_column = column;
|
|
error.end_column = column + 1;
|
|
error.rightmost_column = column + 1;
|
|
push_error(error);
|
|
has_error = true;
|
|
} else {
|
|
Token error = make_error("Cannot use a decimal point in a binary number.");
|
|
error.start_column = column;
|
|
error.leftmost_column = column;
|
|
error.end_column = column + 1;
|
|
error.rightmost_column = column + 1;
|
|
push_error(error);
|
|
has_error = true;
|
|
}
|
|
if (!has_error) {
|
|
_advance();
|
|
|
|
// Consume decimal digits.
|
|
while (_is_digit(_peek()) || _peek() == '_') {
|
|
_advance();
|
|
}
|
|
}
|
|
}
|
|
if (base == 10) {
|
|
if (_peek() == 'e' || _peek() == 'E') {
|
|
has_exponent = true;
|
|
_advance();
|
|
if (_peek() == '+' || _peek() == '-') {
|
|
// Exponent sign.
|
|
_advance();
|
|
}
|
|
// Consume exponent digits.
|
|
if (!_is_digit(_peek())) {
|
|
Token error = make_error(R"(Expected exponent value after "e".)");
|
|
error.start_column = column;
|
|
error.leftmost_column = column;
|
|
error.end_column = column + 1;
|
|
error.rightmost_column = column + 1;
|
|
push_error(error);
|
|
}
|
|
previous_was_underscore = false;
|
|
while (_is_digit(_peek()) || _peek() == '_') {
|
|
if (_peek() == '_') {
|
|
if (previous_was_underscore) {
|
|
Token error = make_error(R"(Only one underscore can be used as a numeric separator.)");
|
|
error.start_column = column;
|
|
error.leftmost_column = column;
|
|
error.end_column = column + 1;
|
|
error.rightmost_column = column + 1;
|
|
push_error(error);
|
|
}
|
|
previous_was_underscore = true;
|
|
}
|
|
_advance();
|
|
}
|
|
}
|
|
}
|
|
|
|
// Detect extra decimal point.
|
|
if (!has_error && has_decimal && _peek() == '.' && _peek(1) != '.') {
|
|
Token error = make_error("Cannot use a decimal point twice in a number.");
|
|
error.start_column = column;
|
|
error.leftmost_column = column;
|
|
error.end_column = column + 1;
|
|
error.rightmost_column = column + 1;
|
|
push_error(error);
|
|
has_error = true;
|
|
} else if (_is_alphanumeric(_peek())) {
|
|
// Letter at the end of the number.
|
|
push_error("Invalid numeric notation.");
|
|
}
|
|
|
|
// Create a string with the whole number.
|
|
int length = _current - _start;
|
|
String number = String(_start, length).replace("_", "");
|
|
|
|
// Convert to the appropriate literal type.
|
|
if (base == 16) {
|
|
int64_t value = number.hex_to_int();
|
|
return make_literal(value);
|
|
} else if (base == 2) {
|
|
int64_t value = number.bin_to_int();
|
|
return make_literal(value);
|
|
} else if (has_decimal || has_exponent) {
|
|
double value = number.to_float();
|
|
return make_literal(value);
|
|
} else {
|
|
int64_t value = number.to_int();
|
|
return make_literal(value);
|
|
}
|
|
}
|
|
|
|
GDScriptTokenizer::Token GDScriptTokenizer::string() {
|
|
enum StringType {
|
|
STRING_REGULAR,
|
|
STRING_NAME,
|
|
STRING_NODEPATH,
|
|
};
|
|
|
|
bool is_multiline = false;
|
|
StringType type = STRING_REGULAR;
|
|
|
|
if (_peek(-1) == '&') {
|
|
type = STRING_NAME;
|
|
_advance();
|
|
} else if (_peek(-1) == '^') {
|
|
type = STRING_NODEPATH;
|
|
_advance();
|
|
}
|
|
|
|
char32_t quote_char = _peek(-1);
|
|
|
|
if (_peek() == quote_char && _peek(1) == quote_char) {
|
|
is_multiline = true;
|
|
// Consume all quotes.
|
|
_advance();
|
|
_advance();
|
|
}
|
|
|
|
String result;
|
|
|
|
for (;;) {
|
|
// Consume actual string.
|
|
if (_is_at_end()) {
|
|
return make_error("Unterminated string.");
|
|
}
|
|
|
|
char32_t ch = _peek();
|
|
|
|
if (ch == '\\') {
|
|
// Escape pattern.
|
|
_advance();
|
|
if (_is_at_end()) {
|
|
return make_error("Unterminated string.");
|
|
}
|
|
|
|
// Grab escape character.
|
|
char32_t code = _peek();
|
|
_advance();
|
|
if (_is_at_end()) {
|
|
return make_error("Unterminated string.");
|
|
}
|
|
|
|
char32_t escaped = 0;
|
|
bool valid_escape = true;
|
|
|
|
switch (code) {
|
|
case 'a':
|
|
escaped = '\a';
|
|
break;
|
|
case 'b':
|
|
escaped = '\b';
|
|
break;
|
|
case 'f':
|
|
escaped = '\f';
|
|
break;
|
|
case 'n':
|
|
escaped = '\n';
|
|
break;
|
|
case 'r':
|
|
escaped = '\r';
|
|
break;
|
|
case 't':
|
|
escaped = '\t';
|
|
break;
|
|
case 'v':
|
|
escaped = '\v';
|
|
break;
|
|
case '\'':
|
|
escaped = '\'';
|
|
break;
|
|
case '\"':
|
|
escaped = '\"';
|
|
break;
|
|
case '\\':
|
|
escaped = '\\';
|
|
break;
|
|
case 'u':
|
|
// Hexadecimal sequence.
|
|
for (int i = 0; i < 4; i++) {
|
|
if (_is_at_end()) {
|
|
return make_error("Unterminated string.");
|
|
}
|
|
|
|
char32_t digit = _peek();
|
|
char32_t value = 0;
|
|
if (digit >= '0' && digit <= '9') {
|
|
value = digit - '0';
|
|
} else if (digit >= 'a' && digit <= 'f') {
|
|
value = digit - 'a';
|
|
value += 10;
|
|
} else if (digit >= 'A' && digit <= 'F') {
|
|
value = digit - 'A';
|
|
value += 10;
|
|
} else {
|
|
// Make error, but keep parsing the string.
|
|
Token error = make_error("Invalid hexadecimal digit in unicode escape sequence.");
|
|
error.start_column = column;
|
|
error.leftmost_column = error.start_column;
|
|
error.end_column = column + 1;
|
|
error.rightmost_column = error.end_column;
|
|
push_error(error);
|
|
valid_escape = false;
|
|
break;
|
|
}
|
|
|
|
escaped <<= 4;
|
|
escaped |= value;
|
|
|
|
_advance();
|
|
}
|
|
break;
|
|
case '\r':
|
|
if (_peek() != '\n') {
|
|
// Carriage return without newline in string. (???)
|
|
// Just add it to the string and keep going.
|
|
result += ch;
|
|
_advance();
|
|
break;
|
|
}
|
|
[[fallthrough]];
|
|
case '\n':
|
|
// Escaping newline.
|
|
newline(false);
|
|
valid_escape = false; // Don't add to the string.
|
|
break;
|
|
default:
|
|
Token error = make_error("Invalid escape in string.");
|
|
error.start_column = column - 2;
|
|
error.leftmost_column = error.start_column;
|
|
push_error(error);
|
|
valid_escape = false;
|
|
break;
|
|
}
|
|
|
|
if (valid_escape) {
|
|
result += escaped;
|
|
}
|
|
} else if (ch == quote_char) {
|
|
_advance();
|
|
if (is_multiline) {
|
|
if (_peek() == quote_char && _peek(1) == quote_char) {
|
|
// Ended the multiline string. Consume all quotes.
|
|
_advance();
|
|
_advance();
|
|
break;
|
|
}
|
|
} else {
|
|
// Ended single-line string.
|
|
break;
|
|
}
|
|
} else {
|
|
result += ch;
|
|
_advance();
|
|
if (ch == '\n') {
|
|
newline(false);
|
|
}
|
|
}
|
|
}
|
|
|
|
// Make the literal.
|
|
Variant string;
|
|
switch (type) {
|
|
case STRING_NAME:
|
|
string = StringName(result);
|
|
break;
|
|
case STRING_NODEPATH:
|
|
string = NodePath(result);
|
|
break;
|
|
case STRING_REGULAR:
|
|
string = result;
|
|
break;
|
|
}
|
|
|
|
return make_literal(string);
|
|
}
|
|
|
|
void GDScriptTokenizer::check_indent() {
|
|
ERR_FAIL_COND_MSG(column != 1, "Checking tokenizer indentation in the middle of a line.");
|
|
|
|
if (_is_at_end()) {
|
|
// Send dedents for every indent level.
|
|
pending_indents -= indent_level();
|
|
indent_stack.clear();
|
|
return;
|
|
}
|
|
|
|
for (;;) {
|
|
char32_t current_indent_char = _peek();
|
|
int indent_count = 0;
|
|
|
|
if (current_indent_char != ' ' && current_indent_char != '\t' && current_indent_char != '\r' && current_indent_char != '\n' && current_indent_char != '#') {
|
|
// First character of the line is not whitespace, so we clear all indentation levels.
|
|
// Unless we are in a continuation or in multiline mode (inside expression).
|
|
if (line_continuation || multiline_mode) {
|
|
return;
|
|
}
|
|
pending_indents -= indent_level();
|
|
indent_stack.clear();
|
|
return;
|
|
}
|
|
|
|
if (_peek() == '\r') {
|
|
_advance();
|
|
if (_peek() != '\n') {
|
|
push_error("Stray carriage return character in source code.");
|
|
}
|
|
}
|
|
if (_peek() == '\n') {
|
|
// Empty line, keep going.
|
|
_advance();
|
|
newline(false);
|
|
continue;
|
|
}
|
|
|
|
// Check indent level.
|
|
bool mixed = false;
|
|
while (!_is_at_end()) {
|
|
char32_t space = _peek();
|
|
if (space == '\t') {
|
|
// Consider individual tab columns.
|
|
column += tab_size - 1;
|
|
indent_count += tab_size;
|
|
} else if (space == ' ') {
|
|
indent_count += 1;
|
|
} else {
|
|
break;
|
|
}
|
|
mixed = mixed || space != current_indent_char;
|
|
_advance();
|
|
}
|
|
|
|
if (mixed) {
|
|
Token error = make_error("Mixed use of tabs and spaces for indentation.");
|
|
error.start_line = line;
|
|
error.start_column = 1;
|
|
error.leftmost_column = 1;
|
|
error.rightmost_column = column;
|
|
push_error(error);
|
|
}
|
|
|
|
if (_is_at_end()) {
|
|
// Reached the end with an empty line, so just dedent as much as needed.
|
|
pending_indents -= indent_level();
|
|
indent_stack.clear();
|
|
return;
|
|
}
|
|
|
|
if (_peek() == '\r') {
|
|
_advance();
|
|
if (_peek() != '\n') {
|
|
push_error("Stray carriage return character in source code.");
|
|
}
|
|
}
|
|
if (_peek() == '\n') {
|
|
// Empty line, keep going.
|
|
_advance();
|
|
newline(false);
|
|
continue;
|
|
}
|
|
if (_peek() == '#') {
|
|
// Comment. Advance to the next line.
|
|
while (_peek() != '\n' && !_is_at_end()) {
|
|
_advance();
|
|
}
|
|
if (_is_at_end()) {
|
|
// Reached the end with an empty line, so just dedent as much as needed.
|
|
pending_indents -= indent_level();
|
|
indent_stack.clear();
|
|
return;
|
|
}
|
|
_advance(); // Consume '\n'.
|
|
newline(false);
|
|
continue;
|
|
}
|
|
|
|
if (line_continuation || multiline_mode) {
|
|
// We cleared up all the whitespace at the beginning of the line.
|
|
// But if this is a continuation or multiline mode and we don't want any indentation change.
|
|
return;
|
|
}
|
|
|
|
// Check if indentation character is consistent.
|
|
if (indent_char == '\0') {
|
|
// First time indenting, choose character now.
|
|
indent_char = current_indent_char;
|
|
} else if (current_indent_char != indent_char) {
|
|
Token error = make_error(vformat("Used \"%s\" for indentation instead \"%s\" as used before in the file.", String(¤t_indent_char, 1).c_escape(), String(&indent_char, 1).c_escape()));
|
|
error.start_line = line;
|
|
error.start_column = 1;
|
|
error.leftmost_column = 1;
|
|
error.rightmost_column = column;
|
|
push_error(error);
|
|
}
|
|
|
|
// Now we can do actual indentation changes.
|
|
|
|
// Check if indent or dedent.
|
|
int previous_indent = 0;
|
|
if (indent_level() > 0) {
|
|
previous_indent = indent_stack.back()->get();
|
|
}
|
|
if (indent_count == previous_indent) {
|
|
// No change in indentation.
|
|
return;
|
|
}
|
|
if (indent_count > previous_indent) {
|
|
// Indentation increased.
|
|
indent_stack.push_back(indent_count);
|
|
pending_indents++;
|
|
} else {
|
|
// Indentation decreased (dedent).
|
|
if (indent_level() == 0) {
|
|
push_error("Tokenizer bug: trying to dedent without previous indent.");
|
|
return;
|
|
}
|
|
while (indent_level() > 0 && indent_stack.back()->get() > indent_count) {
|
|
indent_stack.pop_back();
|
|
pending_indents--;
|
|
}
|
|
if ((indent_level() > 0 && indent_stack.back()->get() != indent_count) || (indent_level() == 0 && indent_count != 0)) {
|
|
// Mismatched indentation alignment.
|
|
Token error = make_error("Unindent doesn't match the previous indentation level.");
|
|
error.start_line = line;
|
|
error.start_column = 1;
|
|
error.leftmost_column = 1;
|
|
error.end_column = column + 1;
|
|
error.rightmost_column = column + 1;
|
|
push_error(error);
|
|
// Still, we'll be lenient and keep going, so keep this level in the stack.
|
|
indent_stack.push_back(indent_count);
|
|
}
|
|
}
|
|
break; // Get out of the loop in any case.
|
|
}
|
|
}
|
|
|
|
void GDScriptTokenizer::_skip_whitespace() {
|
|
if (pending_indents != 0) {
|
|
// Still have some indent/dedent tokens to give.
|
|
return;
|
|
}
|
|
|
|
bool is_bol = column == 1; // Beginning of line.
|
|
|
|
if (is_bol) {
|
|
check_indent();
|
|
return;
|
|
}
|
|
|
|
for (;;) {
|
|
char32_t c = _peek();
|
|
switch (c) {
|
|
case ' ':
|
|
_advance();
|
|
break;
|
|
case '\t':
|
|
_advance();
|
|
// Consider individual tab columns.
|
|
column += tab_size - 1;
|
|
break;
|
|
case '\r':
|
|
_advance(); // Consume either way.
|
|
if (_peek() != '\n') {
|
|
push_error("Stray carriage return character in source code.");
|
|
return;
|
|
}
|
|
break;
|
|
case '\n':
|
|
_advance();
|
|
newline(!is_bol); // Don't create new line token if line is empty.
|
|
check_indent();
|
|
break;
|
|
case '#':
|
|
// Comment.
|
|
while (_peek() != '\n' && !_is_at_end()) {
|
|
_advance();
|
|
}
|
|
if (_is_at_end()) {
|
|
return;
|
|
}
|
|
_advance(); // Consume '\n'
|
|
newline(!is_bol);
|
|
check_indent();
|
|
break;
|
|
default:
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
GDScriptTokenizer::Token GDScriptTokenizer::scan() {
|
|
if (has_error()) {
|
|
return pop_error();
|
|
}
|
|
|
|
_skip_whitespace();
|
|
|
|
if (pending_newline) {
|
|
pending_newline = false;
|
|
if (!multiline_mode) {
|
|
// Don't return newline tokens on multine mode.
|
|
return last_newline;
|
|
}
|
|
}
|
|
|
|
// Check for potential errors after skipping whitespace().
|
|
if (has_error()) {
|
|
return pop_error();
|
|
}
|
|
|
|
_start = _current;
|
|
start_line = line;
|
|
start_column = column;
|
|
leftmost_column = column;
|
|
rightmost_column = column;
|
|
|
|
if (pending_indents != 0) {
|
|
// Adjust position for indent.
|
|
_start -= start_column - 1;
|
|
start_column = 1;
|
|
leftmost_column = 1;
|
|
if (pending_indents > 0) {
|
|
// Indents.
|
|
pending_indents--;
|
|
return make_token(Token::INDENT);
|
|
} else {
|
|
// Dedents.
|
|
pending_indents++;
|
|
Token dedent = make_token(Token::DEDENT);
|
|
dedent.end_column += 1;
|
|
dedent.rightmost_column += 1;
|
|
return dedent;
|
|
}
|
|
}
|
|
|
|
if (_is_at_end()) {
|
|
return make_token(Token::TK_EOF);
|
|
}
|
|
|
|
const char32_t c = _advance();
|
|
|
|
if (c == '\\') {
|
|
// Line continuation with backslash.
|
|
if (_peek() == '\r') {
|
|
if (_peek(1) != '\n') {
|
|
return make_error("Unexpected carriage return character.");
|
|
}
|
|
_advance();
|
|
}
|
|
if (_peek() != '\n') {
|
|
return make_error("Expected new line after \"\\\".");
|
|
}
|
|
_advance();
|
|
newline(false);
|
|
line_continuation = true;
|
|
return scan(); // Recurse to get next token.
|
|
}
|
|
|
|
line_continuation = false;
|
|
|
|
if (_is_digit(c)) {
|
|
return number();
|
|
} else if (_is_alphanumeric(c)) {
|
|
return potential_identifier();
|
|
}
|
|
|
|
switch (c) {
|
|
// String literals.
|
|
case '"':
|
|
case '\'':
|
|
return string();
|
|
|
|
// Annotation.
|
|
case '@':
|
|
return annotation();
|
|
|
|
// Single characters.
|
|
case '~':
|
|
return make_token(Token::TILDE);
|
|
case ',':
|
|
return make_token(Token::COMMA);
|
|
case ':':
|
|
return make_token(Token::COLON);
|
|
case ';':
|
|
return make_token(Token::SEMICOLON);
|
|
case '$':
|
|
return make_token(Token::DOLLAR);
|
|
case '?':
|
|
return make_token(Token::QUESTION_MARK);
|
|
case '`':
|
|
return make_token(Token::BACKTICK);
|
|
|
|
// Parens.
|
|
case '(':
|
|
push_paren('(');
|
|
return make_token(Token::PARENTHESIS_OPEN);
|
|
case '[':
|
|
push_paren('[');
|
|
return make_token(Token::BRACKET_OPEN);
|
|
case '{':
|
|
push_paren('{');
|
|
return make_token(Token::BRACE_OPEN);
|
|
case ')':
|
|
if (!pop_paren('(')) {
|
|
return make_paren_error(c);
|
|
}
|
|
return make_token(Token::PARENTHESIS_CLOSE);
|
|
case ']':
|
|
if (!pop_paren('[')) {
|
|
return make_paren_error(c);
|
|
}
|
|
return make_token(Token::BRACKET_CLOSE);
|
|
case '}':
|
|
if (!pop_paren('{')) {
|
|
return make_paren_error(c);
|
|
}
|
|
return make_token(Token::BRACE_CLOSE);
|
|
|
|
// Double characters.
|
|
case '!':
|
|
if (_peek() == '=') {
|
|
_advance();
|
|
return make_token(Token::BANG_EQUAL);
|
|
} else {
|
|
return make_token(Token::BANG);
|
|
}
|
|
case '.':
|
|
if (_peek() == '.') {
|
|
_advance();
|
|
return make_token(Token::PERIOD_PERIOD);
|
|
} else if (_is_digit(_peek())) {
|
|
// Number starting with '.'.
|
|
return number();
|
|
} else {
|
|
return make_token(Token::PERIOD);
|
|
}
|
|
case '+':
|
|
if (_peek() == '=') {
|
|
_advance();
|
|
return make_token(Token::PLUS_EQUAL);
|
|
} else {
|
|
return make_token(Token::PLUS);
|
|
}
|
|
case '-':
|
|
if (_peek() == '=') {
|
|
_advance();
|
|
return make_token(Token::MINUS_EQUAL);
|
|
} else if (_peek() == '>') {
|
|
_advance();
|
|
return make_token(Token::FORWARD_ARROW);
|
|
} else {
|
|
return make_token(Token::MINUS);
|
|
}
|
|
case '*':
|
|
if (_peek() == '=') {
|
|
_advance();
|
|
return make_token(Token::STAR_EQUAL);
|
|
} else {
|
|
return make_token(Token::STAR);
|
|
}
|
|
case '/':
|
|
if (_peek() == '=') {
|
|
_advance();
|
|
return make_token(Token::SLASH_EQUAL);
|
|
} else {
|
|
return make_token(Token::SLASH);
|
|
}
|
|
case '%':
|
|
if (_peek() == '=') {
|
|
_advance();
|
|
return make_token(Token::PERCENT_EQUAL);
|
|
} else {
|
|
return make_token(Token::PERCENT);
|
|
}
|
|
case '^':
|
|
if (_peek() == '=') {
|
|
_advance();
|
|
return make_token(Token::CARET_EQUAL);
|
|
} else if (_peek() == '"' || _peek() == '\'') {
|
|
// Node path
|
|
return string();
|
|
} else {
|
|
return make_token(Token::CARET);
|
|
}
|
|
case '&':
|
|
if (_peek() == '&') {
|
|
_advance();
|
|
return make_token(Token::AMPERSAND_AMPERSAND);
|
|
} else if (_peek() == '=') {
|
|
_advance();
|
|
return make_token(Token::AMPERSAND_EQUAL);
|
|
} else if (_peek() == '"' || _peek() == '\'') {
|
|
// String Name
|
|
return string();
|
|
} else {
|
|
return make_token(Token::AMPERSAND);
|
|
}
|
|
case '|':
|
|
if (_peek() == '|') {
|
|
_advance();
|
|
return make_token(Token::PIPE_PIPE);
|
|
} else if (_peek() == '=') {
|
|
_advance();
|
|
return make_token(Token::PIPE_EQUAL);
|
|
} else {
|
|
return make_token(Token::PIPE);
|
|
}
|
|
|
|
// Potential VCS conflict markers.
|
|
case '=':
|
|
if (_peek() == '=') {
|
|
return check_vcs_marker('=', Token::EQUAL_EQUAL);
|
|
} else {
|
|
return make_token(Token::EQUAL);
|
|
}
|
|
case '<':
|
|
if (_peek() == '=') {
|
|
_advance();
|
|
return make_token(Token::LESS_EQUAL);
|
|
} else if (_peek() == '<') {
|
|
if (_peek(1) == '=') {
|
|
_advance();
|
|
_advance(); // Advance both '<' and '='
|
|
return make_token(Token::LESS_LESS_EQUAL);
|
|
} else {
|
|
return check_vcs_marker('<', Token::LESS_LESS);
|
|
}
|
|
} else {
|
|
return make_token(Token::LESS);
|
|
}
|
|
case '>':
|
|
if (_peek() == '=') {
|
|
_advance();
|
|
return make_token(Token::GREATER_EQUAL);
|
|
} else if (_peek() == '>') {
|
|
if (_peek(1) == '=') {
|
|
_advance();
|
|
_advance(); // Advance both '>' and '='
|
|
return make_token(Token::GREATER_GREATER_EQUAL);
|
|
} else {
|
|
return check_vcs_marker('>', Token::GREATER_GREATER);
|
|
}
|
|
} else {
|
|
return make_token(Token::GREATER);
|
|
}
|
|
|
|
default:
|
|
return make_error(vformat(R"(Unknown character "%s".")", String(&c, 1)));
|
|
}
|
|
}
|
|
|
|
GDScriptTokenizer::GDScriptTokenizer() {
|
|
#ifdef TOOLS_ENABLED
|
|
if (EditorSettings::get_singleton()) {
|
|
tab_size = EditorSettings::get_singleton()->get_setting("text_editor/indent/size");
|
|
}
|
|
#endif // TOOLS_ENABLED
|
|
}
|