Source code for robot.parsing.lexer.tokens

#  Copyright 2008-2015 Nokia Networks
#  Copyright 2016-     Robot Framework Foundation
#
#  Licensed under the Apache License, Version 2.0 (the "License");
#  you may not use this file except in compliance with the License.
#  You may obtain a copy of the License at
#
#      http://www.apache.org/licenses/LICENSE-2.0
#
#  Unless required by applicable law or agreed to in writing, software
#  distributed under the License is distributed on an "AS IS" BASIS,
#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
#  See the License for the specific language governing permissions and
#  limitations under the License.

from collections.abc import Iterator
from typing import cast, List

from robot.variables import VariableIterator


# Type alias to ease typing elsewhere
StatementTokens = List['Token']


[docs]class Token: """Token representing piece of Robot Framework data. Each token has type, value, line number, column offset and end column offset in :attr:`type`, :attr:`value`, :attr:`lineno`, :attr:`col_offset` and :attr:`end_col_offset` attributes, respectively. Tokens representing error also have their error message in :attr:`error` attribute. Token types are declared as class attributes such as :attr:`SETTING_HEADER` and :attr:`EOL`. Values of these constants have changed slightly in Robot Framework 4.0, and they may change again in the future. It is thus safer to use the constants, not their values, when types are needed. For example, use ``Token(Token.EOL)`` instead of ``Token('EOL')`` and ``token.type == Token.EOL`` instead of ``token.type == 'EOL'``. If :attr:`value` is not given and :attr:`type` is a special marker like :attr:`IF` or `:attr:`EOL`, the value is set automatically. """ SETTING_HEADER = 'SETTING HEADER' VARIABLE_HEADER = 'VARIABLE HEADER' TESTCASE_HEADER = 'TESTCASE HEADER' TASK_HEADER = 'TASK HEADER' KEYWORD_HEADER = 'KEYWORD HEADER' COMMENT_HEADER = 'COMMENT HEADER' INVALID_HEADER = 'INVALID HEADER' FATAL_INVALID_HEADER = 'FATAL INVALID HEADER' TESTCASE_NAME = 'TESTCASE NAME' KEYWORD_NAME = 'KEYWORD NAME' SUITE_NAME = 'SUITE NAME' DOCUMENTATION = 'DOCUMENTATION' SUITE_SETUP = 'SUITE SETUP' SUITE_TEARDOWN = 'SUITE TEARDOWN' METADATA = 'METADATA' TEST_SETUP = 'TEST SETUP' TEST_TEARDOWN = 'TEST TEARDOWN' TEST_TEMPLATE = 'TEST TEMPLATE' TEST_TIMEOUT = 'TEST TIMEOUT' FORCE_TAGS = 'FORCE TAGS' DEFAULT_TAGS = 'DEFAULT TAGS' KEYWORD_TAGS = 'KEYWORD TAGS' LIBRARY = 'LIBRARY' RESOURCE = 'RESOURCE' VARIABLES = 'VARIABLES' SETUP = 'SETUP' TEARDOWN = 'TEARDOWN' TEMPLATE = 'TEMPLATE' TIMEOUT = 'TIMEOUT' TAGS = 'TAGS' ARGUMENTS = 'ARGUMENTS' # Use ´RETURN_SETTING` type instead of `RETURN`. `[Return]` is deprecated and # `RETURN` type will be used with `RETURN` statement in the future. RETURN = 'RETURN' RETURN_SETTING = RETURN # TODO: Change WITH_NAME value to AS in RF 7.0. Remove WITH_NAME in RF 8. WITH_NAME = 'WITH NAME' AS = 'AS' NAME = 'NAME' VARIABLE = 'VARIABLE' ARGUMENT = 'ARGUMENT' ASSIGN = 'ASSIGN' KEYWORD = 'KEYWORD' FOR = 'FOR' FOR_SEPARATOR = 'FOR SEPARATOR' END = 'END' IF = 'IF' INLINE_IF = 'INLINE IF' ELSE_IF = 'ELSE IF' ELSE = 'ELSE' TRY = 'TRY' EXCEPT = 'EXCEPT' FINALLY = 'FINALLY' WHILE = 'WHILE' RETURN_STATEMENT = 'RETURN STATEMENT' CONTINUE = 'CONTINUE' BREAK = 'BREAK' OPTION = 'OPTION' SEPARATOR = 'SEPARATOR' COMMENT = 'COMMENT' CONTINUATION = 'CONTINUATION' CONFIG = 'CONFIG' EOL = 'EOL' EOS = 'EOS' ERROR = 'ERROR' # TODO: FATAL_ERROR is no longer used, remove in RF 7.0 FATAL_ERROR = 'FATAL ERROR' NON_DATA_TOKENS = frozenset(( SEPARATOR, COMMENT, CONTINUATION, EOL, EOS )) SETTING_TOKENS = frozenset(( DOCUMENTATION, SUITE_NAME, SUITE_SETUP, SUITE_TEARDOWN, METADATA, TEST_SETUP, TEST_TEARDOWN, TEST_TEMPLATE, TEST_TIMEOUT, FORCE_TAGS, DEFAULT_TAGS, KEYWORD_TAGS, LIBRARY, RESOURCE, VARIABLES, SETUP, TEARDOWN, TEMPLATE, TIMEOUT, TAGS, ARGUMENTS, RETURN )) HEADER_TOKENS = frozenset(( SETTING_HEADER, VARIABLE_HEADER, TESTCASE_HEADER, TASK_HEADER, KEYWORD_HEADER, COMMENT_HEADER, INVALID_HEADER )) ALLOW_VARIABLES = frozenset(( NAME, ARGUMENT, TESTCASE_NAME, KEYWORD_NAME )) __slots__ = ['type', 'value', 'lineno', 'col_offset', 'error', '_add_eos_before', '_add_eos_after'] def __init__(self, type: 'str|None' = None, value: 'str|None' = None, lineno: int = -1, col_offset: int = -1, error: 'str|None' = None): self.type = type if value is None: value = { Token.IF: 'IF', Token.INLINE_IF: 'IF', Token.ELSE_IF: 'ELSE IF', Token.ELSE: 'ELSE', Token.FOR: 'FOR', Token.WHILE: 'WHILE', Token.TRY: 'TRY', Token.EXCEPT: 'EXCEPT', Token.FINALLY: 'FINALLY', Token.END: 'END', Token.CONTINUE: 'CONTINUE', Token.BREAK: 'BREAK', Token.RETURN_STATEMENT: 'RETURN', Token.CONTINUATION: '...', Token.EOL: '\n', Token.WITH_NAME: 'WITH NAME', Token.AS: 'AS' }.get(type, '') # type: ignore self.value = cast(str, value) self.lineno = lineno self.col_offset = col_offset self.error = error # Used internally be lexer to indicate that EOS is needed before/after. self._add_eos_before = False self._add_eos_after = False @property def end_col_offset(self) -> int: if self.col_offset == -1: return -1 return self.col_offset + len(self.value)
[docs] def set_error(self, error: str): self.type = Token.ERROR self.error = error
[docs] def tokenize_variables(self) -> 'Iterator[Token]': """Tokenizes possible variables in token value. Yields the token itself if the token does not allow variables (see :attr:`Token.ALLOW_VARIABLES`) or its value does not contain variables. Otherwise, yields variable tokens as well as tokens before, after, or between variables so that they have the same type as the original token. """ if self.type not in Token.ALLOW_VARIABLES: return self._tokenize_no_variables() variables = VariableIterator(self.value) if not variables: return self._tokenize_no_variables() return self._tokenize_variables(variables)
def _tokenize_no_variables(self) -> 'Iterator[Token]': yield self def _tokenize_variables(self, variables) -> 'Iterator[Token]': lineno = self.lineno col_offset = self.col_offset remaining = '' for before, variable, remaining in variables: if before: yield Token(self.type, before, lineno, col_offset) col_offset += len(before) yield Token(Token.VARIABLE, variable, lineno, col_offset) col_offset += len(variable) if remaining: yield Token(self.type, remaining, lineno, col_offset) def __str__(self) -> str: return self.value def __repr__(self) -> str: typ = self.type.replace(' ', '_') if self.type else 'None' error = '' if not self.error else f', {self.error!r}' return f'Token({typ}, {self.value!r}, {self.lineno}, {self.col_offset}{error})' def __eq__(self, other) -> bool: return (isinstance(other, Token) and self.type == other.type and self.value == other.value and self.lineno == other.lineno and self.col_offset == other.col_offset and self.error == other.error)
[docs]class EOS(Token): """Token representing end of a statement.""" __slots__ = [] def __init__(self, lineno: int = -1, col_offset: int = -1): super().__init__(Token.EOS, '', lineno, col_offset)
[docs] @classmethod def from_token(cls, token: Token, before: bool = False) -> 'EOS': col_offset = token.col_offset if before else token.end_col_offset return cls(token.lineno, col_offset)
[docs]class END(Token): """Token representing END token used to signify block ending. Virtual END tokens have '' as their value, with "real" END tokens the value is 'END'. """ __slots__ = [] def __init__(self, lineno: int = -1, col_offset: int = -1, virtual: bool = False): value = 'END' if not virtual else '' super().__init__(Token.END, value, lineno, col_offset)
[docs] @classmethod def from_token(cls, token: Token, virtual: bool = False) -> 'END': return cls(token.lineno, token.end_col_offset, virtual)