The lexer, or lexical analyzer, defines how a file's contents are broken into tokens.
The lexer serves as a foundation for nearly all of the features of custom language plugins, from basic syntax highlighting to advanced code analysis features.
The API for the lexer is defined by the
The IDE invokes the lexer in three main contexts, and the plugin can provide different lexer implementations for these contexts:
Syntax highlighting: The lexer is returned from the implementation of the
SyntaxHighlighterFactoryinterface which is registered in the
Building the syntax tree of a file: the lexer is expected to be returned from
ParserDefinition.createLexer(), and the
ParserDefinitioninterface is registered in the
Building the index of the words contained in the file: if the lexer-based words scanner implementation is used, the lexer is passed to the
The lexer used for syntax highlighting can be invoked incrementally to process only the file's changed part. In contrast, lexers used in other contexts are always called to process an entire file or a complete language construction embedded in a different language file.
A lexer that can be used incrementally may need to return its state, which means the context corresponding to each position in a file.
For example, a Java lexer could have separate states for top-level context, comment context, and string literal context.
An essential requirement for a syntax highlighting lexer is that its state must be represented by a single integer number returned from
That state will be passed to the
Lexer.start() method, along with the start offset of the fragment to process, when lexing is resumed from the middle of a file.
Lexers used in other contexts can always return
0 from the
The easiest way to create a lexer for a custom language plugin is to use JFlex.
FlexAdapter adapt JFlex lexers to the Consulo Lexer API.
We have a patched version of JFlex that can be used with the lexer skeleton file located at tools/lexer/idea-flex.skeleton in the IntelliJ IDEA Community Edition source to create lexers compatible with
The patched version of JFlex provides a new command-line option
--charat that changes the JFlex generated code to work with the Consulo skeleton.
--charat option passes the source data for lexing as a
CharSequence and not as an array of characters.
For developing lexers using JFlex, the GrammarKit plugin can be useful. It provides syntax highlighting and other useful features for editing JFlex files.
NOTE Lexers, and in particular JFlex-based lexers, need to be created so that they always match the entire contents of the file, without any gaps between tokens, and generate special tokens for characters which are not valid at their location. Lexers must never abort prematurely because of an invalid character.
Types of tokens for lexers are defined by instances of
Many token types common for all languages are defined in the
Custom language plugins should reuse these token types wherever applicable.
For all other token types, the plugin needs to create new
IElementType instances and associate with the language in which the token type is used.
IElementType instance should be returned every time a particular token type is encountered by the lexer.
An important feature that can be implemented at the lexer level is mixing languages within a file, such as embedding fragments of Java code in some template language.
Suppose a language supports embedding its fragments in another language.
In that case, it needs to define the chameleon token types for different types of fragments that can be embedded, and these token types need to implement the
The enclosing language's lexer needs to return the entire fragment of the embedded language as a single chameleon token, of the type defined by the embedded language.
To parse the contents of the chameleon token, the IDE will call the parser of the embedded language through a call to