C++ single header library which takes a language description as a C++ code and turns it into a LR1 table parser with a deterministic finite automaton lexical analyzer, all in compile time. What's more, the generated parser is actually itself capable of parsing in compile time. All it needs is a C++17 compiler!
- Installation
- Compiler support
- Usage
- Explanation
- Compile Time Parsing
- LR(1) Parser
- Functors - advanced
- External context
- Various features
- Regular expressions
- Diagnostics
- Resulting binary size
It is a single header library. You can just copy the include/ctpg/ctpg.hpp header wherever you want.
Use CMake 3.14+ to build the library:
$ git clone https://github.com/peter-winter/ctpg
$ cmake -S ctpg -B ctpg/build -DCMAKE_BUILD_TYPE=Release
$ cmake --build ctpg/build
$ (cd ctpg/build ; ctest) # optional
$ cmake --install ctpg/build [--prefix /usr/local]
CTPG can be used via the standard find_package
interface. Just link to ctpg::ctpg
!
There is a minimal, self-contained example in examples/language/CMakeLists.txt.
The ctpg-git
AUR package is available for Arch-based distribution users.
- Manual install:
git clone https://aur.archlinux.org/ctpg-git.git
cd ctpg-git
makepkg -si
- Using yay:
yay -S ctpg-git
Tested on:
- GCC 10.3+, 11.x, 12.x
- Clang 12.x, 13.x, 14.x
- MSVC 19.30+
Following code demonstrates a simple parser which takes a comma separated list of integer numbers as argument and prints a sum of them.
readme-example.cpp
#include <ctpg/ctpg.hpp>
#include <iostream>
#include <charconv>
using namespace ctpg;
using namespace ctpg::buffers;
constexpr nterm<int> list("list");
constexpr char number_pattern[] = "[1-9][0-9]*";
constexpr regex_term<number_pattern> number("number");
int to_int(std::string_view sv)
{
int i = 0;
std::from_chars(sv.data(), sv.data() + sv.size(), i);
return i;
}
constexpr parser p(
list,
terms(',', number),
nterms(list),
rules(
list(number) >=
to_int,
list(list, ',', number)
>= [](int sum, char, const auto& n){ return sum + to_int(n); }
)
);
int main(int argc, char* argv[])
{
if (argc < 2)
return -1;
auto res = p.parse(string_buffer(argv[1]), std::cerr);
bool success = res.has_value();
if (success)
std::cout << res.value() << std::endl;
return success ? 0 : -1;
}
Compile and run:
g++ readme-example.cpp -std=c++17 -o example && example "10, 20, 30"
You should see the output : 60. If incorrect text supplied as an argument:
g++ readme-example.cpp -std=c++17 -o example && example "1, 2, 3x"
you should see:
[1:8] PARSE: Unexpected character: x
#include <ctpg/ctpg.hpp>
Namespace ctpg is the top namespace. There are couple of feature namespaces like buffers
using namespace ctpg;
using namespace ctpg::buffers;
Terminal symbols (short: terms) are symbols used in grammar definition that are atomic blocks. Examples of the terms from a C++ language are: identifier, '+' operator, various keywords etc.
To define a term use the one of char_term
, string_term
and regex_term
classes.
Here is the example of a regex_term with a common integer number regex pattern.
constexpr char number_pattern[] = "[1-9][0-9]*";
constexpr regex_term<number_pattern> number("number");
The constructor argument ("number")
indicates a debug name and can be omitted, however it is not advised.
Names are handy to diagnose problems with the grammar. If omitted, the name will be set to the pattern string.
Note: the pattern needs to have a static linkage to be allowed as a template parameter. This is C++17 limitation, and CTPG does not support C++20 features yet.
Other types of terms
char_term
is used when we need to match things like a +
or ,
operator.
string_term
is used when we need to match a whole string, like a language keyword.
Nonterminal symbols (short: nonterms) are essentially all non atomic symbols in the grammar. In C++ language these are things like: expression, class definition, function declaration etc.
To define a nonterm use the nterm
class.
constexpr nterm<int> list("list");
The constructor argument ("list") is a debug name as well, like in the case of regex_term. The difference is in nterms names are neccessary, because they serve as unique identifiers as well. Therefore it is a requirement that nonterm names are unique.
Template parameter <int>
in this case is a value type. More on this concept later.
The parser
class together with its template deduction guides allows to define parsers using 4 arguments:
- Grammar root - symbol which is a top level nonterm for a grammar.
- List of all terms
- List of all nonterms
- List of rules
The parser
object should be declared as constexpr
, which makes all the neccessary calculations of the LR(1) table parser done in compile time.
Let's break down the arguments:
constexpr parser p(
list,
terms(',', number),
nterms(list),
Grammar root.
When the root symbol gets matched (in this case list
) the parse is successful.
Term list.
List of terms enclosed in a terms
call. In our case there are two: number
and a ,
.
Note: the
,
term is not defined earlier in the code. It is an implicitchar_term
. The code implicitly converts the char to thechar_term
class. Thereforechar_terms
(as well asstring_terms
) are allowed not to be defined in advance. Their debug names are assigned to the them by default to a char (or a string) they represent.
Nonterm list.
List of terms enclosed in a nterms
call. In our case, just a single list
nonterm is enough.
Rules
List of rules enclosed in a rules
call.
Each rule is in the form of:
nonterm(symbols...) >= functor
The nonterm
part is what's called a left side of the rule. The symbols are called the right side.
The right side can contain any number of nterm
objects as well as terms (regex_terms
, char_terms
or string_terms
).
Terms can be in their implicit form, like ,
in the example. Implicit string_terms
are in form of "strings".
rules(
list(number)
>= to_int
list(list, ',', number)
>= [](int sum, char, const auto& n)
{ return sum + to_int(n); }
)
The first rule list(number)
indicates that the list
nonterm can be parsed using a single number
regex term.
The second rule uses what's know as a left recurrence. In other words, a list
can be parsed as a list
followed by a ,
and a number
.
Functors
The functors are any callables that can accept the exact number of arguments as there are symbols on the right side and return a value type of the left side. Each nth argument needs to accept a value of a value type of the nth right side symbol.
So in the case of the first to_int
functor, it is required to accept a value type of regex_term
and return an int
.
The second functor is a lambda which accepts 3 arguments: an int
for the list
, a char
for the ,
and auto for whatever is passed as
a value type for the regex_term
.
Note: Functors are called in a way that allows taking advantage of move semantics, so defining it's arguments as a move reference is encouraged.
Value types for terms
Terms unlike nonterms (which have their value types defined as a template parameter to the nterm definition),
have their value types predefined to either a term_value<char>
for a char_term
, and a term_value<std::string_view>
for both regex_term
and string_term
.
The term_value
class template is a simple wrapper that is implicitly convertible to it's template parameter (either a char
or std::string_view
).
That's why when providing functors we can simply declare arguments as either a char
or a std::string_view
.
In our case the to_int
functor has a std::string_view
argument, which accepts a term_value<std::string_view>
just fine.
Of course an auto
in case of lambda will always do the trick.
The advantage of declaring functor arguments as term_value
specialization is that we can access other features (like source tracking) using the term_value
methods.
Use parse
method with 2 argumets:
- a buffer
- an error stream
Buffers
Use a string_buffer from a buffers
namespace to parse a null terminated string or a std::string
.
Error stream
Stream reference like std::cerr
or any other std::ostream
can be pased as a stream argument.
This is the place where the parse
method is going to spit out error messages like a syntax error.
auto res = p.parse(string_buffer(argv[1]), std::cerr);
Parse return value
The parse
method returns an std::optional<T>
, where T
is a value type of the root symbol.
Use the .has_value()
and the .value()
to check and access the result of the parse.
Note: White space characters are skipped by default between consequent terms.
Example code can be easily changed to create an actual constexpr parser.
First, all the functors need to be constexpr.
To achieve this change the to_int
function to:
constexpr int to_int(std::string_view sv)
{
int sum = 0;
for (auto c : sv) { sum *= 10; sum += c - '0'; }
return sum;
}
The function is now constexpr. The <charconv>
header is now unneccessary.
Note: To allow constexpr parsing all of the nonterm value types have to be literal types.
Also change the main to use cstring_buffer
and declare a parse result constexpr.
The error stream argument is also unavailable in constexpr parsing.
int main(int argc, char* argv[])
{
if (argc < 2)
{
constexpr char example_text[] = "1, 20, 3";
constexpr auto cres = p.parse(cstring_buffer(example_text)); // notice cstring_buffer and no std::err output
std::cout << cres.value() << std::endl;
return 0;
}
auto res = p.parse(string_buffer(argv[1]), std::cerr);
bool success = res.has_value();
if (success)
std::cout << res.value() << std::endl;
return success ? 0 : -1;
}
Now when no argument is passed to the program, it prints the compile time result of parsing "1, 20, 3".
g++ readme-example.cpp -std=c++17 -o example && example
should print the number 24.
If the example_text
variable was an invalid input, the code cres.value()
would throw, because the cres
is of type std::optional<int>
with no value.
Changing the parse
call to:
constexpr int cres = p.parse(cstring_buffer(example_text)).value();
would cause compilation error, because throwing std::bad_optional_access
is not constexpr.
CTPG uses a LR(1) parser. This is short from left-to-right and 1 lookahead symbol.
The parser uses a parse table which is somewhat resembling a state machine. Here is pseudo code for the algorithm:
struct entry
int next // valid if shift
int rule_length // valid if reduce
int nterm_nr // valid if reduce
enum kind {success, shift, reduce, error }
bool parse(input, sr_table[states_count][terms_count], goto_table[states_count][nterms_count])
state = 0
states.push(state)
needs_term = true;
while (true)
if (needs_term)
term_nr = get_next_term(input)
entry = sr_table[state, term_nr]
kind = entry.kind
if (kind == success)
return true
else if (kind == shift)
needs_term = true;
state = entry.next
states.push(state)
continue
else if (kind == reduce)
states.pop_n(entry.rule_length)
state = states.top()
state = goto_table[state, entry.nterm_nr]
continue
else
return false
Parser contains a state stack, which grows when the algorithm encounters a shift operation and shrinks on reduce operation.
Aside from a state stack, there is also a value stack for dedicated for parse result calculation. Each shift pushes a value to the stack and each reduce calls an appropriate functor with values from a value stack, removing values from a stack and replacing them with a single value associated with a rule's left side.
Table creation
This topic is out of scope of this manual. There is plenty of material online on LR parsers. Recomended book on the topic: Compilers: Principles, Techniques and Tools
There are situations (parser states) in which when a particualr term is encountered on the input, there is an ambiguity regarding the operation a parser should perform.
In other words a language grammar may be defined in such a way, that both shift and reduce can lead to a successfull parse result, however the result will be different in both cases.
Example 1
Consider a classic expression parser (functors omitted for clarity):
constexpr parser p(
expr,
terms('+', '*', number),
nterms(expr),
rules(
expr(number),
expr(expr, '+', expr),
expr(expr, '*', expr)
)
);
Consider 2 + 2 * 2
input being parsed and a parser in a state after successfully matching 2 + 2
and encountering *
term.
Both shifting a *
term and reducing by the rule expr(expr, '+', expr) would be valid, however would produce different results.
This is a classic operator precedence case, and this conflict needs to be resolved somehow. This is where precedence and associativity take place.
CTPG parsers can resolve such conflict based on precedence and associativity rules defined in a grammar.
Example above can be fixed by explicit term definitions.
Normally, char_terms
can be introduced by implicit definition in the terms
call. However when in need to define a precedence, explicit definition is required.
Simply change the code to:
constexpr char_term o_plus('+', 1); // precedence set to 1
constexpr char_term o_mul('*', 2); // precedence set to 2
constexpr parser p(
expr,
terms(o_plus, o_mul, number),
nterms(expr),
rules(
expr(number),
expr(expr, '+', expr), // note: no need for o_plus and o_mul in the rules, however possible
expr(expr, '*', expr)
)
);
The higher the precedence value set, the higher the term precedence. Default term precedence is equal to 0.
This explicit precedence definition allows a *
operator to have bigger precedence over +
.
Example 2
constexpr char_term o_plus('+', 1); // precedence set to 1
constexpr char_term o_minus('-', 1); // precedence set to 1
constexpr char_term o_mul('*', 2); // precedence set to 2
constexpr parser p(
expr,
terms(o_plus, o_minus, o_mul, number),
nterms(expr),
rules(
expr(number),
expr(expr, '+', expr),
expr(expr, '-', expr), // extra rule allowing binary -
expr(expr, '*', expr),
expr('-', expr) // extra rule allowing unary -
)
);
Binary -
and +
operators have the same precedence in pretty much all languages.
Unary -
however almost always have a bigger precedence than all binary operators.
We can't achieve this by simply defining -
precedence in char_term
definition.
We need a way to tell that expr('-', expr)
has a bigger precedence then all binary rules.
To achieve this override the precedence in a term by a precedence in a rule changing:
expr('-', expr)
to
expr('-', expr)[3]
The []
operator allows exactly this. It explicitly sets the rule precedence so the parser does not have to deduce rule precedence from a term.
So the final code looks like this:
constexpr char_term o_plus('+', 1); // precedence set to 1
constexpr char_term o_minus('-', 1); // precedence set to 1
constexpr char_term o_mul('*', 2); // precedence set to 2
constexpr parser p(
expr,
terms(o_plus, o_minus, o_mul, number),
nterms(expr),
rules(
expr(number),
expr(expr, '+', expr),
expr(expr, '-', expr), // extra rule allowing binary -
expr(expr, '*', expr),
expr('-', expr)[3] // extra rule allowing unary -, with biggest precedence
)
);
Example 3
Consider the final code and let's say the input is 2 + 2 + 2
, parser has read 2 + 2
and is about to read the second +
.
In this case what is the required behaviour? Should the first 2 + 2
be reduced or a second +
should be shifted?
(This may not matter in case of integer calculations, but may have a big difference in situations like expression type deduction in c++ when operator overloading is involved.)
This is the classic associativity case which can be solved by expicitly defining the term associativity.
There are 3 types of associativity available: left to right, right to left and not associative as the default.
To explicitly define a term associativity change the term definitions to:
constexpr char_term o_plus('+', 1, associativity::ltor);
constexpr char_term o_minus('-', 1, associativity::ltor);
constexpr char_term o_mul('*', 2, associativity::ltor);
Now all of these operators are left associative, meaning the reduce will be preferred over shift.
Should the associativity be defined as associativity::rtol
, shift would be preferred.
No associativity prefers shift by default.
Precedence and associativity summary
When a shift reduce conflict is encountered these rules apply in order:
Let r be a rule which is a subject to reduce and t be a term that is encountered on input.
- when explicit r precedence from
[]
operator is bigger than t precedence, perform a reduce - when precedence of last term in r is bigger than t precedence, perform a reduce
- when precedence of last term in r is equal to t precedence and last term in r is left associative, perform a reduce
- otherwise, perform a shift.
Reduce - reduce conflicts
In some cases the language is ill formed and the parser contains a state in which there is an ambiguity between several reduce actions.
Consider example:
constexpr nterm<char> op("op");
constexpr nterm<char> special_op("op");
constexpr parser p(
op,
terms('!', '*', '+'),
nterms(special_op, op),
rules(
special_op('!'),
op('!'),
op('*'),
op('+'),
op(special_op)
)
);
Let's say we parse an input !
. The parser has no way of telling if it should reduce using rule special_op('!')
or op('!')
.
This is an example of reduce/reduce conflict and such parser behaviour should be considered undefined.
There is a diagnostic tool included in CTPG which detects such conflicts so they can be addressed.
Consider a parser matching white space separated names (strings).
constexpr char pattern[] = "[a-zA-Z0-9_]+";
constexpr regex_term<pattern> name("name");
using name_type = std::string_view;
using list_type = std::vector<name_type>;
constexpr nterm<list_type> list("list");
constexpr parser p(
list,
terms(name),
nterms(list),
rules(
list(),
list(list, name)
)
);
How exactly would the functors look for this kind of parser?
The first rule list()
is an example of an empty rule. This means the list can be reduced from no input.
Since the rule's left side is a list
the functor needs to return its value type, which is a list_type
.
The right side is empty so the functor needs to have no arguments.
So let's return an empty vector: [](){ return list_type{}; }
The second rule reduces a list from a name and a list, therefore the functor needs to accept:
list_type
for the first argument: listterm_value<std::string_view>
for the second argument: name- return a
list_type
So let's create a functor:
[](auto&& list, auto&& name){ list.emplace_back(std::move(name)); return list; }
The name
argument will resolve to term_value<std::string_view>&&
, which is convertible to std::string_view&&
.
Now the parser looks like this:
constexpr char pattern[] = "[a-zA-Z0-9_]+";
constexpr regex_term<pattern> name("name");
using name_type = std::string_view;
using list_type = std::vector<name_type>;
constexpr nterm<list_type> list("list");
constexpr parser p(
list,
terms(name),
nterms(list),
rules(
list()
>= [](){ return list_type{}; },
list(list, name)
>= [](auto&& list, auto&& name){ list.push_back(name); return std::move(list); }
)
);
Note: Here we take advantage of move semantics which are supported in the functor calls. This way we are working with the same
std::vector
instance we created as empty using the first rule.
Important Note It is possible for functors to have referrence (both const and not) argument types, however lifetime of the objects passed to functors ends immediately after the functor returns. So it is better to avoid using referrence types as nterm value types.
There are a couple of handy ready to use functor templates:
val
Use when a functor needs to return a value which doesn't depend on left side:
using namespace ctpg::ftors;
constexpr nterm<bool> binary("binary");
constexpr parser p(
binary,
terms('0', '1', '&', '|'),
nterms(binary),
rules(
binary('0')
>= val(false),
binary('1')
>= val(true),
binary(binary, '&', binary)
>= [](bool b1, auto, bool b2){ return b1 & b2; },
binary(binary, '|', binary)
>= [](bool b1, auto, bool b2){ return b1 | b2; },
)
);
create
Use when a functor needs to return a default value of a given type:
// word list parser from one of previous examples
using namespace ctpg::ftors;
constexpr parser p(
list,
terms(name),
nterms(list),
rules(
list()
>= create<list_type>{}, // use instead of a lambda
list(list, name)
>= [](auto&& list, auto&& name){ list.push_back(name); return std::move(list); }
)
);
construct
Use to construct an instance of a given type from an argument:
using namespace ctpg::ftors;
constexpr parser p(
something,
terms(value),
nterms(something),
rules(
something(value)
>= construct<something_type, 1>{} // 1 is the position of value in the rule, this constructs something_type{value}
)
)
element placeholders
Use whenever a rule simply passes nth element from the right side:
using namespace ctpg::ftors;
constexpr char pattern[] = "[1-9][0-9]*";
constexpr regex_term<pattern> number("number");
constexpr to_int(std::string_view x){ /*implement*/ }
constexpr nterm<int> expr("expr");
constexpr parser p(
expr,
terms('+', '(', ')', number),
nterms(expr),
rules(
expr(number)
>= to_int,
expr(expr, '+', expr)
>= [](int i1, auto, int i2){ return i1 + i2; },
expr('(', expr, ')')
>= _e2 // here, just return the second element
)
);
list helpers
Use push_back
or emplace_back
when dealing with common list tasks.
The push_back
calls push_back
on first element passing second element as argument:
list(list, element) = push_back{}
The emplace
back works similarly but supports move semantics.
// word list parser from one of previous examples
using namespace ctpg::ftors;
constexpr parser p(
list,
terms(name),
nterms(list),
rules(
list()
>= create<list_type>{},
list(list, name)
>= push_back{}
)
);
Both push_back
an emplace_back
are class templates, which take two indexes (like in case of element placeholders one-indexed),
denoting element numbers for the list and the value to append.
So in case of comma separeted numbers, we can simply use:
rules(
list(number) >= construct<list_type, 1>{},
list(list, ',', number) >= push_back<1, 3>{} // 1 is the list, 3 is the number
)
There is a situation where the functor can be entirely omitted, that is whenever a left side value type is move constructible from right side value types:
// Example parser, accepts url addresses in for of a protocol and a list of words, like: https://www.example.com
constexpr char word_pattern[] = "[0-9A-Za-z]+";
constexpr regex_term<word_pattern> word("word");
constexpr char protocol_pattern[] = "http://|https://";
constexpr regex_term<protocol_pattern> protocol("protocol");
using list_type = std::vector<std:string_view>;
struct url_type
{
constexpr url_type(std::string_view pr, list_type&& l):
pr(pr), l(std::move(l))
{}
std::string_view pr;
list_type l;
};
constexpr nterm<url_type> url;
constexpr nterm<list_type> list;
constexpr parser p(
url,
terms(word, '.', protocol),
nterms(url, list),
rules(
list(word)
>= [](auto w){ return list_type{w}; },
list(list, '.', word)
>= [](auto&& l, auto, auto w){ l.push_back(w); return std::move(l); },
url(protocol, list)
// skip functor entirely, url_type move constructible from right side value types
)
);
Sometimes it is useful to access some part of program's dynamic data during parsing.
For example, if one wants to parse and evaluate expression a + b
knowing that a = 1
and b = 2
then one of the most intuitive ideas is to pass this mapping of variables directly into the variable-handling functor which should take variable's name and return its value extracted from the mapping.
This idea can be implemented using operator >>=
instead of >=
and invoking context_parse
instead of parse
:
constexpr char_term o_plus('+', 1, associativity::ltor);
constexpr char variable_pattern[] = "[_a-zA-Z][_a-zA-Z0-9]*";
constexpr regex_term<variable_pattern> variable("variable");
constexpr parser p(
expr,
terms(variable, o_plus),
nterms(expr),
rules(
expr(expr, o_plus, expr) >= [](auto l, auto, auto r) { return l + r; },
expr(variable) >>= [](const auto &ctx, auto v) { return ctx.at(v); }
)
);
...
const std::unordered_map<std::string_view, int> context({{"a", 1}, {"b", 2}});
auto res = p.context_parse(context, cstring_buffer("a + b"));
>>=
indicates that the functor accepts the context object as its first argument.
And context_parse
works just like parse
and can accept all the same arguments but it also accepts the context object as its first argument and then redirects it to functors.
Note that >=
can still be used, parser just doesn't pass the context to functors defined using this operator.
Moreover, context object can be passed by non-const reference that allows functors to modify the context:
expr(...) >>= [](auto &ctx, ...) { /* modify ctx */; ... }
...
Context context;
auto res = p.context_parse(context, ...);
Contexts can be used even in compile-time parsing as long as the context object can be constructed in compile time and all referenced methods are constexpr:
constexpr parser p(
expr,
terms(variable, o_plus),
nterms(expr),
rules(
expr(expr, o_plus, expr) >= [](auto l, auto, auto r) { return l + r; },
expr(variable) >>= [](const auto &ctx, auto v)
{
if (v.get_value() == "a")
return ctx.a;
if (v.get_value() == "b")
return ctx.b;
return 0;
}
)
);
...
struct variables {
int a;
int b;
};
constexpr auto res = p.context_parse(variables{1, 2}, cstring_buffer("a + b"));
Note that in all the examples above the context type is specified after the parser is already defined. It is possible because context type substitution is performed inside context_parse
and not inside the parser object constructor. It also allows one to use one parser with different contexts as long as types of these contexts can be safely substituted into functors:
// p is the parser from the previous example.
struct variables {
int a;
int b;
};
struct more_variables {
int a;
int b;
int c;
};
constexpr auto res1 = p.context_parse(variables{1, 2}, cstring_buffer("a + b"));
constexpr auto res2 = p.context_parse(more_variables{1, 2, 3}, cstring_buffer("a + b"));
To change the parse options simply provide a parse_options
instance to the parse
call:
p.parse(parse_options{}, cstring_buffer("abc"), std::cerr);
To set a particular option use on of the set_xxx
methods:
p.parse(parse_options{}.set_verbose(), cstring_buffer("abc"), std::cerr);
Note: The
set_xxx
methods returnparse_options
instance using*this
, so they can be chained together.
The list of available parse options:
- set_skip_whitespace(bool value)
By default parser skips the whitespace characters between the terms, this can be changed using this option.
- set_skip_newline(bool value)
If set_skip_whitespace is set to true
, this options allows to not skip newline (\n
) characters. Value defaults to true.
Note: This option has no effect if skipping whitespaces is disabled.
- set_verbose(bool value)
Sets the parser to verbose mode. More on this in the Verbose output section.
To allow verbose output for debugging purposes call parse
method with such arguments:
p.parse(parse_options{}.set_verbose(), cstring_buffer("abc"), std::cerr);
The default parse_options
is appended with the set_verbose
call, thus changing the verbosity option.
The last argument can be anything convertible to std::ostream
referrence.
The verbose output stream contains alongside usual syntax errors, the detailed process of syntax and lexical analyze. The shift and reduce actions are put to the output which is useful together with the Diagnostics information. The lexical analyzer DFA actions are also printed, again useful during diagnostics.
Source tracking is a feature that makes the parser keep track of source point (that is line and column) it is currently in. This feature is always available and source point information is attached to every term value that is passed to a functor.
To use this information make the functor accept the term_value
type arguments for each term.
For char_terms the value type is term_value<char>
, for both string_term and regtex_term the value type is term_value<std::string_view
.
Each of these types have the source_point
member that can be accessed using get_sp()
method.
The source_point
struct has a line
and column
public members and can be output to a stream using <<
operator.
This is an example of a parser that accepts a whitespace separated words and stores them in a collection together with their source points.
Take a look on the functor that utilises both value and source point of a word using const auto& w
argument by calling get_value()
and get_sp()
respectively.
#include "ctpg.hpp"
#include <iostream>
using namespace ctpg;
using namespace ctpg::ftors;
using namespace ctpg::buffers;
struct word_t
{
std::string w;
source_point sp;
};
using text_t = std::vector<word_t>;
auto&& add_word(text_t&& txt, std::string_view sv, source_point sp)
{
txt.push_back(word_t{std::string(sv), sp});
return std::move(txt);
}
constexpr char word_pattern[] = "[A-Za-z]+";
constexpr regex_term<word_pattern> word("word");
constexpr nterm<text_t> text("text");
constexpr parser p(
text,
terms(word),
nterms(text),
rules(
text() >= create<text_t>{},
text(text, word) >= [](auto&& txt, const auto& w) { return add_word(std::move(txt), w.get_value(), w.get_sp()); }
)
);
int main(int argc, char* argv[])
{
if (argc < 2)
return -1;
auto res = p.parse(string_buffer(argv[1]), std::cout);
if (res.has_value())
{
for (const auto& w : res.value())
{
std::cout << w.w << " at " << w.sp << std::endl;
}
}
return 0;
}
There are currently three types of buffers available:
cstring_buffer
, useful for constexpr parsing static array-like buffers;string_buffer
andstring_view_buffer
for runtime parsing.
It is however easy to add custom types of buffers, there are just couple of requirements for the types to be eligible as buffers.
The buffer needs to expose public iterator
type which should be obtainable by begin
and end
methods and return iterators to the start and past the end of the buffer respectively.
The get_view
member should return a std::string_view
given two iterators, one at the start of the view and the other past the end.
iterator begin() const { return iterator{ data }; }
iterator end() const { return iterator{ data + N - 1 }; }
std::string_view get_view(iterator start, iterator end) const
The iterator type should expose following public member methods:
char operator *() const; // derefference to a pointed char
iterator& operator ++(); // pre and post incrementation
iterator operator ++(int);
bool operator == (const iterator& other) const; // comparison operator
It is possible to define term value types as custom types, not limited to char
or std::string_view
.
It can be achieved using typed_term
class template.
Wrap the usual term definition:
char_term plus('+');
with typed_term
like this:
// a custom type for the plus term
struct plus_tag{};
typed_term plus(char_term('+'), create<plus_tag>{});
The create
is a functor available in the ctpg::ftors
namespace, which simply creates an object of given type using a default constructor of that type
and ignoring all passed arguments to it.
In fact any callble object which accepts std::string_view
can be used instead of create
, this is just an example.
The plus
term has a value type identical to the return type of the functor, plus_tag
in this case.
Take a look at the typed-terms.cpp
in the examples, it uses this feature to create a simple calculator, but instead of the
runtime switch statement on the char value like in the simple-expr-parser.cpp
, the functor object has an overload for each arithmetic operator.
Note: Typed terms cannot use their implicit versions like the basic terms (
char_term
,string_term
) in the rules. They have to be referrenced by the typed_terms object.
If a special error term in a rule is used, the parser tries to recover from syntax error.
Consider the example from error-recovery.cpp example (here, simplified):
constexpr parser p(
exprs,
terms(number, o_plus, ';'),
nterms(exprs, expr),
rules(
exprs() >= create<exprs_type>{},
exprs(exprs, expr, ';') >= push_back<1, 2>{},
exprs(exprs, error, ';') >= _e1,
expr(expr, '+', expr) >= [](int x1, skip, int x2){ return x1 + x2; },
expr(number) >= [](const auto& sv){ return get_int(sv); }
)
);
This rule allows parser to recover from syntax error when the expression is ill formed, the _e1
functor will simply pass expressions parsed to this point:
exprs(exprs, error, ';') >= _e1,
Recovery follows the rules:
- when syntax error occurs a special <error_recovery_token> is presented to the LR algorithm.
- parser states are reverted (popped from a stack) until the state accepting the <error_recovery_token> is encountered.
- if at any point the is no more states to pop, algorithm fails.
- <error_recovery_token> is shifted, and shift action is performed.
- terminals are consumed and ignored until the terminal which would not result in a syntax error is encountered.
- if at any point end of input is encountered, the algorithm fails.
To see how error in rules affect the parse table generation take a look at the diagnostic output and look for the <error_recovery_token> occurrences. See the Diagnostics section for details.
It is possible to define a custom lexical analyzer instead of the one automatically generated from terms.
To achieve this use the 5th argument to the parser definition and pass a use_lexer
object:
constexpr nterm<int> list("list");
constexpr custom_term number("number", [](auto sv){ return int(sv[0]) - '0';} );
constexpr custom_term comma(",", create<no_type>{} );
constexpr parser p(
list,
terms(comma, number),
nterms(list),
rules(
list(number),
list(list, comma, number)
>= [](int sum, skip, int x){ return sum + x; }
),
use_lexer<int_lexer>{}
);
To use custom lexerical analyzer define custom_term
objects instead of traditional terms.
The custom_term
class acts like a typed_term
, you need to provide a functor to it's definition, from which the term
deduces it's value type.
The lexer class needs to implement a match
method:
class int_lexer
{
public:
template<typename Iterator, typename ErrorStream>
constexpr recognized_term match(
match_options options,
source_point sp,
Iterator start,
Iterator end,
ErrorStream& error_stream);
};
The recognized_term
is a simple struct with two members: term_idx
and len
.
When returning this structure from match
return term index (starting from 0 according to terms
call in a parser definition)
and the length of recognized term.
struct recognized_term
{
constexpr recognized_term() = default; // use this overload to indicate lexer error - no term matched
constexpr recognized_term(size16_t term_idx, size_t len); // use this overload to return a match
};
Take a look at the custom-lexer.cpp example for a tutorial on custom lexical analyzer implementation.
When defining a regex pattern for a regex term:
constexpr char number_pattern[] = "[1-9][0-9]*";
constexpr regex_term<number_pattern> number("number");
use the following supported features (in precedence descending order):
Feature | Example | Meaning |
---|---|---|
Single char | a |
character a |
Escaped char | \| |
character | |
Escaped char (hex) | \\x20 |
space character |
Any char | . |
any character |
Char range | [a-z] |
lower case letter |
Char set | [abc] |
a , b , or a c character |
Inverted char range | [^a-z] |
everything but lower case letter |
Inverted char set | [^!] |
everything but ! character |
Complex char set | [_a-zA-Z] |
any letter or underscore character |
Concatenation | ab |
characters a and b in order |
Repetition (zero or more) | a* |
zero or more of a character |
Repetition (one or more) | a+ |
one or more of a character |
Optional | a? |
optional a character |
Repetition (defined number) | a{4} |
4 a characters |
Alternative | a|b |
a character or b character |
Grouping | (a|b)* |
any number of a or b characters |
To diagnose broblems in the parser use the write_diag_str
method which returns a string of output with the parser state machine details:
p.write_diag_str(std::cerr)
The output contains 2 sections: one for syntax analyzer starting with the word PARSER and the other for lexical analyzer starting with LEXICAL ANALYZER.
Parser object size: <number>
Number of states: <number>(cap: <number>)
Max number of situations per state: <number>(cap: <number>)
The size of the parser object may be couple of megabytes for some complex grammars, so consider declaring the parser as a constexpr object rather than on local stack. You may also consider looking at how to reduce the executable binary size.
Next, there is a rule set description in form of:
RULES
nr nterm <- s0 s1 s2 ... s(rule_length)
Next, there is a state machine description in form of:
STATE nr
followed by description of all possible situations in which the parser is when in this state. Each of the situations refer to a single rule and are in form:
nterm <- s0 s1 s2 ... s(n) . s(n+1) ... s(rule_length) ==> lookahead_term
The nterm is the name of the left side nonterm, s0, s1 ... are right side symbols from the same rule.
The .
after the s(n) means the parser is done matching the part of the rule before the .
(all the symbols before the .
).
The lookahead_term is the term expected after the whole rule is matched. If the parser encounters the lookahead term after the rule is matched, the reduce operation is performed.
After the situations there is an action list (in order: goto actions, shift actions and reduce actions):
On <nterm> goto <state_nr>
...
On <term> shift <state_nr>
...
On <term> reduce using (<rule_nr>)
...
Goto and shift actions are basically the same, only difference is goto action refers to a nonterm and shift to a term.
They both refer to the .
in the situation, that is, given the symbol after the .
in this state, parser goes to a new state with a <state_nr>.
Reduce actions occur when the whole rule is matched, hence the reduce actions are present only when the state contains a situation with .
at the end.
What the action means is: given the reduce using rule with a <rule_nr>.
Rules are numbered according to the apearance in the source code (in the rules
call during the parser definition) starting from 0.
Shift/reduce conflicts are presented with lines:
On <term> S/R CONFLICT, prefer reduce(<rule_nr>) over shift
On <term> S/R CONFLICT, prefer shift over reduce(<rule_nr>)
Reduce/reduce conflicts look like this:
On <term> R/R CONFLICT - !!! FIX IT !!!
Section contains deterministic finite automaton which corresponds to all of the terms used in a grammar.
Each line represents a single machine state:
STATE <nr> [recognized <term>] {<char_descr> -> <new_state>} {<char_descr> -> <new_state>}...
The [recognized <term>]
part is optional and means that the DFA in this state could return the recognized term, however it is trying to match longest possible input
so it continues consuming characters. When it reaches an error state (no new state for the character) the last recognized term is returned, or an 'unexpected character' error occurs if no term recognized so far.
The <char_descr> -> <new_state>
represents the DFA transition on a character described by character description <char_descr>.
Character descriptions are in form of a single printable character, or in case of non-printable it's hex representation like : 0x20 for space character.
Character descriptions can also contain character range in form: [start-end]
.
There will be unreachable states in the form:
STATE <nr> (unreachable)
These are leftovers from the regular expression to DFA conversion, just ignore them.
When creating parser for big grammar you may notice the rather big compiled executable binary size. This is because the LR1 table creation algorithm needs to predict various size caps of different collections (max number of states, max number of situations in state). It does that by 'prefer safe over perfect' approach, so it overshoots significantly most of the time.
To address this print out the disgnostic message and see if this is not the case:
PARSER
Parser object size: 509800
Number of states: 89(cap: 2500)
Max number of situations per state: 490(cap: 2500)
If you see numbers significantly lower then the caps, this is the case of an overshoot.
There is a way to address it:
struct custom_limits
{
static const ctpg::size_t state_count_cap = 45;
static const ctpg::size_t max_sit_count_per_state_cap = 30;
};
constexpr parser p(
list,
terms(...),
nterms(...),
rules(...),
use_generated_lexer{},
custom_limits{}
);
Using the 6th argument to parser
definition provide a custom limits structure with state_count_cap
and
max_sit_count_per_state_cap
defined as static const ctpg::size_t
.
The values have to be bigger than what comes out of the diagnostic message. This way you set the lower caps, decreasing the binary size assuring the actual numbers
don't exceed the caps.