kbnf
This crate provides a constrained decoding engine which ensures that a language model's output adheres strictly to the format defined by KBNF (Koishi's BNF), an enhanced variant of EBNF. KBNF includes features that enhance usability, notably embeddable regular expressions.
If you are interested in the design and implementation behind this crate, you may want to check out my blog.
Features
- Supports full context free grammar with worst case O(m*n^3) time complexity, where
n
is the generated text length and m
is the vocabulary size. - Asymptotically fastest for subclasses of context free grammar.
- Guarantees worst case O(m*n) time complexity for every LR(k) grammar(which includes almost all practical grammars)
- Achieves O(n) time complexity with caching eventually given that
n
has a fixed upper bound, or the grammar is regular.
- Vocabulary-independent.
- BPE, BBPE, you-name-it, all types of vocabulary are supported.
- Supports UTF-8 characters in grammar.
- Embeddable regular expressions.
Documentation
Documentation and examples.
Add to your project
Simply add it to your Cargo.toml
or run cargo add kbnf
in your command line.
Performance
One of the goals of this crate is for the constrained decoding engine to be "fast." This can be interpreted both theoretically and practically.
Theoretically, this crate is designed to provide the asymptotically fastest algorithms for each subclass of context free grammar. By implementing an Earley recognizer with Leo optimization, this crate has successfully achieve linear time complexity for every LR(k) grammar and quadratic time complexity for every unambiguous grammar. For general context free grammar, things are more ambiguous(pun intended): while subcubic algorithms exist(although with a large constant), all other general-purpose parsing algorithms(like Earley, GLR, GLL...) are indeed cubic, like ours.
Practically, this crate tries to make the engine be as efficient as possible for grammars used in practice. While many improvements, such as Earley sets compaction and lazy caching, have been made, this is inherently an ongoing process. If you find the engine is a bottleneck in your application, feel free to open an issue.