Tokenpar
Tokenpar is a framework and data model for attaching metadata to tokens within text processing workflows. The term is used in discussions of tokenization, linguistic annotation, and reproducible NLP experiments to denote the parameterization of tokens with features such as lemmas, part-of-speech tags, morphological attributes, and user-defined annotations.
Core features include a standardized token representation, a configuration interface to define fields, and adapters that
In typical workflows, a tokenizer produces tokens; tokenpar enriches them with attributes according to a schema;
History and reception notes describe tokenpar as arising from efforts to standardize token metadata across NLP