De-Tokenize is a Python package that offers efficient and precise reconstruction of tokens into their original sentence structure.
Littin Rajan
- Effortless de-tokenization process
- Lightweight package architecture
- Eliminates external dependencies for hassle-free usage
- Empowers users with robust and adaptable functionality
The source code is currently hosted on GitHub at: https://github.com/littinrajan/detokenize
Binary installers for the latest released version are available at the Python Package Index (PyPI) and on Conda.
# conda
conda install detokenize
# or PyPI
pip install detokenize
from detokenize.detokenizer import detokenize
sample_tokens = ['These', 'are', 'some', 'tokens', '.']
sentence = detokenize(sample_tokens)
Warmly welcome all contributions, including bug reports, bug fixes, documentation enhancements, improvements, new features, and innovative ideas.