Skip to content

De-Tokenize is a Python package which provides fast, accurate structuring of tokens back to original sentence form

License

Notifications You must be signed in to change notification settings

littinrajan/detokenize

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

De-Tokenize


Package PyPI version Author License made-with-python Python 3.7 Downloads

What is it?

De-Tokenize is a Python package that offers efficient and precise reconstruction of tokens into their original sentence structure.

Contributor

Littin Rajan

Main Features

  • Effortless de-tokenization process
  • Lightweight package architecture
  • Eliminates external dependencies for hassle-free usage
  • Empowers users with robust and adaptable functionality

Where to get it?

The source code is currently hosted on GitHub at: https://github.com/littinrajan/detokenize

Binary installers for the latest released version are available at the Python Package Index (PyPI) and on Conda.

# conda
conda install detokenize
# or PyPI
pip install detokenize

Usage

from detokenize.detokenizer import detokenize

sample_tokens = ['These', 'are', 'some', 'tokens', '.']
sentence = detokenize(sample_tokens)

License

MIT

Contributing to De-Tokenize

Warmly welcome all contributions, including bug reports, bug fixes, documentation enhancements, improvements, new features, and innovative ideas.

About

De-Tokenize is a Python package which provides fast, accurate structuring of tokens back to original sentence form

Topics

Resources

License

Stars

Watchers

Forks

Packages

 
 
 

Languages