Skip to content

simonepri/text-tokenizers-colab

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 

Repository files navigation

text-tokenizers-colab

🔪 Tokenize text on the fly on Colab.

Synopsis

Tokenization is the task of splitting a text into meaningful segments, called tokens. This repository contains python notebooks to run some text tokenizers for quick experimentation purposes. Just click on one of the links in the list below and run the notebook.

Do you believe that this is useful? Has it saved you time? Or maybe you simply like it?
If so, support this work with a Star ⭐️.

Notebooks

  • Hugging Face's Transformers Library Tokenizers - Open in Colab
  • Explosion AI spaCy Library Tokenizers - Open in Colab

Authors

See also the list of contributors who participated in this project.

License

This project is licensed under the MIT License - see the license file for details.

Releases

No releases published

Packages

No packages published