A python and rust implementation of SentencePiece (A language-independent subword tokeniser and de-tokeniser developed by Google)
-
Updated
Mar 7, 2025 - Rust
A python and rust implementation of SentencePiece (A language-independent subword tokeniser and de-tokeniser developed by Google)
Add a description, image, and links to the tokeizer topic page so that developers can more easily learn about it.
To associate your repository with the tokeizer topic, visit your repo's landing page and select "manage topics."