Skip to content

de_pytt_bertbasecased_lg-2.1.0

Compare
Choose a tag to compare
@ines ines released this 02 Aug 19:44
· 1480 commits to master since this release

Downloads

Details: https://spacy.io/models/de#de_pytt_bertbasecased_lg

File checksum: 5c5f585bbe23f8c1d565a52cec3400467892cbeb8e1a62ba45335d8c209b275c

Provides weights and configuration for the pretrained transformer model bert-base-german-cased, published by deepset. The package uses HuggingFace's pytorch-transformers implementation of the model. Pretrained transformer models assign detailed contextual word representations, using knowledge drawn from a large corpus of unlabelled text. You can use the contextual word representations as features in a variety of pipeline components that can be trained on your own data.

Requires the spacy-pytorch-transformers package to be installed. A CUDA-compatible GPU is advised for reasonable performance.

Feature Description
Name de_pytt_bertbasecased_lg
Version 2.1.0
spaCy >=2.1.7
Model size 406 MB
Pipeline sentencizer, pytt_wordpiecer, pytt_tok2vec
Sources bert-base-german-cased
License MIT
Author deepset (repackaged by Explosion)

Installation

pip install spacy
spacy download de_pytt_bertbasecased_lg