Skip to content
A lemmatizer for German language text
Python Makefile
Branch: master
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Type Name Latest commit message Commit time
Failed to load latest commit information.
data added gitignore May 19, 2017
eval_texts new evaluation with sample data Jan 30, 2019
germalemma moved germalemma to package with data folder Dec 11, 2019
Makefile moved germalemma to package with data folder Dec 11, 2019
requirements.txt use PatternLite instead Pattern Dec 6, 2019
requirements_dev.txt use PatternLite instead Pattern Dec 6, 2019
tox.ini use PatternLite instead Pattern Dec 6, 2019


December 2019, Markus Konrad / Berlin Social Science Center

A lemmatizer for German language text

Germalemma lemmatizes Part-of-Speech-tagged German language words. To do so, it combines a large lemma dictionary (an excerpt of the TIGER corpus from the University of Stuttgart), functions from the CLiPS "Pattern" package, and an algorithm to split composita.


Easy option: Installing from PyPI via pip

You can install the package from PyPI via pip:

pip install -U germalemma

Downloading and installing from source

In order to use GermaLemma, you will need to install some additional packages (see Requirements section below) and then download the TIGER corpus from the University of Stuttgart. You will need to use the CONLL09 format, not the XML format. The corpus is free to use for non-commercial purposes (see License Agreement).

Then, you should convert the corpus into pickle format for faster loading by executing and passing the path to the corpus file in CONLL09 format:

python tiger_release_[...].conll09

This will place a lemmata.pickle file in the data directory which is then automatically loaded.

Part-of-Speech (POS) Tagging

You will need to apply Part-of-Speech (POS) tagging to your text before you can lemmatize its words. See this blog post on how to do that.


You have set up GermaLemma to use the TIGER corpus (as explained above). You have tokenized your text (e.g. with NLTK). You have POS-tagged your tokens. Now you can use GermaLemma:

from germalemma import GermaLemma

lemmatizer = GermaLemma()

# passing the word and the POS tag ("N" for noun)
lemma = lemmatizer.find_lemma('Feinstaubbelastungen', 'N')
# -> lemma is "Feinstaubbelastung"

Valid POS tags

You can pass POS tags from the STTS tagset, however, only four POS tags can be processed:

  • 'N...' (nouns)
  • 'V...' (verbs)
  • 'ADJ...' (adjectives)
  • 'ADV...' (adverbs)

All other POS tags will result in a ValueError so you should wrap the call to find_lemma in a try-except block.


GermaLemma's accuracy was evaluated using a sample of 696 POS tagged and manually lemmatized words from a sample of paragraphs from proceedings of the European Parliament, Goethe's "Werther", Kafka's "Verwandlung" and a news article from the website of the WZB (see samples in folder "eval_texts").

Under the assumption that the POS tag is correct (only those words were selected), GermaLemma finds the correct lemma in 99.43% of the cases. For comparison, Pattern achieved 95.11% for the same sample.


  • Python 3.6 or newer
  • required package Pyphen
  • optional package PatternLite (This package is optional but highly recommended as it boosts the lemmatizer's accuracy.)


Apache License 2.0. See LICENSE file.

The TIGER corpus is not part of this repository and has to be downloaded separately under separate license conditions.

You can’t perform that action at this time.