The quallmer package is an easy-to-use toolbox to quickly apply AI-assisted qualitative coding to large amounts of texts, images, pdfs, tabular data and other structured data.
Using qlm_code(), users can apply codebook-based qualitative coding
powered by large language models (LLMs) to generate structured,
interpretable outputs. The package includes built-in codebooks for
common applications and allows researchers to create custom codebooks
tailored to their specific research questions using qlm_codebook(). To
ensure quality and reliability of AI-generated coding, quallmer
provides qlm_compare() for evaluating inter-rater reliability and
qlm_validate() for assessing accuracy against gold standards. With
qlm_replicate(), researchers can systematically compare results across
different models and settings to assess sensitivity and reproducibility.
The qlm_trail() function creates complete audit trails following
Lincoln and Guba’s (1985) concept for establishing trustworthiness in
qualitative research.
The quallmer package makes AI-assisted qualitative coding accessible without requiring deep expertise in R, programming or machine learning.
- Creates custom codebooks tailored to specific research questions and data types.
- Uses
instructionsand type specifications from ellmer to define coding instructions and output structure. - Example codebook objects (e.g.,
data_codebook_sentiment) demonstrate how to use built-in codebooks for common qualitative coding tasks. - Extensible framework allows researchers to define domain-specific coding schemes.
- Applies LLM-based coding to qualitative data using a
qlm_codebook. - Works with any LLM supported by ellmer.
- Returns a
qlm_codedobject containing the coded results and metadata for reproducibility.
- Re-executes coding with optional overrides (different models, codebooks, or parameters).
- Tracks provenance chain for comparing results across different configurations.
- Enables systematic assessment of coding reliability and sensitivity to model choices.
- Compares multiple
qlm_codedobjects to assess inter-rater reliability. - Computes agreement metrics including Krippendorff’s alpha, Cohen’s kappa, and Fleiss’ kappa.
- Useful for evaluating consistency across different coders, models, or coding runs.
- Validates LLM-coded output against gold standard human coding.
- Computes classification metrics: accuracy, precision, recall, F1-score, and Cohen’s kappa.
- Supports multiple averaging methods (macro, micro, weighted) and per-class breakdowns.
- Creates complete audit trails following Lincoln and Guba’s (1985) concept for establishing trustworthiness.
- Captures the full decision history: models, parameters, timestamps, and parent-child relationships.
- Stores all coded results for confirmability and dependability.
- Reconstructs branching workflows when multiple coded objects are compared or validated.
- Generates replication materials: environment setup, API configuration, and executable R code.
- Use
qlm_trail(..., path = "filename")to save RDS archive and Quarto report.
For an interactive Shiny application to perform manual coding, review AI-generated annotations, and compute agreement metrics, see the companion package quallmer.app.
The package supports all LLMs currently available with the ellmer
package. For authentication and usage of each of these LLMs, please
refer to the respective
ellmer
documentation and see our tutorial for setting up an OpenAI API
key
or getting started with an open-source Ollama
model.
You can install the development version of quallmer from https://github.com/quallmer/quallmer with:
# install.packages("pak")
pak::pak("quallmer/quallmer")To learn more about how to use the package, please refer to our step-by-step tutorials.
Development of this package was assisted by Claude Code, an AI coding assistant by Anthropic, for code refactoring, documentation updates, and package restructuring.
