Skip to content

mabehrendt/AQuA

Repository files navigation

AQuA 🌊 - Combining Experts’ and Non-Experts’ Views To Assess Deliberation Quality in Online Discussions Using LLMs

AQuA 🌊, an ``Additive deliberative Quality score with Adapters'', is a metric for assessing the quality of individual comments in online discussions based on Adapter predictions for individual deliberative quality indices. This repository contains the code to translate an English dataset to German and calculate AQuA 🌊 scores for each entry.

BibTeX Citation

If you use the AQuA 🌊 score in a scientific publication, we would appreciate using the following citations:

@inproceedings{behrendt-etal-2024-aqua,
    title = "{AQ}u{A} {--} Combining Experts{'} and Non-Experts{'} Views To Assess Deliberation Quality in Online Discussions Using {LLM}s",
    author = "Behrendt, Maike  and Wagner, Stefan Sylvius  and Ziegele, Marc  and Wilms, Lena  and Stoll, Anke  and Heinbach, Dominique  and Harmeling, Stefan",
    editor = "Hautli-Janisz, Annette  and Lapesa, Gabriella  and Anastasiou, Lucas  and Gold, Valentin  and Liddo, Anna De  and Reed, Chris",
    booktitle = "Proceedings of the First Workshop on Language-driven Deliberation Technology (DELITE) @ LREC-COLING 2024",
    month = may,
    year = "2024",
    address = "Torino, Italia",
    publisher = "ELRA and ICCL",
    url = "https://aclanthology.org/2024.delite-1.1",
    pages = "1--12",
    }

Maike Behrendt, Stefan Sylvius Wagner, Marc Ziegele, Lena Wilms, Anke Stoll, Dominique Heinbach, and Stefan Harmeling. 2024. AQuA – Combining Experts’ and Non-Experts’ Views To Assess Deliberation Quality in Online Discussions Using LLMs. In Proceedings of the First Workshop on Language-driven Deliberation Technology (DELITE) @ LREC-COLING 2024, pages 1–12, Torino, Italia. ELRA and ICCL.

Adapters

We trained Adapters similar to Falk & Lapesa (2023).

We trained the following adapters:

  • Relevance: Does the comment have a relevance for the discussed topic?
  • Fact: Is there at least one fact claiming statement in the comment?
  • Opinion: Is there a subjective statement made in the comment?
  • Justification: Is at least one statement justified in the comment?
  • Solution Proposals: Does the comment contain a proposal how an issue could be solved?
  • Additional Knowledge: Does the comment contain additional knowledge?
  • Question: Does the comment include a true, i.e., non-rhetoric question?
  • Referencing Users: Does the comment refer to at least one other user or to all users in the community?
  • Referencing Medium: Does the comment refer to the medium, the editorial team or the moderation team?
  • Referencing Contents: Does the comment refer to content, arguments or positions in other comments?
  • Referencing Personal: Does the comment refer to the person or personal characteristics of other users?
  • Referencing Format: Does the comment refer to the tone, language, spelling or other formal criteria other comments?
  • Polite form of Address: Does the comment contain welcome or farewell phrases?
  • Respect: Does the comment contain expressions of respect or thankfulness?
  • Screaming: Does the comment contain clusters of punctuation or capitalization intended to imply screaming
  • Vulgar: Does the comment contain language that is inappropriate for civil discourse?
  • Insult: Does the comment contain insults towards one or more people?
  • Sarcasm: Does the comment contain biting mockery aimed at devaluing the reference object?
  • Discrimination: Does the comment explicitly or implicitly contain unfair treatment of groups or individuals?
  • Storytelling: Does the commenter include personal stories or personal experiences?

We explain how to calcualte scores on a given dataset step by step in the following.

1. Requirements

To calculate AQuA 🌊 scores, make sure to first install all required python packages using:

$ pip install -r requirements.txt

2. Translation (optional)

The individual Adapters that are the basis for theAQuA 🌊 score are trained on a German dataset. For an evaluation of English data, the data can be translated to Germany by running the translate_to_german.py script:

$ python translate_to_german.py dataset_path dataset_name translation_col output_path --sep (optional)

As an example, to re-create the translated version of the Europolis dataset:

$ python translate_to_german.py data/europolis/europolis_whole.csv europolis cleaned_comment data/europolis/europolis_de_translated.csv

3. Inference

To calculate AQuA 🌊 scores for a German dataset, use the inference_parallel_de.py script (all example csv files are tab separated - the script also expects that the loaded csv file is tab separated):

$ python inference_parallel_de.py inference_data text_col batch_size output_path

To re-create the scores on the Europolis dataset, use:

$ python inference_parallel_de.py data/europolis/europolis_de_translated.csv comment_de 8 output/europolis/europolis_inference.csv

References

Work by Neele Falk and Gabriella Lapesa:

  • Falk, N., & Lapesa, G. (2023). Bridging Argument Quality and Deliberative Quality Annotations with Adapters. In Findings of the Association for Computational Linguistics: EACL 2023 (pp. 2424-2443). doi:10.18653/v1/2023.findings-eacl.187

Please cite the following if you use the datasets contained in this repository:

  • Europolis: Gerber, M., Bächtiger, A., Shikano, S., Reber, S., Rohr, S. (2018). Deliberative Abilities and Influence in a Transnational Deliberative Poll (EuroPolis). British Journal of Political Science, 48(4), 1093–1118. doi:10.1017/S0007123416000144

  • SFU: Kolhatkar, V., Wu, H., Cavasso, L. et al. (2020). The SFU Opinion and Comments Corpus: A Corpus for the Analysis of Online News Comments. Corpus Pragmatics 4, 155–190 doi:10.1007/s41701-019-00065-w

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages