Skip to content

Scripts for the paper "Models of Literary Evaluation and Web 2.0. An Annotation Experiment with Goodreads Reviews", presented at CHR2024

License

Notifications You must be signed in to change notification settings

SimoneRebora/CHR2024

Repository files navigation

CHR2024

Scripts for the paper:
Simone Rebora and Gabriele Vezzani, "Models of Literary Evaluation and Web 2.0. An Annotation Experiment with Goodreads Reviews", CHR2024 (PDF)

Preliminary notes

All ".ipynb" scripts are designed to work with Google Colab. You will need to connect your Drive to the Notebooks and have all datasets saved in the folder "MyDrive/CHR2024".
All ".py" scripts should be run locally (python3 my_script.py). You will need to install the required packages (pip install -r requirements.txt) and setup an OpenAI API key.
Because of copyright limitations, we cannot share the datasets here. If interested in getting access to them, you can contact us by describing your intended use.

Scripts overview

The scripts are linked to different sections of our paper. Here below you can find a brief description and an indication of the section of the paper where the script was used.

  • Compute_Agreement.ipynb (Paper section 3.2). Script used to calculare inter-annotator agreement.
  • Transformers_kfold.ipynb (Paper section 4.1). Script used to fine-tune three Transformer models on the annotated dataset.
  • Transformers_learning_curve.ipynb (Paper section 4.1). Script used to calculate efficiency of the best-performing model with differing amounts of training materials.
  • Transformers_save_model.ipynb (Paper section 4.1). Script used to save the best-performing model when trained on a selected dataset. The model has been published on HuggingFace.
  • GPT_annotation.py (Paper section 4.2). Script used to annotate the selected dataset with GPT-4 (through the OpenAI API).
  • GPT_annotation_fewshot.py (Paper section 4.2). Script used to annotate the selected dataset with GPT-4 (through the OpenAI API) with few-shot strategy.
  • Evaluate_GPT_prompt_engineering.ipynb (Paper section 4.2). Script used to evaluate the efficiency of different GPT4 prompting strategies on the selected dataset.

About

Scripts for the paper "Models of Literary Evaluation and Web 2.0. An Annotation Experiment with Goodreads Reviews", presented at CHR2024

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published