Skip to content

m4tice/qm_project

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

qm_project

Bicycle project crowd evaluation.

This repository is created for evaluating the quality and performance of a new crowd in the project Bicycle crowd evaluation.

For an effortless setup, simply open the file 'qm_project.ipynb' in Jupyter Notebook or Colab. It contains all the description, analyzation and result of all the tasks.

The directory 'main_scripts' contains the scripts used for different analyzing purposes:

  • Task 1:

    • task_1a.py
      • analyze the annotators contributing to the dataset and create the 'annotators.csv'.
    • task_1b.py
      • analyze the annotation time and create the 'annotation_time.csv'.
    • task_1b_by_image.py
      • analyze annotation time grouped by images and create the 'annotation_time_by_image.csv'.
    • task_1b_by_user.py
      • analyze annotation time grouped by users and create the 'annotation_time_by_user.csv'.
    • task_1c.py
      • analyze the amount of results of each annotator and create the 'annotator_result_count.csv'.
    • task_1d.py
      • analyze the highly disagree questions and create the 'question_answers.csv', 'question_groups.csv' and 'highly_disagree_group.csv'.
  • Task 2:

    • task_2.py
      • analyze the occurrence of samples with label 'cant_solve' and 'corrupt_data' marked as True, and create the 'grouped_unsolved_data.csv'
  • Task 3:

    • task_3.py
      • analyze the balance of the reference set
  • Task 4:

    • task_4_export.py
      • The only purpose is to collect data by doing cross-checking all answers with the reference set and create the 'annotators_quality_assessment.csv'.
      • The execution takes approximately 45 minutes to complete; therefore, if the file 'annotators_quality_assessment.csv' already exists, there is no need to execute this script.
    • task_4_visualization.py
      • Load the 'annotators_quality_assessment.csv', using the loaded data to visualize and analyze the answers of all annotators
    • task_4_visualization_extra.py
      • Load the 'annotators_quality_assessment.csv', using the loaded data to visualize and analyze the answers of annotators with sample sizes greater than the standard deviation limit.
    • task_4_good_bad_annotators.py
      • Load the 'annotators_quality_assessment.csv', using the loaded data to retrieve data of a group of wanted annotators

The directory 'files' contains the .csv files created from the .py files
The directory 'pictures' contains the plots created from the .py files

Releases

No releases published

Packages

No packages published