Skip to content

hpbl/READMEClassifier

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

GitHub README Content Classifier

What

This project contains the source code of GitHub README content classifier from the paper "Categorizing the Content of GitHub README Files" (Gede Artha Azriadi Prana, Christoph Treude, Ferdian Thung, Thushari Atapattu, David Lo), published in 2018 in Empirical Software Engineering. DOI: 10.1007/s10664-018-9660-3

How to Use

This project is written in Python 3.

Cross-validation Experiments

  1. Set up file paths in config/config.cfg. By default, CSV files listing the section titles and their labels are in input/. dataset_1.csv contains the section titles and labels for the development set, whereas dataset_2.csv contains the section titles and labels for the evaluation set. The README files corresponding to the CSV files are in input/ReadMes/ directory.
  2. Empty all database tables by running the script script/loading/empty_all_tables.py
  3. Run script/loading/load_section_dataset_25pct.py to extract and load section overview (title text, labels) and content of development set into database.
  4. Run script/loading/load_section_dataset_75pct.py to extract and load section overview (title text, labels) and content of evaluation set into database.
  5. Run the script/experiment/* scripts as required. E.g. script/experiment/classifier_75pct_tfidf.py for the SVM version.

Training Model on Existing Data and CLassifying New Files

  1. Run script/classifier/load_combined_set_and_train_model to extract and load contents and titles listed in combined development and evaluation sets (by default, defined as dataset_combined.csv in config/config.cfg) into the database.
  2. Run script/classifier/load_and_classify_target to extract and load contents of the README files in the directory specified in target_readme_file_dir variable in config/config.cfg.
  3. By default, the resulting section labels will be saved in output/output_section_codes.csv. Classifier will also identify which codes exist for each file, and which codes don't yet exist (i.e. potential for README expansion). This information will be saved in output/output_file_codes.csv

Training Model on Existing Data and Classifying New Files (Partial Steps)

  1. Run script/loading/load_section_dataset_combined.py to extract and load section overview (title text, labels) and content of combined development and evaluation sets (by default, defined as dataset_combined.csv in config/config.cfg) into the database.
  2. Place the README files whose sections are to be classified in the directory specified in target_readme_file_dir variable in config/config.cfg.
  3. Run script/loading/load_target_section_data.py to load the section heading and content data into database.
  4. Run script/classifier/classifier_train_model.py. This script will train SVM model using combined dataset in *combined tables. The resulting model, TFIDF vectorizer, and matrix label binarizer will be saved in model/ directory.
  5. Run script/classifier/classifier_classify_target.py. This script will use the saved model, vectorizer, and binarizer to classify target README files in the directory specified in target_readme_file_dir variable in config/config.cfg.
  6. By default, the resulting section labels will be saved in output/output_section_codes.csv. Classifier will also identify which codes exist for each file, and which codes don't yet exist (i.e. potential for README expansion). This information will be saved in output/output_file_codes.csv

Notes

All scripts will log output (such as F1 score, execution times) into log/ directory. Preprocessed README files (with numbers, mailto: links etc. abstracted out) are saved in temp/ directory. Patterns used for heuristics are listed in doc/Patterns.ods.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%