Skip to content

geohci/side

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

30 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

This is a fork of Facebook's Side project to showcase the verification engine and enable live checking of citations on Wikipedia without relying on the pre-processed WAFER data.

Code

  • verify_wikipedia: original Side code (conf/ and /dpr) but streamlined to only include code relevant to running the verification engine (verifier model) -- i.e. I removed code for building the Sphere indexes and training models. I also added code for running a Flask API for running the verifier model (wsgi.py), fetch passages from external sources (passages/web_source.py), and fetch claims from Wikipedia articles (passages/wiki_claim.py).
  • api_config: Scripts / config for setting up the verifier model Flask API on a Cloud VPS server.
  • tool_ui: code for Flask-based UI for calling the verififer model to be hosted on Toolforge.

External data dependencies

  • The model itself can be downloaded per the original README from this URL (8.0GB). I have not modified the model. That file contains a few important pieces:
    • verifier/outputs/checkpoint.best_validation_acc: model weights (notably, you can delete checkpoint.best_validation_loss and free up 4GB)
    • verifier/predictions/best_validation_acc__wafer_ccnet/checkpoint_cfg.yaml: model config
    • verifier/.hydra/: more config
  • Running the verifier model for the first time will automatically download the base HuggingFace Roberta-large model (1.3GB) and cache it locally.

Test it out

About

The AI Knowledge Editor

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 94.4%
  • Shell 5.6%