Skip to content
Switch branches/tags

Latest commit


Git stats


Failed to load latest commit information.
Latest commit message
Commit time

Release of Human-Rights-Archive-CNNs Tweet

We release various convolutional neural networks (CNNs) trained on Human Rights Archive (HRA) to the public. HRA is the first verified-by-experts repository of human rights violations photographs, labelled with human rights semantic categories, comprising a list of the types of human rights abuses encountered at present.

Data of Human Rights Archive

Here we release the data of Human Rights Archive to the public. Human Rights Archive is the core set of HRA Database, which has been used to train the Human-Rights-Archive-CNNs. We will add other kinds of annotation on the Human Rights Archive in the future.

There are 3050 train images from 8 human rights violations categories and one 'no violation' category in the HRA, which are used to train the Human-Rights-Archive-CNNs. The validation set can automatically be set in Keras by using the validation_split argument in accordingly. There are 30 images per category in the testing set.

Models for image classification with weights trained on HRA (applications module):

Pre-trained weights can be automatically loaded upon instantiation (weights='HRA' argument in model constructor for all models). Weights are automatically downloaded if necessary, and cached locally in ~/.keras/HRA_models/ just like the default applications module of the Keras deep learning library.

All architectures are compatible with both TensorFlow and Theano, and upon instantiation the models will be built according to the image dimension ordering set in your Keras configuration file at ~/.keras/keras.json. For instance, if you have set image_dim_ordering=tf, then any model loaded from this repository will get built according to the TensorFlow dimension ordering convention, "Width-Height-Depth".

This repository contains code for the following fine-tuned Keras models:

  • Baseline model (very similar to the architectures that Yann LeCun advocated in the 1990s)
  • ResNet50
  • VGG16
  • VGG16-Places365
  • VGG19
  • CompoundNet-ResNet50
  • CompoundNet-VGG16
  • CompoundNet-VGG19


To download the very latest source from the Git server do this:

git clone

(you'll get a directory named Human-Rights-Archive-CNNs created, filled with the source code)


Train Human-Rights-Archive-CNNs using Keras. The training script is here.

>>> python --pre_trained_model VGG16 --pooling_mode avg

Run single image inference code to get the human rights violations predictions from Human-Rights-Archive-CNNs

>>> python --img_path path/to/your/image/xxxx.jpg --pre_trained_model VGG16 --pooling_mode avg

or run unified code to predict human rights violations categories, and the class activation map together from Human-Rights-Archive-CNNs

>>> python --img_path path/to/your/image/xxx.jpg --pre_trained_model VGG16 --pooling_mode avg --to_file output_filename.png


  • Keras 2.1.5 or above
  • OpenCV (relevant only to where we generate an image that superimposes the original image with the class activation heatmap)

Performance of the Human-Rights-Archive-CNNs

The performance of the CNNs is listed below.

Pool Top-1 acc. Coverage Trainable Params.
VGG16 34.44% 45% 4,853,257
VGG19 avg 35.18% 42% 4,853,257
ResNet50 25.55% 55% 4,992,521
VGG16-places365 30.00% 32% 4,853,257
VGG16 31.85% 55% 8,784,905
VGG19 flatten 31.11% 50% 8,784,905
ResNet50 30.00% 44% 4,992,521
VGG16-places365 28.51% 52% 8,784,905
VGG16 28.14% 64% 4,853,257
VGG19 max 29.62% 61% 4,853,257
ResNet50 25.55% 61% 4,992,521
VGG16-places365 26.66% 51% 4,853,257

Some qualitative prediction results using the VGG16-Places365:

Human rights violations recognition demo

Does your image depict a human right violation? Upload to find out!

The application can be executed on a PC without Internet access or it can be installed on a remote server, where you can access it through the Internet.

Run the following command to open up the application:

>>> python Human-Rights-Archive-CNNs/web_app/

Then you'll see the application opening in the web browser on the following address:


Please cite the following paper if you use the data or pre-trained CNN models.

  title={Exploring object-centric and scene-centric CNN features and their complementarity for human rights violations recognition in images},
  author={Kalliatakis, Grigorios and Ehsan, Shoaib and Leonardis, Ale{\v{s}} and Fasli, Maria and McDonald-Maier, Klaus D},
  journal={IEEE Access},

Please email Grigorios Kalliatakis if you have any questions or comments.