Skip to content
Animal Detection Network
Branch: master
Clone or download

Animal Detection Network (Andenet)

The goal of Andenet is to provide a standardized dataset to be used as a baseline for developing and evaluating deep machine learning pipelines for projects interested in automatically detecting and labeling animals in trail camera (camera trap) archives.

Andenet-Desktop is a suite of tools for annotating, preparing, and exploring training data from trail cameras (camera traps).

Andenet is actively under development by Peter Ersts and Ned Horning of the Center for Biodiversity and Conservation at the American Museum of Natural History. Additional documentation will be forthcoming.



Andenet is being developed on Ubuntu 18.04 with the following libraries:

  • PyQt5 (5.10.1)
  • TKinter (3.6.7)
  • Pillow (5.4.1)
  • Numpy (1.15.4)

Install GUI libraries:

sudo apt install python3-pyqt5 python3-tk

Install pip3 and install / upgrade dependencies:

sudo apt install python3-pip
sudo -H pip3 install numpy
sudo -H pip3 install --upgrade pillow

Launching Andenet-Desktop

git clone
cd andenet-desktop

This is all you need to do to begin using the base functionality of Andenet-Desktop.

Assisted Annotation and Exporting

Assisted Annotation is ability to load an existing object detection model and use the model's prediction(s) as initial annotated bounding box. Assisted Annotation is useful approach for visually assessing the accuracy and precision of your model as you continue to collect additional training data.

Exporting to some format may require additional libraries / frameworks. Exporting to the TFRecord format, for example, requires some functions from the TensorFlow Object Detection API .

Assisted Annotation with YOLOv3 (Torch)

Additional Dependencies:

  • Torch (1.0.0)
  • Matplotlib (3.0.2)
  • Opencv-python
sudo -H pip3 install torch
sudo -H pip3 install matplotlib
sudo -H pip3 install opencv-python

Clone Ultralytics YOLOv3 repo:



Add Libraries to PYTHONPATH:


Note: This export command needs to run from every new terminal you start. If you wish to avoid running this manually, you can add it as a new line to the end of your ~/.bashrc file.

Assisted Annotation with Tensorflow and TFRecord Export

Additional Dependencies:

  • lxml (4.3.0)
  • TensorFlow (1.12.0)
sudo -H pip3 install --upgrade lxml
sudo -H pip3 install matplotlib

For detailed steps to install TensorFlow, follow the TensorFlow installation instructions. A typical user can install Tensorflow using one of the following commands:

# For CPU
sudo -H pip3 install tensorflow
# For GPU
sudo -H pip3 install tensorflow-gpu

Clone TensorFlow's models repo:

git clone

Protobuf Compilation

The TensorFlow Object Detection API uses Protobufs to configure model and training parameters. Before the framework can be used, the Protobuf libraries must be compiled. This should be done by running the following command from the models/research/ directory:

sudo apt install protobuf-compiler
# From models/research/
protoc object_detection/protos/*.proto --python_out=.

Add Libraries to PYTHONPATH

When running locally, the tensorflow/models/research/ and slim directories should be appended to PYTHONPATH. This can be done by running the following from tensorflow/models/research/:

# From tensorflow/models/research/
export PYTHONPATH=`pwd`:`pwd`/slim:`pwd`/object_detection

Note: This export command needs to run from every new terminal you start. If you wish to avoid running this manually, you can


add the export command as a new line to the end of your ~/.bashrc file.

You can’t perform that action at this time.