Skip to content
This repository has been archived by the owner. It is now read-only.
This Project aims to train a model able of recognizing six different places in our university (SRU).
Jupyter Notebook Python JavaScript Vue CSS
Branch: master
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Type Name Latest commit message Commit time
Failed to load latest commit information.
Usage Sample

Computer Vision and Machine Learning

SRU:Places Recognizer Project

Mahya Mahdian and Mohammad Hassan Sattarian

This Project aims to train a model able of recognizing six different places in our university (SRU).

Chosen places (Model Classes) :

  • Computer Campus
  • Architecture Campus
  • Field
  • buffet
  • Self
  • Culture house

Quick Start

Download the APP connect to server bellow and take picture or choose from gallery to predict:



Clone the repository open a terminal and enter:

cd "Usage Sample"
python [ImagesAddress]


Model is a fine tuning implementation based on VGG16: Places365 which is (obviously) a VGG16 network pre-trained with places images, more specifically places365 dataset, using only convolutional layers of base model with having 5 last convolutional layers unfrozen and trained connected to a 2-layered fully-connected network with 256 and and 6 nodes, having, respectively Relu and Softmax as activation functions for detecting nonlinearities and coding result in 6 classes.


Dataset used to train this model contains of 4800 images in 6 classes, each class representing a specified place in SRU university. images divided into approximately 3000 images for training and 1800 images for testing. each class has 500 images as train set and 300 images as test set.

for collecting dataset pictures and videos have captured from foresaid places in university from different distances, angles and times not just from the front view of the place but all around it to cover as most predict cases as possible. images then reviewed, videos frames extracted and proper pictures selected also processed, rotated and resized to feed out model. images are resized to 108x192 pixels so the model is not that heavy and still has enough features to predict well.


For better accessibility and providing a graphical user interface (GUI) to use the model, an Android App is designed and created. using the App user can predict an image, taken in the app or chosen from the gallery, by writing the prediction server address. image then sent to the server and prediction result would be shown after few moments.

You can’t perform that action at this time.