1st place solution to the MIIA Deep Learning Hackathon
Switch branches/tags
Nothing to show
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Failed to load latest commit information.


Pothole Detection

This repository contains all the files and code necessary to reproduce my winning solution to the pothole detection problem given at the MIIA - Deep Learning Hackathon. I hope this can be helpful to the participants new to the field of Machine/Deep Learning (or anyone else).

Note, since this was a hackathon, we had <10 hours to solve the problem. I'm sure there is still plenty of room for improvement and wouldn't be surprised if the code contains some bugs. I'll leave that to you to discover. Feel free to contact me if you have any questions and/or suggestions (maraisjandre9@gmail.com).


I strongly recommend the cloud service, Crestle. It takes 3 minutes to create an account and two clicks to open up a Jupyter Notebook with all the necessary libraries pre-installed and access to a powerful GPU.

Once your server is set up, clone this repo with the terminal command

git clone https://github.com/jandremarais/PotholeDetection.git 

Since the data is too big to save in a Github repo, you need to download it from this Google Drive link. If using a cloud server, you can download the data with this command (thanks to this SO answer)

python gddown.py 0B1IZ6xxwxyvTdkV0dm9hWm9pX0k train.zip

and then extract files with jar -xf train.zip. You can do the same for the test files, python gddown.py 0B1IZ6xxwxyvTSEQzaVVkYXNKdzQ test.zip

Note the directory structure:

├── test  
│   └── unknown  
│       └── QCiNkpJJqhUrhJw.JPG  
├── train  
│   ├── negative  
│   │   └── QBvfRmpwrRbuhxE.JPG  
│   └── positive  
│       └── QBJSjQDEIQPuaxb.JPG  
└── valid  
    ├── negative  
    │   └── QbKxDKiHtsYBtZq.JPG  
    └── positive
        └── QPJVBLDFqTEnvrc.JPG 

The images inside the folders of this repo are only samples and should be deleted before downloading the data.

data_crop follows the same structure. Its contents will be created by a code cell in the solution.ipynb notebook.

Edit: Since the labels of the test images are now made available, the test folder from the Google Drive link also have 'positive' and 'negative' subdirectories.

The folder custom_layers and the files densenet121.py and resnet_101.py are from this awesome repo. It contains code to create common ConvNet architectures in Keras, including pretrained ImageNet weights. See the README.md of the original repo to get access to the pretrained weights and save them in the models directory (again using the gddown.py script). I edited the model scripts for compatibility with Keras 2.0.

Now you are ready to run the code in the solution.ipynb notebook. I've included some explanations and comments in-between the code cells to increase readability.