Skip to content
Color Your B/W Images !
Branch: master
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Type Name Latest commit message Commit time
Failed to load latest commit information.


Taking inspiration from the work by Emil Walner on how to colorize black and white images, we've built a neural network which has four components.


We split the network we had before into an encoder and a decoder. Between them, we’ll use a fusion layer.
In parallel to the encoder, the input images also run through one of today’s most powerful classifiers — the Inception ResNet v2 . This is a neural network trained on 1.2M images. We extract the classification layer and merge it with the output from the encoder.
By transferring the learning from the classifier to the coloring network, the network can get a sense of what’s in the picture. Thus, enabling the network to match an object representation with a coloring scheme.

The main difference in colorization from other visual neural networks is the importance of pixel location. In coloring networks, the image size or ratio stays the same throughout the network.


The original model posted by Emil Walner had been trained over 9.5k test images, 256x256 in size, for 2000 epochs. Any free services, didn't suffice for us to train our model online with such configs. Training our model locally allowed us to use only 1024 images at a time. On an average an epoch took 135s on a Macbook Pro having an inbuilt Intel Iris Plus Graphics 650 card, utilizing ~80% active memory. A single set of 1024 images, trained over 200 epochs, took 7.5 hrs. A total of 10 sets of images were trained (~75 hours), iteratively. This allowed us to get intermediary results. For a fully trained model, visit this

Intermediary Results

1. Black & White Image | 2. Our Locally Trained Model | 3. State of the Art Coloring Model

Screenshot 2019-04-23 at 10 42 59 PM

Screenshot 2019-04-23 at 10 43 08 PM

Screenshot 2019-04-23 at 10 43 21 PM

Local Setup

Clone the repository
Download the models from here and place them in api/ml_models directory.

cd color_em
npm install 
npm start

choose from a set of given images in directory testing_images

What now ?

Would be awesome if someone could help us with a free and reliable online service where we can train our model further.
The ML code lies in the jupyter file. Any tweakings to the script which could benefit us will be appreciated.

You can’t perform that action at this time.