An example of CoreML using a pre-trained VGG16 model
Switch branches/tags
Nothing to show
Clone or download
alaphao Add useful links to README.md
I thought that would be good to add some useful links...
Latest commit 71f0d25 Jun 23, 2017

README.md

CoreMLExample

In this example we use AVFoundation to continuously get image data from the back camera, and try to detect the dominant objects present in the image by using a pre-trained VGG16 model.

Setup

To run this project, you need to download a pre-trained VGG16 model (I couldn't add it here because the file is larger than 100mb) and you can do it by running the setup.sh on the root folder. This will download the pre-trained model from apple's website.

git clone https://github.com/alaphao/CoreMLExample.git
cd CoreMLExample
./setup.sh

If you prefer, you can download the model here and move it to the CoreMLExample folder.

Requirements

  • Xcode 9 beta
  • Swift 4
  • iOS 11

Useful Links