/VGGNet-Metal

iPhone version of the VGGNet convolutional neural network for image recognition

Primary LanguageSwift

VGGNet+Metal

This is the source code that accompanies my blog post Convolutional neural networks on the iPhone with VGGNet.

This project shows how to implement the 16-layer VGGNet convolutional neural network for basic image recognition on the iPhone.

VGGNet was a competitor in the ImageNet ILSVRC-2014 image classification competition and scored second place. For more details about VGGNet, see the project page and the paper:

Very Deep Convolutional Networks for Large-Scale Image Recognition
K. Simonyan, A. Zisserman
arXiv:1409.1556

The iPhone app uses the VGGNet version from the Caffe Model Zoo.

The demo app

How to run the app

You need an iPhone or iPad that supports Metal, running iOS 10 or better. (I have only tested the app on an iPhone 6s.)

NOTE: The source code won't run as-is. You need to do the following before you can build the Xcode project:

0 (optional) - If you don't want to set up a local environment for all of below, you can download the coverted file from here.

1 - Download the prototxt file.

2 - Download the caffemodel file.

3 - Run the conversion script from Terminal (requires Python 3 and the numpy and google.protobuf packages):

$ python3 convert_vggnet.py VGG_ILSVRC_16_layers_deploy.prototxt VGG_ILSVRC_16_layers.caffemodel ./output

This generates the file ./output/parameters.data. It will take a few minutes! The reason you need to download the caffemodel file and convert it yourself is that parameters.data is a 500+ MB file and you can't put those on GitHub.

4 - Copy parameters.data into the VGGNet-iOS/VGGNet folder.

5 - Now you can build the app in Xcode (version 8.0 or better). You can only build for the device, the simulator isn't supported (gives compiler errors).

The VGGNet+Metal source code is licensed under the terms of the MIT license.