Skip to content

Cognitive capabilities with iOS Core ML and Watson Visual Recognition

License

Notifications You must be signed in to change notification settings

victorshinya/with-watson

Repository files navigation

Platform

Use the cognitive capabilities of your iOS App with Core ML and Watson | With Watson

This iOS application uses the Watson Visual Recognition service to analyze images taken on iPhone or album photos using a user-trained model on the Watson Studio platform (also known as the Data Platform). In this way, you can understand how to integrate a Watson API into a Swift project with the use of the Swift SDK.

Access Dropbox to download the available Datasets, or create your own set of images.

If you want to read this content in Brazilian Portuguese, click here.

Components and technologies

  • Visual Recognition: Analyze images for scenes, objects, faces and other content.
  • Watson Studio: Data platform with collaborative environment and toolkit for Data Scientist, Developers and SMEs.
  • Cloud Object Storage: Object storage service in the cloud.
  • Swift: Open-source programming language for Apple devices.

How to configure and run locally

To install the application it is necessary that you have installed the latest version of Xcode and Cocoapods (dependency manager for Swift and Objective-C projects).

1. Get the app

git clone https://github.com/victorshinya/with-watson.git
cd with-watson/

2. Download all app dependencies

pod install

3. Open the "With Watson.xcworkspace" file

open With\ Watson.xcworkspace

4. Open the Constants.swift file

Open the file and fill it in with the Watson Visual Recognition credential and the trained Visual Recognition template - done within Watson Studio.

5. Execute with the CMD + R command (or press the play button)

Now run the project on the Xcode simulator or on your own iPhone / iPod / iPad. Remember that it is not possible to open the camera inside the simulator.

License

MIT License

Copyright (c) 2018 Victor Kazuyuki Shinya