In three clicks, make an iPhone app that runs image recognition on every video frame coming live from the camera
This app will create an iPhone app that uses Tensorflow image recognition, trained on your own images. You can make a trail mix calorie counter, a skin cancer detector, a cat in sofa detector, a cake recognizer, or whatever you can come up with.
- Install Xcode 8 or higher
- Install Docker
- Install Git
- Make sure Docker and git are running
- To make training faster, open Docker
Preferences
>Advanced
, and increaseCPUs
andmemory
- Start Trainer
- Click
Browse
- You can train your model with both images and videos
- To train with images, go to
images/originals
- Create a new folder named after the object you want the app to recognize, e.g. "burgers"
- Put pictures of burgers into the
burgers
folder - To train with videos, go to
videos
- Add your videos. Rename the video files e.g.
burgers1.mov
,burgers2.mov
, etc
- Click Start Training. The first training session will take about 30 minutes, depending on how many videos and images you have.
- Click the Xcode icon. Creating your app will take about 45 minutes.
- Go to ~/projects/tf_files/tensorswift_ios/tensorswift
- Open the xcode project and open
Config.swift
- Add your labels and which web page the app should open when it recognizes them
- Add the images to
images/originals
andvideos
- Click
Start Training
- In Xcode's left hand file panel, delete the files in the folder 'model'
- Click
Reveal model files
and drag the files into themodel
folder in Xcode
Feel free to grab one of these, or get in touch about any of them:
- Optimize the model for inference
- quantisize_graph for smaller model file size
- Show Tensorboard vital stats directly in the app
- Add UI for hyperparameters
- Fix
reset
- it doesn't seem to do anything. Should remove bottlenecks and resize logs - Make it more clear how to retrain the model