Team DeepThings (Mez Gebre and I) won the Best Product Category at the Deep Learning Hackathon in San Francisco. We developed in three days a real-time system capable of identifying objects and speaking what it sees, thinking about making a useful tool for the visually impaired, as it could make navigation easier. Proof of concept on a laptop, final model running on Android.
This is only the first prototype for Windows.
- Get the Webcam feed without bottlenecks.
- Recognize images using Inception v3.
- Text to speech with Google TTS API.
- Making a functional model.
- Tuninning the parameters.
- Output visual display of the results.
This module requires:
Just run:
python classify_real_time_v2.py
The output should look like this:
For more information, check my medium post here
This proyect is Copyright © 2016-2017 Lucas Gago. It is free software, and may be redistributed under the terms specified in the MIT Licence.