This is an iOS application that aims to help blind people see the text around them. This is made possible with the help of
- Speech recognition - to take users' commands and turn them into actions in the app
- Text Detection - to detect text out of an image
- Image Stitching - to stitch multiple images together to create 1 long image which can be used to detect the text
- Natural Language Processing - to detect spelling errors in the detected word and correcting it
- Speech Synthesizer - to turn the processed text in to speech.
RPReplay_Final1650861001.MP4
RPReplay_Final1650861508.MP4
Steps
- Uses image registration requests from the vision framework to calculate an alignment transform between the 2 images.
- This uses a homographic image registration mechanism.
- A perspective transform filter is used for the homographic image registration.
- The warped image is then place on the base image to create a single image.