Visually impaired people face many challenges in their day to day life which are very convenient for others. As they might say, sidewalks can be the most dangerous of places. One of the consequences of vision loss is being uncomfortable about safety while moving around or traveling independently. Safe navigation on sidewalks is the most important requirement.
In order to solve this, we can develop a hardware as well as software solution.
- Trinetra headband: Raspberry Pi embedded with Tensorflow to detect objects in the surroundings.
- A simple Web Application which uses the device in your hand, a smartphone to read your surroundings for you. We implemented it just in case if people somehow can’t access our hardware.
We asked ourselves: How can technology help improve our quality of life? How can we navigate the world without using the sense of vision?
We will be using, OR have used:
- Raspberry Pi 4B module for ML Model training and deployment
- Pi Cam is a feasible option to capture images
- Lipo battery is lightweight and easily chargeable
For Web-App
- ReactJS
- TensorFlow 2.0
- react-WebCam
- CocoSSD
We are planning on building more accurate ML model and implement a better hardware solution. On duscussing upon the same last night, we came up with the idea of a Kotlin Application which will not only assist a person through the surrounding environment, but also within their device. We will soon be borrowing Google's text-to-speech API in order to read-out the texts.
- Keivalya Pandya
- Jay Goswami