Skip to content

team-eyespace/Eyespace

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Eyespace Build Status

Inspiration

Generally, the elderly and visually impaired people walk with someone by their side. But what if nobody is around to help? We were very concerned about the fact that help may not be always available so we wanted to make a positive impact on the elderly and visually impaired people. Nowadays, the elderly want to have independence and not always want people around to monitor them, so we also catered to their independence.

What does Eyespace do?

Our web-app takes input in the form of images through the raspberry pi camera, processes them and through the ultrasonic sensor, it gives accurate distance calculation between an object and the person with the camera. Also, if the distance from the object to the person is small, a warning signal is sent out to the person in the form of speech through their speaker or headset. If the person does not respond to this message within 1 minute, an emergency call is sent out to 911 and another emergency contact of the user's choice (registered in the firebase database). Currently, with our prototype, we get images in the form of streams from the phone's camera, we do real-time object detection on these images and the results are read out to the user when they ask our very own personal assistant a question about their surrounding. We have written a cloud function to carry out the operations in the back-end.

How did we build it?

At Hack Harvard, we started off with setting up the Raspberry Pi 3 and it's method of input. We used multiple Python scripts to get input from the pi camera, the ultrasonic sensors and accelerometers. The data from the camera was passed through the Clarifai API which generates tags based on the image after it has been processed. We used Python to refine the results that we received from the API. The inputs from the ultrasonic sensors and accelerometers are processed using our Python script, and the results are calculated and returned to the user on the Web-App. We integrated all the components to get a web-app and a device to go along with it.

Currently, we are prototyping using a native mobile app created using Google's Flutter. We get the image from the camera and use real time object detection which is read out to the user by our own personal assistant. We have a trained AutoML model for pothole detection as well. Alongside that, all our operations are created using Google Cloud Technologies.

What were the challeneges we ran into?

We had to learn Flutter and Dart from scratch as we switched onto a native mobile app. We also ran into some issues dealing with Image Streams and their formats.

Accomplishments that we're proud of!

Firstly, we are extremely proud of this prototype of our product. Secondly we got past all challenges of the project as a team and helped each other out whenever necessary.

What did we learn from this experience?

We learned how to use and manipulate Google Cloud Technologies. Most importantly, my team learned better teamwork and cooperation among each other.

Technologies Used:

Google Cloud Vision, Firebase, AutoML, Flutter, Android, iOS, Google Cloud Functions, Google DialogFlow API, OCR.

Next stages of this project...

Our passion for helping the elderly and visually impaired will always exist, and we will continue to take this project on further in order to make it a successful product and make a change in society foe a greater good.

We thank Google for the support given to us during this project.