Need Advice on Training a Model for Blind Navigation App (iOS)

Hey everyone!

I’m currently working on a project for my master’s degree, and I’d love to get some feedback, guidance, or even collaboration if anyone’s interested!

The idea is to develop an iOS navigation app for blind users, which uses the phone’s camera and machine learning to analyze the environment in real time, and provide audio guidance to help users navigate safely.

How it works (in theory):
The app will use the phone camera to detect important objects and features in the environment — things like:

  • Pedestrian paths
  • Traffic lights
  • Dangerous areas (e.g., stairs, escalators)
  • Braille signs/text (if possible)

It will also combine this visual input with the user’s GPS location to help guide them toward their destination.

I’m still defining what the ML model should be able to recognize and how to structure everything. Right now, I’m trying to figure out:

  • Should I use object detection, image segmentation, or a combination of both?
  • How can I find or build datasets for training (e.g., traffic lights, stairs, Braille)?
  • Is this realistic as a single project, or should I drop some features for now?

If anyone has experience with similar projects, or knows good resources, tutorials, or datasets — I’d be super grateful for any tips or advice!

Thanks a lot in advance!