Microsoft’s Seeing AI app scans and narrates the world. It allows blind people or people with low vision to have more access to information around them. Now, the app can take advantage of the LiDAR scanners in the iPhone 12 and iPhone 12 Pro Max. With LiDAR support, the app can utilize spatial audio to scan rooms.
Microsoft has shown off the Seeing AI app several times over the last few years, including when it began as a prototype and in real-world examples.VPN Deals: Lifetime license for $16, monthly plans at $1 & more
LiDAR support and the app’s latest update improve the capabilities of the app. Here’s the complete changelog as found on the App Store:
The new World channel, available on devices with a LiDAR scanner running iOS 14, enables you to explore an unfamiliar space in 3D, using spatial audio. When wearing headphones, you will hear objects around you announced from their location in the room. You can also find a particular object by placing an audio beacon on it. We are keen to hear your feedback on this early experiment, and invite you to work with us as we explore this new area together with the community.
On iPhone 12 Pro and Pro Max, the haptic proximity sensor enables you to point the LiDAR scanner and feel the distance to things around you
The main screen has been visually redesigned to improve contrast and widen the camera’s field of view
Improvements to image descriptions on the Scene channel, and when browsing photos on your phone
Improved text recognition accuracy on the Document channel
Seeing AI is now available in seven additional languages: Czech, Danish, Finnish, Greek, Hungarian, Polish, and Swedish
Plus, various bug fixes under the hood
In addition to using the LiDAR scanner to help people explore rooms, the app can use the technology to place audio beacons onto objects. It also has a new feature that uses the haptic proximity sensor to feel the distance that objects are away from you.
We may earn a commission for purchases using our links. Learn more.