The company has announced that the innovative capabilities of Gemini Nano will soon enhance its accessibility feature, TalkBack. This initiative exemplifies how generative AI is being utilized to make software more inclusive for a broader range of users.
Gemini Nano is the compact version of Google’s large-language-model platform, designed for seamless on-device operation. This means it functions without the need for an internet connection. In this context, the program will generate auditory descriptions of objects, specifically catering to users who are blind or have low vision.
For instance, in a recent demo, TalkBack described an item as, “A close-up of a black and white gingham dress. The dress is short, with a collar and long sleeves, and is cinched at the waist with a large bow.”
According to the company, TalkBack users face approximately 90 unlabeled images daily. With the implementation of LLMs, the system aims to provide insights into these visuals, significantly reducing the need for manual data entry.
“This update will help fill in missing information,” emphasized Sameer Samat, President of the Android ecosystem. “Whether it’s providing more context for photos shared by family or friends or describing the style and cut of clothing when shopping online, this enhancement will be invaluable.”
This feature is expected to roll out on Android later this year. If the performance matches that of the demonstration, it could revolutionize the experience for individuals with visual impairments.
We’re excited to announce our new AI newsletter! Sign up here to start receiving it directly in your inbox on June 5.