Google Bard Supports Multimodal Interaction
Google Bard Now Understands Images Alongside Text
(Google Bard Supports Multimodal Interaction)
Google upgraded its Bard AI tool today. Bard gained multimodal interaction capabilities. Users can now submit images with text prompts. Bard analyzes both and responds with text answers.
This feature works in English initially. People worldwide using English can access it now. Support for additional languages is coming later. Google confirmed the update in an official blog post.
The multimodal function simplifies complex tasks. Users might photograph a painting and ask about its artist. Bard identifies the artwork and provides historical context. Travelers could snap a landmark photo requesting visiting hours. Bard extracts text from images like receipts or documents too.
Bard relies on Google Lens technology for image analysis. Safety measures prevent misuse. Bard automatically blurs faces in uploaded photos. Offensive image filters activate during processing.
Google VP Sissie Hsiao stated the enhancement makes Bard more intuitive. She emphasized real-world problem-solving applications.
Bard remains an experimental project. Google encourages user feedback to refine performance. This move follows rival AI systems adding visual features recently.
(Google Bard Supports Multimodal Interaction)
The update requires no special settings. English users see a new image upload icon in Bard’s interface.