Navigation
Recherche
|
iOS 26: Visual Intelligence Now Searches On-Screen Content
lundi 9 juin 2025, 19:40 , par MacRumors
![]() The feature, which previously worked only with the camera to identify real-world objects, now analyzes on-screen content. Users can ask ChatGPT questions about what they're viewing or search Google, Etsy, and other supported apps for similar items and products. Visual Intelligence recognizes specific objects within apps – like highlighting a lamp to find similar items online. The system also detects events on screen and suggests adding them to Calendar, automatically extracting dates, times, and locations. Accessing the feature appears straightforward: users press the same button combination used for screenshots. They can then choose to save or share the screenshot, or explore further with Visual Intelligence. The update basically makes Visual Intelligence more of a universal search and action tool across the entire iPhone experience. Apple says the feature builds on Apple Intelligence's on-device processing approach, maintaining user privacy while delivering contextual assistance across apps. Tags: Visual Intelligence, WWDC 2025This article, "iOS 26: Visual Intelligence Now Searches On-Screen Content" first appeared on MacRumors.comDiscuss this article in our forums
https://www.macrumors.com/2025/06/09/ios-26-visual-intelligence-now-searches-on-screen-content/
Voir aussi |
59 sources (15 en français)
Date Actuelle
lun. 9 juin - 23:40 CEST
|