Apple Introduces Visual Intelligence with iOS 18.2

Apple’s new Visual Intelligence feature revolutionizes image-based AI for iPhone 16 Pro users.

The release of iOS 18.2 marks a significant milestone in Apple’s AI advancements, introducing long-promised features like Genmoji and Image Playground. Among these, Visual Intelligence stands out as a transformative tool designed to enhance how users interact with the world through their devices. Currently exclusive to the iPhone 16 Pro and Pro Max, this feature was unveiled at Apple’s September event.

What is Visual Intelligence?

Visual Intelligence serves as Apple’s counterpart to Google Lens, combining advanced camera capabilities with AI to analyze images in real time. It allows users to access detailed information about objects, businesses, or text simply by pointing their device’s camera. From translating text and summarizing content to integrating with ChatGPT and Google Image Search, the feature opens new possibilities for discovery and interaction.

Limitations and Access

Two key restrictions accompany Visual Intelligence. First, the feature is currently only available on the iPhone 16 Pro and Pro Max. Apple has hinted at potential availability for older devices in the future, acknowledging the enduring popularity of tools like Google Lens since its 2017 debut.

Second, accessing Visual Intelligence requires joining a waitlist. Users can navigate to “Apple Intelligence & Siri” in their settings to join. Once approved, the tool is ready for use.

How to Use Visual Intelligence

Activating Visual Intelligence involves long-pressing the Camera Control button, located at the bottom right of the device’s screen. This opens the Visual Intelligence interface, where users can analyze objects, businesses, or text through several interactive options.

  • Interacting with Text: Point the camera at text and tap the circle at the bottom of the screen to analyze it. Options include translating, summarizing, or having the text read aloud by Siri. Contact details, like phone numbers and email addresses, are recognized and actionable for tasks such as making a call or creating an event.

  • Interacting with Businesses: Pointing the camera at a business sign provides details like hours of operation, menus, services, or the ability to place orders and make reservations. This feature is currently limited to users in the United States.

Partner Integrations: ChatGPT and Google Image Search

Visual Intelligence leverages both ChatGPT and Google Image Search for extended functionality:

  • ChatGPT: Aimed at providing information about objects, users can point their camera at an item and tap the ChatGPT icon. For example, scanning a hand cream bottle reveals product details and allows follow-up questions such as pricing or purchase locations.

  • Google Image Search: Selecting this option displays similar images from the web, useful for comparison shopping or finding deals. While it identifies price points, users must complete purchases manually.

Visual Intelligence positions Apple at the forefront of mobile AI integration, offering users a powerful, intuitive tool to explore and interact with their surroundings. As the rollout continues, Apple’s ongoing commitment to expanding access and functionality promises even greater innovations ahead.

Next
Previous