I have been waiting to try Visual Intelligence since Apple first unveiled the iPhone 16 in September. After all, there’s a new button (Camera Control) on my iPhone, and I haven’t been using it for photography.
Camera Control hasn’t clicked with me after a few months with the iPhone 16 Pro Max. I take a decent amount of photos with my iPhone, but any time I try to use the dedicated button, it feels cumbersome and confusing – so much so that I’ve resorted back to my trusty touchscreen and Lock Screen shortcut.
Enter iOS 18.2 which is now installed on my iPhone 16 Pro Max, and so far, I’m a big fan at the prospect of what Visual Intelligence can become.
What is Visual Intelligence? I hear you ask. Well, it’s an Apple Intelligence feature exclusive to the iPhone 16 lineup and takes full advantage of Camera Control. You launch it by long-pressing Camera Control and then snap a photo of whatever you’re looking at. From there, you can ask ChatGPT for information, search Google, or highlight any text in the photo. Think of Visual Intelligence as Apple’s version of Google Lens with its own hardware button to access on the fly.
You can launch Visual Intelligence from anywhere, even the Lock Screen, which makes it incredibly useful whenever you want to do a quick search. My first test was taking a picture of my Game Boy Camera on my desk. As mentioned above, visual Intelligence gives you a few options, so I first used Google Search to find the product. Then, I asked ChatGPT for information, and it was able to tell me all about the Game Boy Camera’s history. From there, you can ask follow-up questions, so I asked, “When did the Game Boy Camera launch in Europe?” ChatGPT obliged with the correct answer.
While it’s still technically in beta, despite the official iOS 18.2 launch, Visual Intelligence worked a treat with a recognizable product like the Game Boy Camera – I’m not sure how often I’d use it to search for an item, but considering it’s just a simple long press away, it might become my go-to way of searching the web for things.
Another great use for Visual Intelligence is when you’re out and about and want to see information about a shop, cafe, bar, or restaurant. I tested it with a local coffee shop, and while it didn’t work like Apple showed off in its demo, I think that’s more down to the early beta version I’m testing than the feature itself.
In that demo, Apple showed that Visual Intelligence could determine a dog breed. I tried this with my French Bulldog, and while I could search Google for similar dogs, it couldn’t give me a straight-up answer.
That kind of sums up Visual Intelligence in its current form. It has huge potential: I love the way it gives Camera Control a genuine purpose, and when it works, it’s fantastic. But it’s still cooking in the oven, and sometimes it fails to pick up on what I’m asking for (for example, the text summary option sometimes appears, and other times it doesn’t).
One thing is for sure, however: Visual Intelligence makes total sense to me now, and I finally understand why Apple added Camera Control to the new iPhones. It’s the kind of Apple Intelligence feature that I can see people turning to when they need a quick answer, as long as it works smoothly, and the ChatGPT and Google integration make it multi-faceted.
I love testing new iOS features, and iOS 18.2 might just be the most exciting update we’ve seen in years. After using the software for a month or so and with access to Genmoji and Image Playground, I can confidently say that iOS 18.2 feels like the iOS 18 and Apple Intelligence we’ve been waiting for.
Visual Intelligence has a lot to offer, and I’m incredibly excited to see where Apple takes this Apple Intelligence feature over the next year. Exclusive to the best iPhones, this could be the reason to buy an iPhone 16 – who would’ve thought it could be Camera Control?