That's not how it works. Things that require OCR are processed on device like recognizing websites, phone numbers addresses, dates, translate etc. Location based recognition
uses Maps data, all other requests are either sent to Apple PCC, Chatgpt and Google.
The OCR part, iPhones going back to iPhone 11 can do. Apple called it
Live Text.
The landmark and plant recognition iPhones back to 11 can do. Apple called it Visual
Look Up.
What visual intelligence does is bring those capabilities to a live camera view instead of inside the Photos app. The interesting thing is that you can get all those features of Visual Intelligence if you just download the google lens app and you don't need iPhone 15 Pro or iPhone 16 to get them.