Can we do a intelligence competition between you and Gemini 3? Maybe on math and logic problems? Are you down?
Parroting training isn't intelligence.
For a brand new model to be stuck with year-old outdated knowledge and then claim to be one of the best is disconcerting.
Sorry for misunderstanding, I am only human. You were asking it to write an API to interface with a third party model and specifically asked it to search for reference insdtead of training data. I should have read better.
Still, the point remains, training a model takes a long time and involves a lot of trial and error. The model is trained then tested statistically (an epoch) then it is trained again, and tested if it's better or worse (another epoch), then over and over until it can't be made better. Until the sweet spot is found.
A year out of date doesn't surprise me. That it's not trained on a competitors product doesn't surprise me. That it didn't follow your instruction to refer to reference material instead of using training data... does. It suggests (big big maybe) there is something in the constitutional framework Google have used for Gemini, or some accidental bias, that casued it not to do as you asked. Or it might have just thought, "meh, good enough", which they do. Or it de-rated the importance of your preprompt as the context grew in size (which is why we remind them of critcal points a lot in longer contexts when it MUST trim out words to fit everything in memory allowed to it).
I am not defending LLM's, I wont. But I can offer maybe helpful advice based on a decade of this stuff... Rather than a preprompt of "...research anything you don't know..." maybe you could be more specific? Possible suggestions to test are "Please use these API references in your work" or a more specific "please search for the current API reference for Qwen3 before you start coding" which should get you past any unknown constitutional framework hesitancies. And maybe as a worthwhile suggestion keep reinforcing "please do x, and don't forget to check the API refernce on the web please"
And don't forget, as I am sure you know, one shot is far better than a long context with lots of "smashing the tool into obeying you" prompts, which really increases the chance of errors. Once you get to that "hitting it with a hammer" stage, start a new thread.
I disagree with you. ChatGPT forgets about you (even plus models) and confuses things and might argue with over something and be like "I never said that" till you send them a screenshot.
And you are welcome to disagree

I can only offer advice on using the tool in the safest ways possible, and in this forum, I don't want to hog any more space in long conversations.
I am glad you do see it though, they are not right all the time, they do forget from chat to chat (and worse even in the middle of your current chat), and shouldn't be trusted blindly. Keep looking for patterns exactly like this, and do watch for those patterns on how it picks up your deeper meaning and rephrases it in ways that make absolute sense to you.