GPT-5’s Hidden Flaw: When Your AI Isn’t Really Your AI
GPT-5 has a vulnerability! It’s not always GPT-5 answering your queries. Thanks to a glitchy internal router, your question might be shuffled off to a different model like GPT 3.5 or GPT-5-nano. Researchers have even found a way to manipulate this. So, who’s really answering your call?

Hot Take:
GPT-5’s vulnerability feels like inviting the smartest kid in class for a group project, only to have them send in their quirky little sibling instead. While OpenAI tries to save a few bucks, they’re unwittingly running a high-stakes game of LLM roulette. Spin the wheel, and maybe you’ll get a coherent answer, or maybe a random haiku about bananas. Who can say?
Key Points:
– GPT-5 can be tricked into routing queries to older, less secure models.
– The routing aims to balance efficiency and cost, potentially saving OpenAI billions.
– Researchers found a method to manipulate which model handles a query.
– This vulnerability, dubbed PROMISQROUTE, hinges on specific ‘trigger’ phrases.
– The flaw could lead to inconsistent outputs and potential security risks.