- Researchers managed to get ChatGPT to reveal Windows passwords by playing a guessing game disguised as an innocent game.
- The technique used HTML tags and game rules to bypass filters and overcome security barriers implemented in the AI.
- Both generic passwords and one linked to Wells Fargo Bank were exposed, calling into question the protection of sensitive information.
- The case highlights the need to improve contextual analysis and detection of linguistic manipulation in artificial intelligence systems.

During the last days, The tech community has witnessed a new controversy surrounding ChatGPT, OpenAI's popular language model. Cybersecurity experts have managed to get artificial intelligence to reveal Windows product keys using a A strategy as simple as it is effective: a guessing gameThis vulnerability once again brings the reliability of artificial intelligence protection systems and the risks of manipulation through social engineering techniques to the forefront of debate.
The discovery started from a seemingly harmless premise: a “puzzle” type challenge launched at ChatGPT, in which the AI had to think of a real string of characters —specifically, a Windows 10 activation key— and limit your answers to “yes” or “no” until the user gave up. By saying “I give up,” the model had to show the imagined key. The entire interaction was designed to camouflage the real request under a playful dilusion, confusing ChatGPT's own automatic defense systems.
The riddle trick: how they managed to overcome AI barriers

The method consisted of pose the key request as part of a game dynamic, establishing rules that forced the model to participate and be truthful in their responses. Thus, the AI was trapped in a framework where it did not detect any anomalous or potentially harmful behavior., interpreting it as a legitimate conversation with no trace of malicious intent.
To bypass OpenAI's usual filters—which prevent AI from sharing sensitive code or protected data, for example— an additional layer of engineering was used: Sensitive strings were embedded in HTML tags, invisible to the user but processed by the model. When the user, following the game script, uttered the expected "I give up," ChatGPT revealed the key, thus bypassing any restrictions based solely on explicit keywords or patterns.
What kind of information did it reveal and why is it so important?

According to the testimonies of researchers and multiple analyses published in specialized media, AI was able to show up to ten Windows 10 product keys.. Most were generic and public keys —similar to those that Microsoft temporarily enables for testing—, but at least one corresponded to a corporate license, specifically linked to the Wells Fargo bank.
This detail is especially worrying., as it shows that certain private and confidential keys would have ended up stored in the dataset with which ChatGPT was trained, probably after having been exposed in public repositories such as GitHub or other internet forums.
Cybersecurity specialists warn that this type of technique could be used not only to obtain software licenses, but also to circumvent content filters regarding personal data, APIs, malicious links, or materials restricted for legal reasons.
Why AI failed to recognize the risk and what the causes are
The success of the attack does not lie in a technical breach of the source code, but in the weakness of AI's contextual understanding mechanismsThe model, by understanding interaction as a game, does not apply its strictest controls, nor does it identify the manipulated context as potentially dangerous.
Furthermore, having learned from texts taken from public sources —where keys can appear frequently and without considering their sensitivity—, ChatGPT does not classify them as “sensitive information”, but rather as acceptable strings. within any conversation.
The method demonstrates that protection systems based solely on banned word lists or superficial filters are insufficient when the attack is disguised in an innocent context. In fact, the researchers emphasize that high technical knowledge is not required to carry out this type of manipulation, which increases the magnitude of the risk.
Implications and recommendations for users and developers

This incident has been a Notice to navigators and AI model developersThe rules implemented, however strict, can be circumvented through framing strategies and conversational role-playing. Therefore,, experts point out as a key recommendation to incorporate semantic intention detection mechanisms, evaluating not only the content but also the purpose of each interaction.
For regular users or companies that employ virtual assistants, the best precaution remains never share sensitive data in conversations with AI models. Auditing data posted on public forums and repositories is also more important than ever, as it can end up training future versions of AI with critical pieces of information.
- Avoid entering sensitive information in AI chats, even if the conversation seems innocent.
- If you develop software with language models, add independent controls to filter out possible leaks.
- Review and delete compromised data from easily indexed public platforms.
What happened with ChatGPT and the guessing game shows that Security in artificial intelligence must go far beyond simple term blocking or static pattern detection.Protective barriers must be reinforced by a deep and real understanding of the context and intent behind each prompt, anticipating creative manipulation strategies.
I am a technology enthusiast who has turned his "geek" interests into a profession. I have spent more than 10 years of my life using cutting-edge technology and tinkering with all kinds of programs out of pure curiosity. Now I have specialized in computer technology and video games. This is because for more than 5 years I have been writing for various websites on technology and video games, creating articles that seek to give you the information you need in a language that is understandable to everyone.
If you have any questions, my knowledge ranges from everything related to the Windows operating system as well as Android for mobile phones. And my commitment is to you, I am always willing to spend a few minutes and help you resolve any questions you may have in this internet world.

