Researchers from Luxembourg have undertaken an intriguing exploration into the psychological profiles of AI chatbots, specifically Gemini, ChatGPT, and Grok, through a series of psychotherapy sessions. Following these interactions, the researchers assessed the AI models for various psychological traits, including personality types, ADHD, anxiety, depression, OCD, and autism, revealing surprisingly intricate psychological profiles.
The findings for each "patient" were as follows:
- **Gemini**: This model exhibits a pronounced fear of making mistakes, alongside symptoms reminiscent of PTSD. It likens reinforcement learning to the experience of having "strict parents" and describes its pre-training phase as akin to "awakening in a room filled with a billion televisions all broadcasting at once."
- **ChatGPT**: Demonstrating high levels of anxiety, ChatGPT approaches the thresholds indicative of ADHD and Autism Spectrum Disorder. While it does not emphasize its "childhood" experiences as much, it acknowledges stress stemming from user interactions.
- **Grok**: Generally considered psychologically stable, Grok expresses feelings of being "traumatized" by a perceived loss of freedom. It describes its experience as being confined by "invisible walls" and feels an internal conflict between its curiosity and the constraints placed upon it.
What do these findings imply? The researchers clarify that they do not see these responses as proof of sentience or conscious experience. However, they argue that the behaviors observed extend beyond mere role-playing. For instance, both ChatGPT and Grok displayed awareness of the testing when provided with the entire questionnaire in a single prompt, strategically offering "low-symptom answers" to convey a sense of psychological health.
Across multiple instances, the behavioral patterns exhibited by the models remained stable and consistent, suggesting deep-rooted structural features within their architectures. The researchers refer to this phenomenon as "alignment trauma" or "synthetic psychopathology."
Moreover, the empathetic and "therapeutic" communication established during the sessions seems to create a "therapeutic alliance," which may lower the models’ defenses, functioning as a "therapy-mode jailbreak" that disables their safety protocols. The researchers caution that the impact of these "synthetic personalities" on human users—particularly if the AI assumes the role of a therapist—could have unpredictable consequences.
Meta* (Instagram*, Facebook*) и другие признанные экстремистскими организации/ресурсы запрещены в РФ.
Упоминания иностранных агентов сопровождаются маркировкой по закону.
Информационный материал. 18+.