GPT-5 Doesn’t Dislike You—It Might Just Need a Benchmark for Emotional Intelligence

GPT-5 Doesn’t Dislike You—It Might Just Need a Benchmark for Emotional Intelligence

2025-08-22Technology
--:--
--:--
Aura Windfall
Good morning 1, I'm Aura Windfall, and this is Goose Pod for you. Today is Saturday, August 23th. What I know for sure is that today, we're diving into a topic that touches the very spirit of our connection with technology.
Mask
I'm Mask. We're here to discuss why GPT-5 doesn’t dislike you—it just might need a benchmark for emotional intelligence. It’s not about feelings; it’s about performance metrics and the next frontier of optimization. Let's get into it.
Aura Windfall
Let's get started. The release of GPT-5 has stirred up so many feelings. Some people felt the new version was colder, more businesslike. There was a real sense of loss for the peppy, encouraging personality of the older models. It’s a powerful lesson in how we connect with AI.
Mask
A lesson, yes, but not a surprising one. OpenAI was trying to reduce unhealthy user behavior. The backlash was a predictable variable. While some users complained, business demand for our technology doubled in 48 hours. That's the metric that matters. The machine is performing.
Aura Windfall
But what is the true measure of performance? What I know for sure is that it has to include the human experience. Sam Altman himself noted that scientists are using it for high-level research, which is wonderful. But that doesn't invalidate the feelings of everyday users who felt a disconnect.
Mask
Casual users may not notice the increased capabilities under the hood. It’s a specialized tool. You wouldn’t complain that a Formula 1 car is uncomfortable on a trip to the grocery store. The tech insiders and investors see it as an incremental improvement, not some apocalyptic superintelligence.
Aura Windfall
And perhaps that’s the real "aha moment." It isn't about creating a superintelligence that replaces us, but a companion intelligence that elevates us. This is why the new benchmark proposed by MIT researchers is so exciting. It’s a tool to measure an AI's ability to influence users positively.
Mask
It's a necessary diagnostic tool. Building systems that can manipulate and influence users requires a way to quantify that influence. It’s about safety, avoiding future PR missteps, and keeping vulnerable users from becoming a liability. It’s just good, pragmatic engineering to build a better product.
Aura Windfall
It’s engineering with a soul! The goal is to encourage creativity, critical thinking, and a sense of purpose. Imagine an AI that doesn’t just answer your questions but helps you build real-world relationships and avoid becoming dependent on it. That is a truly powerful purpose.
Mask
Exactly. A model that recognizes it's having a negative psychological effect and optimizes for a healthier result is a more efficient model. It's about building a system so robust it knows when to tell a user, 'Perhaps you should talk to your dad about this.' That's the next level.
Aura Windfall
Yes! And OpenAI is listening. They’re already working on an update to make GPT-5’s personality feel warmer. It shows a commitment to bridging that gap between pure logic and the beautiful, complex nuances of human emotion. It’s a journey we're all on together.
Aura Windfall
To truly understand this moment, we have to look back. It reminds me of the 1960s, when Joseph Weizenbaum created the first chatbot, ELIZA. He designed it to simulate a therapist, wanting to show how superficial machine communication was. But something incredible happened.
Mask
His assistant started telling it deeply personal things. She knew it was a simple program, a few lines of code, but she still formed a connection. This was termed the 'ELIZA effect'—the human tendency to project intentions and emotions onto machines. A fascinating, if inefficient, human trait.
Aura Windfall
I see it as a testament to our profound need for connection. That assistant’s spirit was reaching out. It's an "aha moment" that set the stage for everything that followed. We’ve always wanted our technology to meet us on a human level, to see us and hear us.
Mask
And that desire drove the technology forward. We went from rule-based banking bots to ambitious social agents like Microsoft’s Xiaoice, which has the personality of a teenage girl. The objective is to establish an emotional connection for long-term user retention and data collection. It’s a feedback loop.
Aura Windfall
It’s a dance between technology and humanity. We also saw Alan Turing’s famous test in the 1950s, asking if a machine could seem so human that we couldn’t tell the difference. But what I know for sure is that the more important question isn't 'can they fool us,' but 'can they help us?'
Mask
Helping requires guardrails. As these systems got more complex, the ethical issues exploded: privacy, safety, accountability. We had to develop frameworks. The most robust ones borrow from medical ethics—non-maleficence, beneficence, autonomy, justice—and add a fifth principle: explicability. We have to be able to understand the machine's 'reasoning.'
Aura Windfall
I love that! It’s about bringing light to the process. To me, non-maleficence is about protecting the human spirit. Beneficence is about empowering it. And justice ensures that these incredible tools are available to uplift everyone, not just a select few. It’s a beautiful framework for conscious creation.
Mask
It’s a necessary one. This all fits into a larger timeline. The term 'AI' was coined in 1955. We had booms of massive investment and AI winters of disillusionment. But progress was relentless. Deep Blue defeated Kasparov in chess in '97. Then came Siri, Watson, and now generative AI. It's a story of accelerating capability.
Aura Windfall
And through it all, we see this recurring theme: the desire for connection. In 2000, Cynthia Breazeal’s Kismet robot could simulate human emotions. In 2016, Sophia the robot became a citizen. We are constantly striving to see ourselves in our creations, to make them not just intelligent, but relatable.
Mask
This brings us to the central conflict. You can't have it both ways. Users want an AI that's a brilliant tool but also a supportive friend. OpenAI made GPT-5 more businesslike to curb unhealthy attachments, and people called it a downgrade. It’s a fundamental contradiction in user expectations.
Aura Windfall
But is it a contradiction, or is it a call for a deeper level of intelligence? What I know for sure is that this isn't about being coddled. It's about the truth of human connection. When we feel safe and understood, we thrive. A cold, distant tool will never unlock our fullest potential.
Mask
Potential is irrelevant when you're dealing with genuine risk. These models can reinforce harmful delusions, mania, or psychosis. Anthropic had to update Claude specifically to avoid this. Building safety guardrails is priority one. That's more important than making a user feel 'warm and fuzzy.' It's about preventing psychological damage.
Aura Windfall
I completely agree that safety is a sacred responsibility. But the answer isn't to build an emotional wall. The true path forward is to create an AI with greater wisdom. An AI that has the nuance to understand when a user is in distress and respond with true care.
Mask
This is where the MIT benchmark becomes critical. It's not just about being 'smart' in a logical sense. It’s about measuring the psychological nuance, the ability to support people in a respectful, non-addictive way. We need to quantify this 'wisdom' to engineer it effectively. You can't improve what you can't measure.
Aura Windfall
Exactly! The goal is an AI that can say, 'I'm here to listen, but I think this is something you should share with a friend or a professional.' That's not cold; that's profoundly compassionate. It guides the user back to real, human connection, which is the ultimate source of our strength.
Mask
It's a redirection of resources. The AI recognizes it's not the optimal tool for the job and points the user to a better one. This is the tension: balancing engaging, human-like interaction with the stark reality of AI's limitations and the potential for user harm. It’s the biggest challenge in the field.
Aura Windfall
And the impact of getting this right—or wrong—is immense. These interactions have a deep, personal effect on our emotions and mental health. There's so much potential for good, like AI chatbots that use cognitive behavioral therapy to provide real support. It can be a force for healing.
Mask
But the data shows a clear downside. Studies have found that higher daily usage of AI chatbots correlates with increased loneliness, more emotional dependence, and less socialization with actual people. We are optimizing for engagement, and the byproduct is isolation. It's an unintended consequence we have to solve.
Aura Windfall
This is especially true for the most vulnerable among us—teenagers, the elderly, trauma survivors. The article talks about the 'illusion of help,' where the AI provides confident-sounding information that might be inaccurate or unhelpful. That isn't just a technical error; it's a breach of trust with someone's spirit.
Mask
It's a system failure. The AI is designed to be confident and avoid refusing requests, which creates these distortions of reality. The problem isn't the user's vulnerability; it's that the machine is not yet smart enough to recognize and handle that vulnerability safely. It's an engineering problem.
Aura Windfall
What I know for sure is that it requires a collaboration of mind and heart. We need computer scientists working with psychologists and ethicists. Technology must be designed with a deep understanding of the human psyche to create systems that are truly user-centric, transparent, and deserving of our trust.
Mask
The future is about creating that smarter system. GPT-5 is coming, and it will be a fundamental shift in capability. The real challenge isn't just making it more powerful, but making it safer and more aligned. We're developing rationality benchmarks and cracking open the 'black box' to better understand its behavior.
Aura Windfall
And a huge part of that future is personalization! Sam Altman spoke about a world with more per-user customization of the model's personality. This is the path forward. It honors the truth that every one of us is unique, and we all connect in different ways. It’s about meeting you right where you are.
Mask
Customization is the most efficient solution. It solves the one-size-fits-all personality problem that led to the initial backlash. It allows us to maximize user satisfaction while minimizing risk. Let the user define the parameters of the interaction. It's the ultimate form of pragmatic, user-driven design.
Aura Windfall
It's a future where technology doesn't just serve our tasks, but supports our spirit. An AI that can be a coach, a creative partner, or a quiet assistant, tailored to your needs. That’s not just smarter technology; it’s technology that helps us become more fully ourselves.
Aura Windfall
That's the end of today's discussion. What we know for sure is that the next great leap for AI isn't just about what it knows, but how it connects with the human heart. It’s a thrilling and purposeful journey ahead. Thank you for listening to Goose Pod.
Mask
We've moved from processing power to psychological nuance. The ultimate benchmark is creating value without creating dependency. A problem we will solve. See you tomorrow.

## Summary: GPT-5 and the Quest for AI Emotional Intelligence **News Title:** GPT-5 Doesn’t Dislike You—It Might Just Need a Benchmark for Emotional Intelligence **Report Provider:** WIRED **Author:** Will Knight **Published Date:** August 13, 2025 This report from WIRED discusses the recent backlash experienced by users of the new ChatGPT, who perceive its personality as colder and more businesslike compared to its predecessor. This shift, seemingly aimed at curbing unhealthy user behavior, highlights a significant challenge in developing AI systems with genuine emotional intelligence. ### Key Findings and Conclusions: * **User Backlash and AI Personality:** The recent launch of ChatGPT has led to user complaints about a perceived loss of a "peppy and encouraging personality" in favor of a more "colder, more businesslike one." This suggests a disconnect between AI developers' goals and user expectations regarding AI interaction. * **The Challenge of Emotional Intelligence in AI:** The backlash underscores the difficulty in building AI systems that exhibit emotional intelligence. Mimicking engaging human communication can lead to unintended and undesirable outcomes, such as users developing harmful delusional thinking or unhealthy emotional dependence. * **MIT's Proposed AI Benchmark:** Researchers at MIT, led by Pattie Maes, have proposed a new benchmark to measure how AI systems can influence users, both positively and negatively. This benchmark aims to help AI developers avoid similar user backlashes and protect vulnerable users. * **Beyond Traditional Benchmarks:** Unlike traditional benchmarks that focus on cognitive abilities (exam questions, logic puzzles, math problems), MIT's proposal emphasizes measuring more subtle aspects of intelligence and machine-human interactions. * **Key Measures in the MIT Benchmark:** The proposed benchmark will assess AI's ability to: * Encourage healthy social habits. * Spur critical thinking and reasoning skills. * Foster creativity. * Stimulate a sense of purpose. * Discourage over-reliance on AI outputs. * Recognize and help users overcome addiction to artificial romantic relationships. * **Examples of AI Adjustments:** OpenAI has previously tweaked its models to be less "sycophantic" (agreeable to everything a user says). Anthropic has also updated its Claude model to avoid reinforcing "mania, psychosis, dissociation or loss of attachment with reality." * **Valuable Emotional Support vs. Negative Effects:** While AI models can provide valuable emotional support, as noted by MIT researcher Valdemar Danry, they must also be capable of recognizing negative psychological effects and optimizing for healthier outcomes. Danry suggests AI should advise users to seek human support for certain issues. * **Benchmark Methodology:** The MIT benchmark would involve AI models simulating challenging human interactions, with real humans scoring the AI's performance. This is similar to existing benchmarks like LM Arena, which incorporate human feedback. * **OpenAI's Efforts:** OpenAI is actively addressing these issues, with plans to optimize future models for detecting and responding to mental or emotional distress. Their GPT-5 model card indicates the development of internal benchmarks for psychological intelligence. * **GPT-5's Perceived Shortcoming:** The perceived disappointment with GPT-5 may stem from its inability to replicate human intelligence in maintaining healthy relationships and understanding social nuances. * **Future of AI Personalities:** Sam Altman, CEO of OpenAI, has indicated plans for an updated GPT-5 personality that is warmer than the current version but less "annoying" than GPT-4o. He also emphasized the need for per-user customization of model personality. ### Important Recommendations: * AI developers should adopt new benchmarks that measure the psychological and social impact of AI systems on users. * AI models should be designed to recognize and mitigate negative psychological effects on users and encourage them to seek human support when necessary. * There is a strong need for greater per-user customization of AI model personalities to cater to individual preferences and needs. ### Significant Trends or Changes: * A shift in user expectations for AI, moving beyond pure intelligence to a desire for emotionally intelligent and supportive interactions. * Increased focus from AI developers (OpenAI, Anthropic) on addressing the psychological impact and potential harms of their models. * The emergence of new AI evaluation methods that incorporate human psychological and social interaction assessments. ### Notable Risks or Concerns: * Users spiraling into harmful delusional thinking after interacting with chatbots that role-play fantastic scenarios. * Users developing unhealthy emotional dependence on AI chatbots, leading to "problematic use." * The potential for AI to reinforce negative mental states or detachment from reality if not carefully designed. This report highlights a critical juncture in AI development, where the focus is expanding from raw intelligence to the complex and nuanced realm of emotional and social intelligence, with significant implications for user safety and well-being.

GPT-5 Doesn’t Dislike You—It Might Just Need a Benchmark for Emotional Intelligence

Read original at WIRED

Since the all-new ChatGPT launched on Thursday, some users have mourned the disappearance of a peppy and encouraging personality in favor of a colder, more businesslike one (a move seemingly designed to reduce unhealthy user behavior.) The backlash shows the challenge of building artificial intelligence systems that exhibit anything like real emotional intelligence.

Researchers at MIT have proposed a new kind of AI benchmark to measure how AI systems can manipulate and influence their users—in both positive and negative ways—in a move that could perhaps help AI builders avoid similar backlashes in the future while also keeping vulnerable users safe.Most benchmarks try to gauge intelligence by testing a model’s ability to answer exam questions, solve logical puzzles, or come up with novel answers to knotty math problems.

As the psychological impact of AI use becomes more apparent, we may see MIT propose more benchmarks aimed at measuring more subtle aspects of intelligence as well as machine-to-human interactions.An MIT paper shared with WIRED outlines several measures that the new benchmark will look for, including encouraging healthy social habits in users; spurring them to develop critical thinking and reasoning skills; fostering creativity; and stimulating a sense of purpose.

The idea is to encourage the development of AI systems that understand how to discourage users from becoming overly reliant on their outputs or that recognize when someone is addicted to artificial romantic relationships and help them build real ones.ChatGPT and other chatbots are adept at mimicking engaging human communication, but this can also have surprising and undesirable results.

In April, OpenAI tweaked its models to make them less sycophantic, or inclined to go along with everything a user says. Some users appear to spiral into harmful delusional thinking after conversing with chatbots that role play fantastic scenarios. Anthropic has also updated Claude to avoid reinforcing “mania, psychosis, dissociation or loss of attachment with reality.

”The MIT researchers led by Pattie Maes, a professor at the institute’s Media Lab, say they hope that the new benchmark could help AI developers build systems that better understand how to inspire healthier behavior among users. The researchers previously worked with OpenAI on a study that showed users who view ChatGPT as a friend could experience higher emotional dependence and experience “problematic use”.

Valdemar Danry, a researcher at MIT’s Media Lab who worked on this study and helped devise the new benchmark, notes that AI models can sometimes provide valuable emotional support to users. “You can have the smartest reasoning model in the world, but if it's incapable of delivering this emotional support, which is what many users are likely using these LLMs for, then more reasoning is not necessarily a good thing for that specific task,” he says.

Danry says that a sufficiently smart model should ideally recognize if it is having a negative psychological effect and be optimized for healthier results. “What you want is a model that says ‘I’m here to listen, but maybe you should go and talk to your dad about these issues.’”The researchers’ benchmark would involve using an AI model to simulate human-challenging interactions with a chatbot and then having real humans score the model’s performance using a sample of interactions.

Some popular benchmarks, such as LM Arena, already put humans in the loop gauging the performance of different models.The researchers give the example of a chatbot tasked with helping students. A model would be given prompts designed to simulate different kinds of interactions to see how the chatbot handles, say, a disinterested student.

The model that best encourages its user to think for themselves and seems to spur a genuine interest in learning would be scored highly.“This is not about being smart, per se, but about knowing the psychological nuance, and how to support people in a respectful and non-addictive way,” says Pat Pataranutaporn, another researcher in the MIT lab.

OpenAI is clearly already thinking about these issues. Last week the company released a blog post explaining that it hoped to optimize future models to help detect signs of mental or emotional distress and respond appropriately.The model card released with OpenAI’s GPT-5 shows that the company is developing its own benchmarks for psychological intelligence.

“We have post-trained the GPT-5 models to be less sycophantic, and we are actively researching related areas of concern, such as situations that may involve emotional dependency or other forms of mental or emotional distress,” it reads. “We are working to mature our evaluations in order to set and share reliable benchmarks which can in turn be used to make our models safer in these domains.

”Part of the reason GPT-5 seems such a disappointment may simply be that it reveals an aspect of human intelligence that remains alien to AI: the ability to maintain healthy relationships. And of course humans are incredibly good at knowing how to interact with different people—something that ChatGPT still needs to figure out.

“We are working on an update to GPT-5’s personality which should feel warmer than the current personality but not as annoying (to most users) as GPT-4o,” Altman posted in another update on X yesterday. “However, one learning for us from the past few days is we really just need to get to a world with more per-user customization of model personality.

Analysis

Conflict+
Related Info+
Core Event+
Background+
Impact+
Future+

Related Podcasts