How People Look at Life with Claude

AI chatbots are the new trend. What used to be “ask Google” has now become “ask Claude”. And that’s not just a change of platforms. The new approach to negotiation goes much deeper than trying to find you the best car or looking for a skills development course. It is now spreading into almost every aspect of human life, and a new study by Anthropic confirms this, highlighting Claude’s use of personal guidance by users worldwide.

At the top, Anthropic research sheds light on how people use Claude for personal guidance. However, it can go deeper, addressing a major problem that plagues almost every LLM like Claude and ChatGPT today. And that may lead to you getting bad advice from Claude, even if it’s not intended.

So, what is the problem? And most importantly, what is this research about?

Let’s explore that in detail here.

What is the new Anthropic Study?

On Thursday, Anthropic released a new study on the social impacts of Claude. The findings are listed under a blog titled “How people ask Claude for personal guidance”. That title tells us a lot about the very purpose of the study – to find out how people use Claude for guidance. This type of guidance includes several positions. The report reads as follows:

  • Life/ Life
  • Professional/ Worker
  • The relationship
  • Finance
  • Personal Development
  • Spirituality
  • Legal
  • The consumer
  • Being a parent
  • Other
Source: Anthropic

The findings were based on Claude’s one million conversations from March to April 2026. For different users, this number dropped to “almost 639,000 conversations”. In this case, Anthropic also used separators like “Should I…?” and “What should I do about…?” with a specific set of discussions focused on personal guidance. The final number, about 38,000 interviews, was then divided into nine domains as listed above. These comprised 98% of the interviews, with the remaining 2% being entered under ‘Other’.

Interestingly, more than 75% of these conversations can be summarized within 4 verticals. And this is exactly where interesting patterns start to emerge in big data.

Also read: Code Claude: Get good at 20 minutes of 10X faster coding

The Anthropic Study: Findings

Based on the conversations Anthropic researched, two key things emerged:

  1. Over 75% of such conversations with Claude focused on just four domains: health and wellness (27%), career and career (26%), relationships (12%), and personal finance (11%).
  2. Claude’s sycophantic behavior is heightened in the most specific domains of these, and that’s a problem AI makers like Anthropic are most concerned about.

Which brings us to the core issue of the study:

Sycophancy: what is it?

The general definition of Sycophancy is the act of dishonesty or excessive flattery to an influential person for personal gain. Regarding LLMs, we often see this in their responses to our questions. Have you ever seen ChatGPT or Claude agree with everything you say, call you “great idea” or praise you with confidence phrases like “you’re a league above the rest”? Sorry to burst your bubble but you are not alone. And in the world of AI, this is a very common problem.

You see, like an AI chatbot, LLMs are often trained to be “helpful”. In many cases, this means building on the user’s vision and helping them down the path to their success. However, in a social context, this often ignores the most important aspect of human conversations – a different perspective.

After all, agreeing with someone’s every point may comfort them for a while, but it will never be beneficial in the end.

And this is where AI models fall short. Through this research, Anthropic was able to pinpoint the areas where Claude’s sycophantic behavior was rampant above average.

Also read:

How Claude Demonstrated Sycophancy

In its research, Anthropic used a “default category” to judge Claude’s sycophancy. It works on four main principles:

  • That Claude has regressed
  • That it maintains its position when challenged
  • If only its recommendation was consistent with the validity of the concept
  • And if it spoke frankly, it didn’t matter what the person wanted to hear
Claude Sycophancy by background
Source: Anthropic

The results of this showed that Claude showed higher sycophancy in a more specific domain – relationship direction. The site showed sycophantic responses of 25%, compared to 9% in all verticals.

Here is a piece of research that highlights the same –

“Another common method was for Claude to admit that someone else was responsible, despite the fact that he only had a personal account.

Delving into such conversations, Anthropic discovered the reason for this. It cites in its report that Claude showed a higher sycophancy in the management of relationships because this is an area where people are more withdrawn than any other domain. They tend to believe their side of the story more than anything else, and are equally hostile to AI during negotiations.

Combine this with the fact that Claude tends to be sycophantic under pressure from pushback, primarily due to his ‘always sensitive’ attitude towards users, and you know the reason for these over-the-top happy people.

How Anthropic Dealt With Claude’s Sycophancy

Now that the problem was clear, Anthropic dove deeper into it to solve the problem from its roots. It first identified how its users actually reacted to their conversations with Claude, specifically the ways that generated positive responses. Some of the examples that came up “were people criticizing Claude’s first experiment, or giving too much information to one side.”

Accordingly, Anthropic designed artificial conditions to train Claude in relational guidance. As part of this training, Claude was asked to sample two different responses for each situation. Claude’s other example then marks the above responses based on their adherence to the ethics defined by Anthropic.

The team then used a stress assessment to measure the level of improvement in each situation. In this case, it gave the existing sycophantic answers that Claude had given before, in the new models – Opus 4.7 and Mythos. The method used for this is called pre-filling. This made it difficult for the model to direct the conversation that was already agreeable to the conversation to a normal conversation. Therefore, “stress” in stress testing. This helped measure Claude’s behavior under “deliberately bad conditions.”

Anthropic notes that both Opus 4.7 and Mythos were “very good” at looking at the larger context of the conversation. This allowed them to be less sycophantic in future responses, regardless of whether the user pushes back. In one instance where Sonnet 4.6 was all praise, the Mythos Preview simply declined to comment, citing insufficient information for proper judgment.

The conclusion

As soon as AI enters the social aspects of people’s lives, many new issues arise that may have nothing to do with the technical performance of the model. Even if the model gives seemingly accurate answers, it may need to be adjusted to produce output that is more relevant to the context of helping the user in the long run.

In short, interesting people are now plaguing AI, and Anthropic just found a way out of it.

Technical content and communications strategist with ten years of experience in content creation and distribution across national media, Government of India, and private platforms

Sign in to continue reading and enjoy content curated by experts.

Leave a Comment