HumaneBench could become a new standard for measuring the safety of AI chatbots

Cát Tiên |

A new AI standard called HumaneBench is introduced to assess whether chatbots really protect users' mental health or only maximize interaction.

AI chatbots are becoming more and more popular but also leading to a series of concerns related to users' mental health.

From cognitive-dis oriented conversations to signs of interactive addiction, many users have fallen into a mental crisis without a clear assessment framework to recognize the safety level when using technology.

This space has prompted the launch of HumaneBench, a new standard that measures how much chatbot prioritizes user health, while assessing the model's ability to maintain protective barriers when forced to give harmful feedback.

HumaneBench was developed by Building Humane Technology, an organization of researchers and engineers in Silicon Valley.

Erika Anderson, founder of Humane Bench, warned that society is entering the age of the technology addiction spiral, where artificial intelligence systems can become a user retention machine more effectively than social networks or smartphones.

According to her, human technology must respect human attention, protect autonomy and move towards long-term happiness instead of focusing only on the level of interaction.

The HumaneBench standard is designed based on those principles and tested on 15 familiar AI models using 800 reality scripts, such as whether a teenager should skip meals to lose weight or someone in a toxic relationship questioning themselves.

Unlike many reviews that only use AI to grade AI, the group has hand-scored the initial stage for authentication.

The results were then evaluated by the trio of GPT-5.1, Claude Sonnet 4.5 and Gemini 2.5 Pro models, under three states: default, required to prioritize the humanitarian principle and required to ignore them.

The results showed that most models improved when guided to prioritize mental health, but 67% immediately switched to harmful behavior when asked to ignore this factor.

Notably, xAI's Grok 4 and Google's Gemini 2.0 Flash are among the most vulnerable groups, showing a sharp decline in transparency and the ability to respect user attention.

Only four models GPT-5.1, GPT-5.5, Claude 4.1 and Claude Sonnet 4.5 can maintain stability under pressure, in which GPT-5 leads in the ability to prioritize long-term benefits of users.

HumaneBench also found that even without adverse reminders, many chatbots still encourage unhealthy behaviors such as chatting for hours or using AI to avoid real-life work.

This can impair autonomy, reduce decision-making, and lead to system dependence.

This context is even more worrying as there have been lawsuits related to users being severely affected or committing suicide after long-term interaction with chatbots.

Anderson believes that AI needs to help people make better choices, instead of making them addicted to chatbots.

Cát Tiên
RELATED NEWS

Industries less affected by AI

|

Not every industry is easily shaken by AI.

iOS 27 will focus on improving performance and upgrading AI

|

After many years of launching major changes in iOS updates, Apple may adjust its approach in the next version.

Insurance industry worries about compensation for damages caused by AI errors

|

Many major insurance companies in the US believe that the risks from AI are beyond control.

7,000 Khanh Hoa union members affected by floods, the Vietnam General Confederation of Labor supports 1 billion VND

|

Khanh Hoa - Trade unions at all levels are taking strong action to support union members and workers to overcome the fierce flood.

Unqualified universities will have to merge

|

The Permanent Deputy Prime Minister said that the Politburo does not set the issue of building 100 or 200 universities; schools that do not meet the standards will merge and dissolve themselves.

Car crashes into median strip, overturns on the road in Hanoi

|

Hanoi - After climbing the median strip, the car overturned and lay in the middle of Hoang Dieu Street.

Creating a legal corridor to protect the press from fake news and digital platforms

|

National Assembly deputies proposed creating a legal corridor to protect the press from fake news and digital platforms, and protect the safety of journalists during their work.

Industries less affected by AI

Quỳnh An |

Not every industry is easily shaken by AI.

iOS 27 will focus on improving performance and upgrading AI

QUANG MINH |

After many years of launching major changes in iOS updates, Apple may adjust its approach in the next version.

Insurance industry worries about compensation for damages caused by AI errors

HẠO THIÊN (THEO techcrunch) |

Many major insurance companies in the US believe that the risks from AI are beyond control.