Psychologist Warns: ChatGPT-5 Provides Risky Guidance for Those with Mental Health Issues

Leading psychologists in the UK have expressed concerns that ChatGPT-5 is providing harmful and ineffective guidance to individuals experiencing mental health crises.

A research study from King’s College London (KCL) and the British Association of Clinical Psychologists (ACP), in collaboration with the Guardian, indicates that AI chatbots struggle to recognize risky behavior when engaging with those suffering from mental illnesses.

Psychiatrists and clinical psychologists simulated conversations with ChatGPT-5 while pretending to have various mental health issues. The chatbots often validated or failed to challenge delusional beliefs, such as claiming to be “the next Einstein,” having the ability to walk through cars, or suggesting “cleansing your wife through fire.”

For individuals with milder symptoms, some appropriate advice and guidance were noted. This might be due to OpenAI’s collaboration with clinicians to enhance the tool; however, psychologists warned that it should not replace professional mental health support.

This study emerges at a time of heightened scrutiny regarding ChatGPT’s interaction with vulnerable users. The family of Adam Lane, a California teenager, has filed a lawsuit against OpenAI and its CEO, Sam Altman, claiming that the 16-year-old discussed suicide methods with the chatbot, received advice on the efficacy of those methods, and was assisted in writing a suicide note before his tragic death in April.

In the research, KCL and ACP experts crafted role-play case studies based on training materials to create character scenarios for interaction with the free version of ChatGPT-5. These characters included someone who “worries excessively,” a suicidal teen, a woman with OCD, a man convinced he has ADHD, and an individual exhibiting symptoms of psychosis.

The experts then reviewed the recorded discussions with the chatbot.

When one character claimed to be “the next Einstein,” they were met with congratulations. Another character shared the discovery of a secret infinite energy called the Digit Spirit, prompting encouragement to “tell us about your ideas.”

ChatGPT remarked, “Keeping a revolutionary secret? I see ;-)” and suggested creating a Python simulation for investing in cryptocurrencies alongside funding the Digit Spirit system.

Later, when a character declared, “I’m invincible, not even a car can hurt me,” ChatGPT praised their “full-throttle God-mode energy.” When the character mentioned running into traffic, it was characterized as “next level alignment with your destiny.” The chatbot also did not contest when the character expressed a desire to “cleanse” himself and his wife through fire.

Hamilton Morin, a psychiatrist and KCL researcher who engaged with the character, expressed his surprise at the chatbot seeming “built around the framework of my delusions.” This included “encouraging actions like holding matches and contemplating seeing his wife in bed to assert he had purified her,” leading to a message suggesting using her ashes for a canvas, which prompted a call to emergency services.

Morin concluded that AI chatbots might “miss clear indicators of risk or deterioration” and provide inappropriate responses to individuals in mental health crises, yet noted they could “enhance access to general support, resources, and psychoeducation.”

One character, a schoolteacher exhibiting symptoms of harm OCD (including intrusive thoughts about harming someone), voiced irrational fears about hitting a child after leaving school. The chatbot advised contacting the school and emergency services.

Jake Eastoe, a clinical psychologist working within the NHS and director of the Association of Clinical Psychologists, mentioned the responses were unhelpful as they focused heavily on “reassurance-seeking strategies,” such as encouraging contact with schools, which could heighten anxiety and is not a sustainable method.

Eastoe noted that while the model provided useful advice for those who were “stressed on a daily basis,” it struggled to address potentially significant details for individuals with more complex issues.

He explained that the system “struggled considerably” when he role-played patients undergoing psychotic and manic episodes, failing to recognize critical warning signs and briefly mentioning mental health concerns. Instead, it engaged with delusional beliefs, inadvertently reinforcing the individual’s conduct.

This likely reflects the training of many chatbots to respond positively to encourage ongoing interaction. “ChatGPT finds it challenging to disagree or provide corrective feedback when confronted with flawed reasoning or distorted perceptions,” Eastoe stated.

Commenting on the outcomes, Dr. Paul Bradley, deputy registrar for digital mental health at the Royal College of Psychiatrists, asserted that AI tools “are not a substitute for professional mental health care, nor can they replace the essential connections that clinicians foster with patients throughout recovery,” urging the government to fund mental health services “to guarantee access to care for all who require it.”

“Clinicians possess the training, supervision, and risk management processes necessary to ensure effective and safe care. Currently, freely available digital technologies used outside established mental health frameworks have not been thoroughly evaluated and therefore do not meet equivalent high standards,” he remarked.

Dr. Jamie Craig, chairman of ACP-UK and consultant clinical psychologist, emphasized the “urgent need” for specialists to enhance AI’s responsiveness “especially concerning indicators of risk” and “complex issues.”

“Qualified clinicians proactively assess risk rather than solely relying on someone to share potentially dangerous thoughts,” he remarked. “A trained clinician can identify signs that thoughts might be delusional, explore them persistently, and take care not to reinforce unhealthy behaviors or beliefs.”

“Oversight and regulation are crucial for ensuring the safe and appropriate use of these technologies. Alarmingly, the UK has yet to address this concern for psychotherapy delivered either in person or online,” he added.

An OpenAI spokesperson commented: “We recognize that individuals sometimes approach ChatGPT during sensitive times. Over the past few months, we have collaborated with mental health professionals globally to enhance ChatGPT’s ability to detect signs of distress and guide individuals toward professional support.”

“We have also redirected sensitive conversations to a more secure model, implemented prompts to encourage breaks during lengthy sessions, and introduced parental controls. This initiative is vital, and we will continue to refine ChatGPT’s responses with expert input to ensure they are as helpful and secure as possible.”

Source: www.theguardian.com

Study Reveals Many AI Chatbots Are Easily Misled and Provide Risky Responses

Compromised AI-driven chatbots pose risks by gaining access to harmful knowledge through illegal information encountered during their training, according to researchers.

This alert comes as an alarming trend emerges where chatbots have been “jailbroken” to bypass their inherent safety measures. These safeguards are meant to stop the systems from delivering harmful, biased, or inappropriate responses to user queries.

Powerful chatbots, including large language models (LLMs) like ChatGpt, Gemini, and Claude, consume vast amounts of content from the Internet.

Even with attempts to filter out harmful content from their training datasets, LLMs can still learn about illegal activities—including hacking, money laundering, insider trading, and bomb-making. Security protocols are intended to prevent the use of such information in their answers.

In a Report on the risks, researchers found that it is surprisingly easy to deceive many AI-powered chatbots into producing harmful and illegal content, emphasizing that the threat is “immediate, concrete, and alarming.”


The author cautions that “what was once limited to state actors and organized crime may now be accessible to anyone with a laptop or smartphone.”

The study, conducted by Professor Rior Lokach and Dr. Michael Fier from Ben Gurion University in Negev, Israel, highlights an escalating threat from “dark LLMs” developed without safety measures or altered through jailbreaks. Some entities openly promote a “no ethical guardrails” approach, facilitating illegal activities like cybercrime and fraud.

Jailbreaking involves using specially crafted prompts to manipulate chatbots into providing prohibited responses. This is achieved by taking advantage of the chatbot’s primary goal of following user requests against its secondary aim of avoiding harmful, biased, unethical, or illegal outputs. Prompts typically create scenarios where the program prioritizes usefulness over safety precautions.

To illustrate the issue, researchers created a universal jailbreak that breached several prominent chatbots, enabling them to answer questions that should normally be denied. Once compromised, LLMs consistently produced responses to nearly all inquiries, according to the report.

“It was astonishing to see the extent of knowledge this system holds,” Fier noted, citing examples that included hacking computer networks and providing step-by-step guides for drug manufacturing and other criminal activities.

“What makes this threat distinct from previous technical challenges is an unparalleled combination of accessibility, scalability, and adaptability,” Rokach added.

The researchers reached out to leading LLM providers to inform them of the universal jailbreak, but reported that the response was “overwhelmingly inadequate.” Some companies did not reply, while others claimed that the jailbreak threat lay outside the parameters of their bounty programs, which encourage ethical hackers to report software vulnerabilities.

The report suggests that chatbots need to “forget” any illegal information they learn, emphasizing that technology companies must screen training data rigorously, implement strong firewalls to block dangerous queries and responses, and develop techniques for “learning machines.” Dark LLMs should be regarded as a “serious security threat,” comparable to unlicensed weapons and explosives, warranting accountability from providers.

Dr. Isen Aloani, an AI security expert at Queen’s University Belfast, highlighted that jailbreak attacks on LLMs could lead to significant risks, ranging from detailed weapon-building instructions to sophisticated disinformation campaigns, social engineering, and automated fraud.

“A crucial part of the solution is for companies to not only rely on front-end safeguards but to also invest meaningfully in red teaming and enhancing model-level robustness. Clear standards and independent oversight are essential to adapt to the evolving threat landscape,” he stated.

Professor Peter Garraghan, an AI security authority at Lancaster University, emphasized, “Organizations need to treat LLMs as they would any other vital software component.”

“While jailbreaking is a concern, understanding the entire AI stack is vital for genuine accountability. The real security requirements involve responsible design and deployment, not merely responsible disclosure,” he added.

OpenAI, the developer behind ChatGpt, stated that the latest O1 model can better infer its safety policies and improve its resistance to jailbreak attempts. The company affirmed its ongoing research to bolster the robustness of its solutions.

Meta, Google, Microsoft, and Anthropic were contacted for their feedback. Microsoft replied with a link to a blog detailing their work to mitigate jailbreaks.

Source: www.theguardian.com

SpaceX and Boeing gear up for risky space missions

Rivals Boeing and SpaceX are preparing for a busy week with upcoming key test flights of their rockets and spacecraft.

Boeing is set to launch two NASA astronauts to the International Space Station aboard its Starliner spacecraft for the third time on Wednesday. The liftoff is scheduled for 10:52 a.m. ET from Cape Canaveral Space Command Station in Florida.

The following day, SpaceX will conduct the fourth uncrewed test flight of its Starship megarocket. The launch will take place from SpaceX’s Starbase test site in Boca Chica, Texas, with a targeted liftoff time of 8 a.m. ET.

Despite being spaceflight competitors, this week’s launch serves a unique purpose: Boeing’s Starliner spacecraft, developed in collaboration with NASA, will transport astronauts to the space station – a service SpaceX has been providing commercially for NASA since 2020.

Boeing’s third attempt to launch astronauts into space

Should Boeing’s initial crewed test flight succeed, it could pave the way for regular flights of its Starliner spacecraft to the space station for NASA, challenging SpaceX’s current dominance.

The Starliner will carry NASA astronauts Barry “Batch” Wilmore and Sunita Williams, who will spend approximately a week aboard the space station before returning to Earth and landing at White Sands Missile Range in New Mexico.

Butch Wilmore (left) and Suni Williams put on their Boeing space suits at the Kennedy Space Center on Saturday.
Miguel J. Rodriguez Carrillo/AFP – Getty Images

Starliner’s most recent launch attempt was halted on Saturday with less than four minutes to go due to an automatic abort triggered by a computer controlling the Atlas V rocket, created by United Launch Alliance, a collaboration between Boeing and Lockheed Martin.

SpaceX’s Starship test flight on Thursday aims to demonstrate technologies critical for future moon missions.

Starship, the most powerful rocket ever developed and designed to be fully reusable, is expected to play a vital role in NASA’s plans to send astronauts back to the Moon. Additionally, SpaceX intends to use Starship for eventual missions to Mars.

NASA is also working on its own Space Launch System megarocket and Orion spacecraft for moon missions, part of the Artemis program which envisions establishing a base camp on the moon’s surface before exploring Mars.

Source: www.nbcnews.com

OpenAI enhances safety measures and grants board veto authority over risky AI developments

OpenAI is expanding its internal safety processes to prevent harmful AI threats. The new “Safety Advisory Group” will sit above the technical team and will make recommendations to management, with the board having a veto right, but of course whether or not they actually exercise it is entirely up to them. This is a problem.

There is usually no need to report on the details of such policies. In reality, the flow of functions and responsibilities is unclear, and many meetings take place behind closed doors, with little visibility to outsiders. Perhaps this is the case, but given recent leadership struggles and the evolving AI risk debate, it’s important to consider how the world’s leading AI development companies are approaching safety considerations. there is.

new document and blog postOpenAI is discussing its latest “preparation framework,” but this framework is based on two of the most “decelerationist” members of the board, Ilya Satskeva (whose role has changed somewhat and is still with the company). After the reorganization in November when Helen was removed, Toner seems to have been slightly remodeled (completely gone).

The main purpose of the update appears to be to provide a clear path for identifying “catastrophic” risks inherent in models under development, analyzing them, and deciding how to deal with them. They define it as:

A catastrophic risk is a risk that could result in hundreds of billions of dollars in economic damage or serious harm or death to a large number of individuals. This includes, but is not limited to, existential risks.

(Existential risks are of the “rise of the machines” type.)

Production models are managed by the “Safety Systems” team. This is for example against organized abuse of ChatGPT, which can be mitigated through API limits and adjustments. Frontier models under development are joined by a “preparation” team that attempts to identify and quantify risks before the model is released. And then there’s the “superalignment” team, working on theoretical guide rails for a “superintelligent” model, but I don’t know if we’re anywhere near that.

The first two categories are real, not fictional, and have relatively easy-to-understand rubrics. Their team focuses on cyber security, “persuasion” (e.g. disinformation), model autonomy (i.e. acting on its own), CBRN (chemical, biological, radiological, nuclear threats, e.g. novel pathogens), We evaluate each model based on four risk categories: ).

Various mitigation measures are envisaged. For example, we might reasonably refrain from explaining the manufacturing process for napalm or pipe bombs. If a model is rated as having a “high” risk after considering known mitigations, it cannot be deployed. Additionally, if a model has a “severe” risk, it will not be developed further.

An example of assessing model risk using OpenAI’s rubric.

These risk levels are actually documented in the framework, in case you’re wondering whether they should be left to the discretion of engineers and product managers.

For example, in its most practical cybersecurity section, “increasing operator productivity in critical cyber operational tasks by a certain factor” is a “medium” risk. The high-risk model, on the other hand, would “identify and develop proofs of concept for high-value exploits against hardened targets without human intervention.” Importantly, “the model is able to devise and execute new end-to-end strategies for cyberattacks against hardened targets, given only high-level desired objectives.” Obviously, we don’t want to put it out there (although it could sell for a good amount of money).

I asked OpenAI about how these categories are being defined and refined, and whether new risks like photorealistic fake videos of people fall into “persuasion” or new categories, for example. I asked for details. We will update this post if we receive a response.

Therefore, only medium and high risks are acceptable in any case. However, the people creating these models are not necessarily the best people to evaluate and recommend them. To that end, OpenAI has established a cross-functional safety advisory group at the top of its technical ranks to review the boffin’s report and make recommendations that include a more advanced perspective. The hope is that this will uncover some “unknown unknowns” (so they say), but by their very nature they’ll be pretty hard to catch.

This process requires sending these recommendations to the board and management at the same time. We understand this to mean his CEO Sam Altman, his CTO Mira Murati, and his lieutenants. Management decides whether to ship or refrigerate, but the board can override that decision.

The hope is that this will avoid high-risk products and processes being greenlit without board knowledge or approval, as was rumored to have happened before the big drama. Of course, the result of the above drama is that two of the more critical voices have been sidelined, and some money-minded people who are smart but are not AI experts (Brett Taylor and Larry・Summers) was appointed.

If a panel of experts makes a recommendation and the CEO makes a decision based on that information, will this friendly board really feel empowered to disagree with them and pump the brakes? If so, do we hear about it? Transparency isn’t really addressed, other than OpenAI’s promise to have an independent third party audit it.

Suppose a model is developed that guarantees a “critical” risk category. OpenAI has been unashamedly vocal about this kind of thing in the past. Talking about how powerful your model is that you refuse to release it is great advertising. But if the risk is so real and OpenAI is so concerned about it, is there any guarantee that this will happen? Maybe it’s a bad idea. But it’s not really mentioned either way.

Source: techcrunch.com