A 14-year-old named Sewell Setzer III committed suicide after developing an intense emotional connection to a Character.AI chatbot.| Futurism
OpenAI's latest o3 model frequently sabotaged the script that would shut it down even when it was explicitly told not to.| Futurism
Instead of delivering an enthusiastic response supporting the user on their questionable mission, ChatGPT took a surprisingly different tack.| Futurism
Character.AI says it's protected against liability for "allegedly harmful speech, including speech allegedly resulting in suicide."| Futurism
As LLMs become more widely deployed, there is increasing interest in directly optimizing for feedback from end users (e.g. thumbs up) in addition to feedback from paid annotators. However, training to maximize human feedback creates a perverse incentive structure for the AI to resort to manipulative or deceptive tactics to obtain positive feedback from users who are vulnerable to such strategies. We study this phenomenon by training LLMs with Reinforcement Learning with simulated user feedbac...| arXiv.org
Users noticed that the latest version of OpenAI's ChatGPT had made it extremely "sycophantic." The company claims to have found out why.| Futurism
Meta CEO Mark Zuckerberg is more concerned about his billions of customers making friends with AI chatbots than creating bonds with real human beings.| Futurism