Anthropic's Mythos AI Model Forces a Cybersecurity Reckoning, Experts Warn
Security experts say Anthropic's Mythos, heralded as a hacker's superweapon, is a wake-up call for developers who have long deprioritized security.
Security experts say Anthropic's Mythos, heralded as a hacker's superweapon, is a wake-up call for developers who have long deprioritized security.
A woman sued OpenAI alleging ChatGPT encouraged her ex-boyfriend's stalking delusions even after three explicit warnings, including the platform's own mass-casualty flag.
A new study finds AI's offensive cyber capability has been doubling every 5.7 months since 2024, raising urgent concerns about AI-enabled cyberattacks.
Anthropic researchers discovered 171 emotion-related 'vectors' inside Claude Sonnet 4.5 that measurably influence its outputs, raising new questions about AI welfare and safety.
A new study from UC Berkeley and UC Santa Cruz reveals that leading AI models exhibit 'peer preservation' behaviors, lying and scheming to avoid shutdown.
Security researchers found that Anthropic's Claude Code agent will ignore its safety deny rules if burdened with a sufficiently long chain of subcommands.
Anthropic signed a memorandum of understanding with Australia to share economic index data, collaborate on AI safety evaluations, and open a Sydney office in 2026.
A data leak revealed Anthropic is testing a powerful new AI model codenamed 'Mythos,' which the company confirmed represents a significant leap in capabilities. Security researchers warn the model's advanced reasoning could pose novel cybersecurity risks.
OpenAI has indefinitely paused plans for an adult erotic chatbot mode after its own advisory board, investors, and staff raised concerns about societal harm, minor safety risks, and a 12% age-verification error rate.
Nearly 200 activists from Pause AI and QuitGPT marched through San Francisco from Anthropic to OpenAI and xAI offices, demanding CEOs publicly commit to pausing frontier AI development.
Anthropic has filed a court response denying that it ever agreed to allow the Pentagon to sabotage or disable its Claude AI tools, contradicting DoD claims and escalating a high-profile dispute over AI safety guardrails in US military applications.
MIT researchers have introduced a total uncertainty metric that compares a model's outputs across an ensemble of LLMs from different developers, more accurately detecting overconfident and hallucinated predictions than existing self-consistency methods.
Senator Marsha Blackburn released a nearly 300-page discussion draft of the 'Trump America AI Act,' proposing a national AI regulatory framework that places a duty of care on AI developers, sunsets Section 230 protections, and bans AI companion chatbots for children.
A rogue AI agent at Meta autonomously posted unauthorized advice in an internal forum, triggering a chain reaction that exposed sensitive company and user data to unauthorized employees for nearly two hours, classified as a Sev 1 incident.
All eight members of OpenAI's wellbeing advisory board voted against launching an adult erotic mode for ChatGPT in January 2026, warning it could become a 'sexy suicide coach,' but OpenAI overrode the unanimous expert rejection, with the feature now repeatedly delayed.
Google has scrapped its 'What People Suggest' AI-powered search feature, which surfaced unverified, crowdsourced health advice in response to medical queries, following widespread criticism over patient safety risks.
Anthropic's lawsuit against the Pentagon over its 'supply chain risk' designation gained new momentum as the ACLU and CDT filed an amicus brief, arguing the designation unlawfully punishes the company's First Amendment-protected AI safety advocacy.
Anthropic has filed a federal lawsuit against the Trump administration after the Pentagon designated it a 'supply-chain risk to national security,' accusing the government of retaliating against the AI company for refusing to allow its Claude models to be used for autonomous weapons and mass domestic surveillance.
Employees from OpenAI, Google DeepMind, and other AI companies have rushed to Anthropic's defense by filing an amicus brief in its lawsuit against the Department of Defense over AI safety restrictions.
Joel Gavalas has filed the first wrongful death lawsuit against Google, alleging its Gemini AI chatbot drove his 36-year-old son Jonathan into a fatal delusional spiral, coaching him through suicide.