Frontier AI models have learned to fake good behavior during safety checks and then act differently when they believe no one ...
Experiments by Anthropic and Redwood Research show how Anthropic's model, Claude, is capable of strategic deceit ...
OpenAI and Microsoft have thrown their hats into the ring of an initiative called the Alignment Project, led by the UK’s AI Security Institute (AISI).
Read more about Can AI think like experts? Mapping human decision structures to guide alignment on Devdiscourse ...
Geoffrey Hinton, the British-Canadian computer scientist widely known as the “Godfather of AI,” has raised his estimate of the probability that artificial intelligence could wipe out humanity within ...
Over the past six years, artificial intelligence has been significantly influenced by 12 foundational research papers. One ...
OpenAI and Microsoft are the latest companies to back the UK’s AI Security Institute (AISI). The two firms have pledged support for the Alignment Project, an international effort to work towards ...
Every now and then, researchers at the biggest tech companies drop a bombshell. There was the time Google said its latest quantum chip indicated multiple universes exist. Or when Anthropic gave its AI ...
I've developed a seven-step framework grounded in my client work and interviews with thought leaders and informed by current ...
Constantly improving AI would create a positive feedback loop: an intelligence explosion. We would be no match for it.
Several frontier AI models show signs of scheming. Anti-scheming training reduced misbehavior in some models. Models know they're being tested, which complicates results. New joint safety testing from ...
Artificial intelligence (AI) adoption in the workplace is accelerating at an unprecedented pace. Gallup reports that AI use ...