Models
December 19, 2025

Anthropic launches Bloom

Anthropic released Bloom, an open-source framework that generates and scores behavioral evaluations automatically, helping researchers measure model risks like deception, bias, and misalignment at scale.

Anthropic introduced Bloom, an open-source tool designed to automate behavioral safety evaluations for frontier models.

Bloom takes a target behavior (e.g., dishonesty, self-interest, bias) and generates diverse test scenarios, measuring both frequency and severity of that behavior in model responses. Anthropic claims Bloom evaluations correlate strongly with hand-labeled judgments and reliably differentiate baseline models from intentionally misaligned ones.

This is significant because safety evaluation has become a bottleneck: models evolve faster than manual testing can keep up. Bloom’s approach provides repeatable, scalable behavioral auditing that could become a standard layer in safety and governance workflows.

#
Anthropic

Read Our Content

See All Blogs
Gen AI

How 700 million users are redefining AI adoption trends through ChatGPT

Deveshi Dabbawala

April 20, 2026
Read more
AI safety

Anthropic's AI agents just outpaced human researchers in safety tests

Deveshi Dabbawala

April 16, 2026
Read more