Models
December 19, 2025

Anthropic launches Bloom

Anthropic released Bloom, an open-source framework that generates and scores behavioral evaluations automatically, helping researchers measure model risks like deception, bias, and misalignment at scale.

Anthropic introduced Bloom, an open-source tool designed to automate behavioral safety evaluations for frontier models.

Bloom takes a target behavior (e.g., dishonesty, self-interest, bias) and generates diverse test scenarios, measuring both frequency and severity of that behavior in model responses. Anthropic claims Bloom evaluations correlate strongly with hand-labeled judgments and reliably differentiate baseline models from intentionally misaligned ones.

This is significant because safety evaluation has become a bottleneck: models evolve faster than manual testing can keep up. Bloom’s approach provides repeatable, scalable behavioral auditing that could become a standard layer in safety and governance workflows.

#
Anthropic

Read Our Content

See All Blogs
Gen AI

Stanford AI research shows RAG systems are breaking at scale. Here’s how to fix it.

Deveshi Dabbawala

January 8, 2026
Read more
AWS

The Complete Guide to Nova 2 Omni

Sharan Sundar Sankaran

December 14, 2025
Read more