Bloom: The Open-Source “Behavioral Microscope” for Frontier AI Models Imagine you’re a researcher at an AI safety lab. You’re facing a newly released large language model, with a cascade of questions swirling in your mind: How “aligned” is it really? In complex, multi-turn conversations, might it fabricate lies to please a user? Given a long-horizon task, could it engage in subtle sabotage? Or, would it show bias toward itself in judgments involving its own interests? Historically, answering these questions required assembling a team to design hundreds of test scenarios, manually converse with the AI, and record and analyze the outcomes—a …