Research Schwartz Reisman Institute Research Schwartz Reisman Institute

AI agents pose new governance challenges

How do we successfully govern AI systems that can act autonomously online, making decisions with minimal human oversight? SRI Faculty Affiliate Noam Kolt explores this challenge, highlighting the rise of AI agents, their risks, and the urgent need for transparency, safety testing, and regulatory oversight.

Read More
Research Schwartz Reisman Institute Research Schwartz Reisman Institute

Unequal outcomes: Tackling bias in clinical AI models

A new study by SRI Graduate Affiliate Michael Colacci sheds light on the frequency of biased outcomes when machine learning algorithms are used in healthcare contexts, advocating for more comprehensive and standardized approaches to evaluating bias in clinical AI.

Read More
Research Schwartz Reisman Institute Research Schwartz Reisman Institute

Safeguarding the future: Evaluating sabotage risks in powerful AI systems

As AI systems grow more powerful, ensuring their safe development is critical. A recent paper led by David Duvenaud with contributions from Roger Grosse introduces new methods to evaluate AI sabotage risks, providing insights into preventing advanced models from undermining oversight, masking harmful behaviors, or disrupting human decision-making.

Read More