About this Event
1520 Middle Dr, Knoxville, TN 37916
https://trustai-seminar.github.io/2026spring.htmlAbstract:
Generative AI systems increasingly influence human writing, thoughts, and actions, yet our ability to measure and control the behavior of these systems is scant. In this talk, I will describe some ways to measure biases in generative AI systems and what we learned from testing state-of-the-art models. Then, I will advocate for auditing strategies that depend on analysis and manipulation of internal representations, and show how a simple inference-time intervention can be used to mitigate gender bias and control model censorship without degrading overall model utility.
Bio:
David Evans (https://www.cs.virginia.edu/evans/) is the Olsen Bicentennial Professor of Engineering and a Professor of Computer Science at the University of Virginia where he leads research on security and privacy with a recent focus on understanding and mitigating risks associated with artificial intelligence. He is the author of an open computer science textbook, a book on secure computation, and a tragicomic children's book on combinatorics and computability. He was Program Co-Chair for the 24th ACM Conference on Computer and Communications Security (CCS 2017) and the 30th (2009) and 31st (2010) IEEE Symposia on Security and Privacy, where he initiated the Systematization of Knowledge papers. He has SB, SM and PhD degrees in Computer Science from MIT and has been a faculty member at the University of Virginia since 1999.
Co-Sponsors: TRUST-AI Distinguished Seminar Series (FNU Suya, organizer) and Center for Social Theory
0 people are interested in this event