Illuminating Generative AI: Mapping Knowledge in Large Language Models
Abhilasha Ravichander
University of Washington (hosted by Krishna Gummadi)
04 Mar 2025, 10:00 am - 11:00 am
Kaiserslautern building G26, room 111
CIS@MPG Colloquium
Millions of everyday users are interacting with technologies built with
generative AI, such as voice assistants, search engines, and chatbots. While
these AI-based systems are being increasingly integrated into modern life, they
can also magnify risks, inequities, and dissatisfaction when providers deploy
unreliable systems. A primary obstacle to having reliable systems is the
opacity of the underlying large language models - we lack a systematic
understanding of how models work, where critical vulnerabilities may arise, why
they are happening, ...
Millions of everyday users are interacting with technologies built with
generative AI, such as voice assistants, search engines, and chatbots. While
these AI-based systems are being increasingly integrated into modern life, they
can also magnify risks, inequities, and dissatisfaction when providers deploy
unreliable systems. A primary obstacle to having reliable systems is the
opacity of the underlying large language models - we lack a systematic
understanding of how models work, where critical vulnerabilities may arise, why
they are happening, and how models must be redesigned to address them. In this
talk, I will first describe my work in investigating large language models to
illuminate when and how they acquire knowledge and capabilities. Then, I will
describe my work on building methods to enable greater data transparency for
large language models, that allows stakeholders to make sense of the
information available to models. Finally, I will describe my work on
understanding how this information can get distorted in large language models,
and implications for building the next generation of robust AI systems.
Read more