Illuminating Generative AI: Mapping Knowledge in Large Language Models
                                                    Abhilasha Ravichander
                                                Max Planck Institute for Software Systems 
			
                    
                
                03 Dec 2025, 12:15 pm - 1:15 pm            
            Saarbrücken building E1 5, room 002
            Joint Lecture Series
		Millions of everyday users are interacting with technologies built 
with generative AI, such as voice assistants, search engines, and chatbots. 
While these AI-based systems are being increasingly integrated into modern 
life, they can also magnify risks, inequities, and dissatisfaction when 
providers deploy unreliable systems. A primary obstacle to having more reliable 
systems is the opacity of the underlying large language models— we lack a 
systematic understanding of how models work, where critical vulnerabilities may 
arise, why they are happening, ...
                Millions of everyday users are interacting with technologies built 
with generative AI, such as voice assistants, search engines, and chatbots. 
While these AI-based systems are being increasingly integrated into modern 
life, they can also magnify risks, inequities, and dissatisfaction when 
providers deploy unreliable systems. A primary obstacle to having more reliable 
systems is the opacity of the underlying large language models— we lack a 
systematic understanding of how models work, where critical vulnerabilities may 
arise, why they are happening, and how models must be redesigned to address 
them. In this talk, I will first describe my work in investigating large 
language models to illuminate when models acquire knowledge and capabilities. 
Then, I will describe my work on building methods to enable data transparency 
for large language models, that allows practitioners to make sense of the 
information available to models. Finally, I will describe work on understanding 
why large language models produce incorrect knowledge, and implications for 
building the next generation of responsible AI systems. 
                                        Read more