Loading Events

« All Events

ASSET Seminar: “Demystifying the Inner Workings of Language Models”

February 19 at 12:00 PM - 1:15 PM

Abstract:

Large language models (LLMs) power a rapidly-growing and increasingly impactful suite of AI technologies. However, due to their scale and complexity, we lack a fundamental scientific understanding of much of LLMs’ behavior, even when they are open source. The “black-box” nature of LMs not only complicates model debugging and evaluation, but also limits trust and usability. In this talk, I will describe how my research on interpretability (i.e., understanding models’ inner workings) has answered key scientific questions about how models operate. I will then demonstrate how deeper insights into LLMs’ behavior enable both 1) targeted performance improvements and 2) the production of transparent, trustworthy explanations for human users.

Zoom Link (if unable to attend in-person): https://upenn.zoom.us/j/99113576305

Sarah Wiegreffe

Postdoctoral Researcher

Sarah Wiegreffe is a postdoctoral researcher at the Allen Institute for AI (Ai2) and the Allen School of Computer Science and Engineering at the University of Washington. She has worked on the explainability and interpretability of neural networks for NLP since 2017, with a focus on understanding how language models make predictions in order to make them more transparent to human users. She has been honored as a 3-time Rising Star in EECS, Machine Learning, and Generative AI. She received her PhD in computer science from Georgia Tech in 2022, during which time she interned at Google and Ai2 and won the Ai2 outstanding intern award. She frequently serves on conference program committees, receiving outstanding area chair awards at ACL 2023 and EMNLP 2024.

Details

Date:
February 19
Time:
12:00 PM - 1:15 PM
Event Tags:
, ,

Venue

Amy Gutmann Hall, Room 414
3333 Chestnut Street
Philadelphia, 19104 United States
+ Google Map