Weijia Shi (UW): Computer Science and DSI Joint Colloquium
Please join us for CS and DSI joint colloquium.
Monday, February 9
2:30pm – 3:30pm
Data Science Institute, Room 105
5460 S University Ave, Chicago, IL 60615
Abstract: Language models (LMs) are typically monolithic: a single model storing all knowledge and serving every use case. This design presents significant challenges; they often generate factually incorrect statements, require costly retraining to add or remove information, and face serious privacy and copyright issues. In this talk, I will discuss how to break this monolith by introducing modular architectures and training algorithms that separate capabilities across composable components. I’ll cover two forms of modularity: (1) External modularity, which augments LMs with external tools like retrievers to improve factuality and reasoning; and (2) internal modularity, which builds inherently modular LMs from decentrally trained components to enable flexible composition and an unprecedented level of control.
Bio: Weijia Shi is a Ph.D. candidate at the University of Washington, where she is advised by Luke Zettlemoyer and Noah Smith. Her research focuses on developing augmented and modular architectures and training algorithms that enable language models to be more controllable, collaboratively developed, and factual. She received an Outstanding Paper Award at ACL 2024 and was recognized as Rising Stars in Machine Learning in 2023 and in Data Science in 2024.
Marco Biroli (UChicago): Schmidt AI in Science Speaker Series
Jiaqi Zhang (MIT): Statistics and DSI Joint Colloquium
Mateo Díaz (Johns Hopkins): Statistics and DSI Joint Colloquium