- About
- Events
- Calendar
- Graduation Information
- Cornell Learning Machines Seminar
- Student Colloquium
- BOOM
- Fall 2024 Colloquium
- Conway-Walker Lecture Series
- Salton 2024 Lecture Series
- Seminars / Lectures
- Big Red Hacks
- Cornell University - High School Programming Contests 2024
- Game Design Initiative
- CSMore: The Rising Sophomore Summer Program in Computer Science
- Explore CS Research
- ACSU Research Night
- Cornell Junior Theorists' Workshop 2024
- People
- Courses
- Research
- Undergraduate
- M Eng
- MS
- PhD
- Admissions
- Current Students
- Computer Science Graduate Office Hours
- Advising Guide for Research Students
- Business Card Policy
- Cornell Tech
- Curricular Practical Training
- A & B Exam Scheduling Guidelines
- Fellowship Opportunities
- Field of Computer Science Ph.D. Student Handbook
- Graduate TA Handbook
- Field A Exam Summary Form
- Graduate School Forms
- Instructor / TA Application
- Ph.D. Requirements
- Ph.D. Student Financial Support
- Special Committee Selection
- Travel Funding Opportunities
- Travel Reimbursement Guide
- The Outside Minor Requirement
- Diversity and Inclusion
- Graduation Information
- CS Graduate Minor
- Outreach Opportunities
- Parental Accommodation Policy
- Special Masters
- Student Spotlights
- Contact PhD Office
To the Test Set and Beyond: Generalizing Far From Your Training Data with Compositional Energy Based Models
Abstract: A critical issue when training machine learning models in the embodied environments is a lack of comprehensive data. In my talk, I'll illustrate how by factoring the data distribution into a set of learnable subdistributions, we can learn models that generalize far from the training distribution. I'll first introduce the concept of Energy Based Models, and illustrate the algebra through which these models enable compositionality. I'll then discuss how such compositionality enables generalization far from the training dataset in both vision and robotics domains. Finally, I'll discuss how such operations can be used to construct agents with multiple-sources of commonsense knowledge, by combining different foundation models at prediction time.
Bio: Yilun Du is PhD student at MIT EECS advised by Leslie Kaelbling, Tomas Lozano-Perez, and Joshua Tenenbaum. Previously, he has done a research fellowship at OpenAI, and held visiting researcher/research internship positions at FAIR and Google Deepmind. He is supported by the NSF Graduate Fellowship, and has received outstanding paper awards at NeurIPS and ICRA workshops.