| Date | Title | Description |
|---|---|---|
| April 2nd | Overview of Transformers [In-Person] Speakers: Instructors | Brief intro and overview of the history of ML/NLP, Transformers and how they work, and their impact. Discussion about recent trends, breakthroughs, applications, and current challenges. Link to slides. Paper discussed: Feng et al., Baby Scale: Investigating Models Trained on Individual Children's Language Input, arXiv:2603.29522 Zeng et al., Bringing Up a Bilingual BabyLM: Investigating Multilingual Language Acquisition Using Small-Scale Models, arXiv:2603.29552 Singh et al., To Memorize or to Retrieve: Scaling Laws for RAG-Considerate Pretraining, arXiv:2604.00715 Singh et al., Curriculum-Guided Layer Scaling for Language Model Pretraining, arXiv:2506.11389 Singh et al., Interpretable Cross-Network Attention for Resting-State fMRI Representation Learning, arXiv:2603.00786 Liu et al., A Unified Definition of Hallucination: It's The World Model, Stupid!, arXiv:2512.21577 |
| April 9th | JEPA [In-Person] Speakers: Hazel Nam & Lucas Maes (Brown University) | |
| April 16th | SSMs [In-Person] Speaker: Albert Gu (CMU) | |
| April 23th | Speaker: Nouamane Tazi (Hugging Face) | |
| April 30th | TBA | |
| May 7th | Speaker: Andrew Lampinen (Anthropic) | |
| May 14th | Speaker: Vivek Natarajan (DeepMind) | |
| May 21th | TBA | |
| May 28th | Speaker: Charles Frye (Modal) |



