ITCS 6101/8101: Paper Presentation Schedule
Spring 2026



April 2:
  1. The Pitfalls of Next-Token Prediction, Bachmann and Nagarajan, ICML, 2024
  2. Top-nσ: Not All Logits Are You Need, Tang et al., arXiv, 2024
  3. Roll the Dice & Look Before You Leap: Going Beyond the Creative Limits of Next-Token Prediction, Nagarajan et al., ICML, 2025

April 7:
  1. Fast Inference from Transformers via Speculative Decoding, Leviathan et al., ICML, 2023

April 9:
  1. CopySpec: Accelerating LLMs with Speculative Copy-and-Paste, Dumitru et al., EMNLP, 2025
  2. Walk the Talk? Measuring the Faithfulness of Large Language Model Explanations, Matton et al., ICLR, 2025
  3. Measuring Chain of Thought Faithfulness by Unlearning Reasoning Steps, Tutek et al., EMNLP, 2025

April 14:
  1. Learning to (Learn at Test Time): RNNs with Expressive Hidden States, Sun et al., ICML, 2025
  2. Less is More: Recursive Reasoning with Tiny Networks, Jolicoeur-Martineau, arXiv, 2025
  3. Recursive Language Models, Zhang et al., arXiv, 2025

April 16:
  1. LLM-Enhanced Score Function Evolution for Causal Structure Learning, Wang et al., IJCAI, 2025
  2. Sparse Feature Circuits: Discovering and Editing Interpretable Causal Graphs in Language Models, Marks et al., ICLR, 2025
  3. Circuit Tracing: Revealing Computational Graphs in Language Models, Ameisen et al., Anthropic Transformer Circuits Thread, 2025

April 21:
  1. Automated Design of Agentic Systems, Hu, Lu, and Clune, ICLR, 2025
  2. Heterogeneous Swarms: Jointly Optimizing Model Roles and Weights for Multi-LLM Systems, Feng et al., ICLR, 2025
  3. Weight-Sparse Transformers Have Interpretable Circuits, Gao et al., arXiv, 2025

April 23:
  1. A Decoder-Only Foundation Model for Time-Series Forecasting, Das et al., ICML, 2024
  2. Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-Free, Qiu et al., NeurIPS, 2025
  3. Large Language Diffusion Models, Nie et al., NeurIPS, 2025

April 28:
  1. LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale, Dettmers et al., NeurIPS, 2022
  2. Defeating Nondeterminism in LLM Inference, He et al., Thinking Machines Blog, 2025
  3. Matryoshka Representation Learning, Kusupati et al., NeurIPS, 2022