The International Conference on Learning Representations (ICLR) 2023 is being hosted in Kigali, Rwanda from May 1st - May 5th. We’re excited to share all the work from SAIL that’s being presented, and you’ll find links to papers, videos and blogs below. Feel free to reach out to the contact authors directly to learn more about the work that’s happening at Stanford!
List of Accepted Papers
A Control-Centric Benchmark for Video Prediction
Authors: Stephen Tian, Chelsea Finn, Jiajun Wu
Contact: tians@stanford.edu
Links: Paper | Website
Keywords: benchmarking, video prediction, visual mpc, manipulation
Ask Me Anything: A simple strategy for prompting language models
Authors: Simran Arora*, Avanika Narayan*, Mayee F. Chen, Laurel Orr, Neel Guha, Kush Bhatia, Ines Chami, Frederic Sala, Christopher Ré
Contact: avanikan@stanford.edu
Award nominations: Spotlight, top 25% of acceptances
Links: Paper | Website
Keywords: prompting, foundation models
Asymptotic Instance-Optimal Algorithms for Interactive Decision Making
Authors: Kefan Dong, Tengyu Ma
Contact: kefandong@stanford.edu
Links: Paper
Keywords: instance optimality, reinforcement learning theory
Beyond Confidence: Reliable Models Should Also Quantify Atypicality
Authors: Beyond Confidence: Reliable Models Should Also Quantify Atypicality
Contact: merty@stanford.edu
Award nominations: Oral Presentation
Links: Paper
Keywords: trustworthy machine learning, reliable machine learning, uncertainty, calibration
Characterizing Intrinsic compositionality in Transformers with Tree Projections
Authors: Shikhar Murty, Pratyusha Sharma, Jacob Andreas, Christopher D Manning
Contact: smurty@cs.stanford.edu
Links: Paper
Keywords: compositionality, emergent syntax, generalization
Diagnosing and Rectifying Vision Models using Language
Authors: Yuhui Zhang, Jeff Z. HaoChen, Shih-Cheng Huang, Kuan-Chieh Wang, James Zou, Serena Yeung
Contact: yuhuiz@stanford.edu
Links: Paper | Video | Website
Keywords: model diagnosis, multi-modal contrastive learning, vision and language
Extreme Q-Learning: MaxEnt RL Without Entropy
Authors: Div Garg*, Joey Hejna*, Matthieu Geist, Stefano Ermon
Contact: jhejna@cs.stanford.edu
Award nominations: notable top 5%
Links: Paper | Website
Keywords: reinforcement learning, offline reinforcement learning, statistical learning, extreme value analysis, maximum entropy rl, gumbel
First Steps Toward Understanding the Extrapolation of Nonlinear Models to Unseen Domains
Authors: Kefan Dong, Tengyu Ma
Contact: kefandong@stanford.edu
Links: Paper
Keywords: extrapolation of nonlinear models, theory, structured domain shift, gaussian kernel
FlexVDW: A machine learning approach to account for protein flexibility in ligand docking
Authors: Patricia Suriana, Joseph M. Paggi, Ron O. Dror
Contact: psuriana@stanford.edu
Links: Paper
Keywords: deep learning, structural biology, protein ligand docking, protein flexibility
Hungry Hungry Hippos: Towards Language Modeling with State Space Models
Authors: Daniel Y. Fu*, Tri Dao*, Khaled K. Saab, Armin W. Thomas, Atri Rudra, Christopher Ré
Contact: danfu@cs.stanford.edu
Award nominations: Notable top-25% (spotlight)
Links: Paper | Blog Post | Website
Keywords: language modeling, state space models, convolution, fft, io-aware
Is a Caption Worth a Thousand Images? A Controlled Study for Representation Learning
Authors: Shibani Santurkar, Yann Dubois, Rohan Taori, Percy Liang, Tatsunori B Hashimoto
Contact: rtaori@stanford.edu
Links: Paper
Keywords: clip, transfer learning, contrastive learning, multi-modal
Learning Controllable Adaptive Simulation for Multi-resolution Physics
Authors: Tailin Wu, Takashi Maruyama, Qingqing Zhao, Gordon Wetzstein, Jure Leskovec
Contact: tailin@cs.stanford.edu
Award nominations: notable-top-25% (spotlight)
Links: Paper | Website
Keywords: learned simulation, adaptive, multi-scale, error vs. computation, controllable
Parameter-Efficient Fine-Tuning Design Spaces
Authors: Jiaao Chen, Aston Zhang, Xingjian Shi, Mu Li, Alex Smola, Diyi Yang
Contact: jchen896@gatech.edu
Links: Paper
Keywords: parameter-efficient fine-tuning, design spaces
Pitfalls of Gaussians as a noise distribution in NCE
Authors: Holden Lee, Chirag Pabbaraju, Anish Sevekari, Andrej Risteski
Contact: cpabbara@stanford.edu
Links: Paper
Keywords: nce, noise contrastive estimation, generative models, statistical efficiency, theory
Post-hoc Concept Bottleneck Models
Authors: Mert Yuksekgonul, Maggie Wang, James Zou
Contact: merty@stanford.edu
Award nominations: Spotlight (Top 25%)
Links: Paper
Keywords: concepts, interpretability, concept bottleneck models, model editing
Reward Design with Language Models
Authors: Minae Kwon, Sang Michael Xie, Kalesha Bullard, Dorsa Sadigh
Contact: minae@cs.stanford.edu
Links: Paper | Video
Keywords: alignment, reinforcement learning, foundation models, reward design
Simplified State Space Layers for Sequence Modeling
Authors: Jimmy T.H. Smith*, Andrew Warrington*, Scott W. Linderman
Contact: jsmith14@stanford.edu
Award nominations: Notable-top-5% (In-person Oral Presentation)
Links: Paper | Website
Keywords: deep learning, sequence model, state space model, s4
Surgical Fine-Tuning Improves Adaptation to Distribution Shifts
Authors: Yoonho Lee*, Annie S. Chen*, Fahim Tajwar, Ananya Kumar, Huaxiu Yao, Percy Liang, Chelsea Finn
Contact: asc8@stanford.edu
Links: Paper
Keywords: transfer learning, fine-tuning, parameter freezing, distortion of pre-trained models
The Asymmetric Maximum Margin Bias of Quasi-Homogeneous Neural Networks
Authors: Daniel Kunin, Atsushi Yamamura, Chao Ma, Surya Ganguli
Contact: kunin@stanford.edu, atsushi3@stanford.edu
Award nominations: ICLR 2023 notable top 25%
Links: Paper
Keywords: margin, maximum-margin, implicit regularization, neural networks, neural collapse, gradient flow, implicit bias, robustness, homogeneous, symmetry, classification
When and why vision-language models behave like bags-of-words, and what to do about it?
Authors: Mert Yuksekgonul, Federico Bianchi, Pratyusha Kalluri, Dan Jurafsky, James Zou
Contact: merty@stanford.edu
Award nominations: Oral (Top 5%)
Links: Paper | Blog Post
Keywords: vision-language models, clip, contrastive learning, retrieval, vision-language pretraining, multimodal representation learning
We look forward to seeing you at ICLR!