
The International Conference on Learning Representations (ICLR) 2026 is being hosted in Rio de Janeiro from April 23rd – April 27th. We’re excited to share all the work from SAIL being presented, with links to papers, websites, blog posts, and videos below. Feel free to reach out to the contact authors directly to learn more about the work happening at Stanford!
List of Accepted Papers
AbdCTBench: Learning Clinical Biomarker Representations from Abdominal Surface Geometry
Authors: Muhammad Ahmed Chaudhry, Suhana Bedi, Pola Lydia Lagari, Brian T Layden, William Galanter, Ayis Pyrros, Sanmi Koyejo
Contact: mahmedch@stanford.edu
Links: Paper | Website
Keywords: computer vision for healthcare, radiology, computed tomography (CT), vision transformers, CNNs
AccelOpt: A Self-Improving LLM Agentic System for AI Accelerator Kernel Optimization
Authors: Genghan Zhang, Shaowei Zhu, Anjiang Wei, Zhenyu Song, Allen Nie, Zhen Jia, Nandita Vijaykumar, Yida Wang, Kunle Olukotun
Contact: zgh23@stanford.edu
Award nominations: ICLR 2026 MALGAI workshop (Oral)
Links: Paper | Website | Blog Post
Keywords: self-improving LLM agents, kernel optimization, AI accelerator
Addressing Divergent Representations from Causal Interventions on Neural Networks
Authors: Satchel Grant, Simon Jerome Han, Alexa R. Tartaglini, Christopher Potts
Contact: grantsrb@stanford.edu
Award nominations: ICLR 2026 Oral Presentation
Links: Paper | Website
Keywords: causal interventions, computational pathways, DAS, patching
An Information Theoretic Perspective on Agentic System Design
Authors: Shizhe He, Avanika Narayan, Ishan S. Khare, Scott W. Linderman, Christopher Ré, Dan Biderman
Contact: shizhehe@stanford.edu
Links: Paper | Blog Post
Keywords: information bottleneck, rate-distortion theory, agentic collaboration, large language models, scaling laws
ATLAS: Adaptive Transfer Scaling Laws for Multilingual Pretraining, Finetuning, and Decoding the Curse of Multilinguality
Authors: Shayne Longpre, Sneha Kudugunta, Niklas Muennighoff, I Hsu, Isaac Caswell, Alex Pentland, Sercan Arik, Chen-Yu Lee, Sayna Ebrahimi
Contact: slongpre@media.mit.edu
Links: Paper
Keywords: scaling laws, large language models
AutoMetrics: Approximate Human Judgements with Automatically Generated Evaluators
Authors: Michael J. Ryan, Yanzhe Zhang, Amol Salunkhe, Yi Chu, Di Xu, Diyi Yang
Contact: mryan0@stanford.edu
Links: Paper | Website
Keywords: evaluation, metrics, data-constrained, human-centered, LLM-as-a-judge
Blending Training and Deployment Data with Weighted Expert Ensembles for Post-hoc LLM Calibration
Authors: Aishwarya Mandyam, Wenhui Sophia Lu, Wing Hung Wong, John Duchi, Barbara E. Engelhardt
Contact: am2@stanford.edu
Links: Paper
Keywords: calibration, distribution shift
Can Large Language Models Match the Conclusions of Systematic Reviews?
Authors: Christopher Polzak, Alejandro Lozano, Min Woo Sun, James Burgess, Yuhui Zhang, Kevin Wu, Chia-Chun Chiang, Jeffrey J Nirschl, Serena Yeung-Levy
Contact: clcp@cs.stanford.edu
Links: Paper | Website
Keywords: benchmarks, multi-document reasoning, medical AI
Causal Interpretation of Neural Network Computations with Contribution Decomposition
Authors: Joshua Brendan Melander, Zaki Alaoui, Shenghua Liu, Surya Ganguli, Stephen A. Baccus
Contact: sliu24@stanford.edu
Links: Paper
Keywords: mechanistic interpretability, neuroscience, XAI, AI safety, tool development
Collaborative Gym: A Framework for Enabling and Evaluating Human-Agent Collaboration
Authors: Yijia Shao, Vinay Samuel, Yucheng Jiang, John Yang, Diyi Yang
Contact: shaoyj@stanford.edu
Links: Paper | Website | Video
Keywords: language model agents, human-AI collaboration, human-in-the-loop, evaluation
Cosmos Policy: Fine-Tuning Video Models for Visuomotor Control and Planning
Authors: Moo Jin Kim, Yihuai Gao, Tsung-Yi Lin, Yen-Chen Lin, Yunhao Ge, Grace Lam, Percy Liang, Shuran Song, Ming-Yu Liu, Chelsea Finn, Jinwei Gu
Contact: moojink@stanford.edu
Links: Paper | Website | Video
Keywords: world models, robotics, manipulation, model-based planning, imitation learning, video generation
Cost-of-Pass: An Economic Framework for Evaluating Language Models
Authors: Mehmet Hamza Erol, Batu El, Mirac Suzgun, Mert Yuksekgonul, James Y Zou
Contact: mhamza@stanford.edu
Links: Paper | Website
Keywords: economic evaluation framework, language-model evaluation, cost-performance trade-off, inference time techniques
CTC-DRO: Robust Optimization for Reducing Language Disparities in Speech Recognition
Authors: Martijn Bartelds, Ananjan Nandi, Moussa Koulako Bala Doumbouya, Dan Jurafsky, Tatsunori Hashimoto, Karen Livescu
Contact: bartelds@stanford.edu
Links: Paper | Website | Video
Keywords: distributionally robust optimization, deep learning, robustness, speech recognition
DeepSearch: Overcome the Bottleneck of Reinforcement Learning with Verifiable Rewards via Monte Carlo Tree Search
Authors: Fang Wu, Weihao Xuan, Heli Qi, Ximing Lu, Aaron Tu, Li Erran Li, Yejin Choi
Contact: fangwu97@stanford.edu
Links: Paper | Website
Keywords: MCTS, RLVR
Discrete Diffusion Trajectory Alignment via Stepwise Decomposition
Authors: Jiaqi Han, Austin Wang, Minkai Xu, Wenda Chu, Meihua Dang, Haotian Ye, Huayu Chen, Yisong Yue, Stefano Ermon
Contact: jiaqihan@stanford.edu
Keywords: discrete diffusion, language models, reinforcement learning
Distributional Machine Unlearning via Selective Data Removal
Authors: Youssef Allouah, Rachid Guerraoui, Sanmi Koyejo
Contact: yallouah@stanford.edu
Links: Paper
Keywords: unlearning, theory, privacy, sample complexity, machine learning, statistical learning
DSL-Monkeys: Self-Generated In-Context Examples for Low-Resource GPU DSL Kernels
Authors: Nathan Paek*, Simon Guo*, Vishnu Sarukkai*, Willy Chan*, William Hu, Ethan Boneh, Simran Arora, Ludwig Schmidt, Kayvon Fatahalian, Azalia Mirhoseini
Contact: simonguo@stanford.edu
Links: Paper
Keywords: GPU, in-context learning, low-resource data
ENACT: Evaluating Embodied Cognition with World Modeling of Egocentric Interaction
Authors: Qineng Wang*, Wenlong Huang*, Yu Zhou, Hang Yin, Tianwei Bao, Jianwen Lyu, Weiyu Liu, Ruohan Zhang, Jiajun Wu, Li Fei-Fei, Manling Li (*Equal Contribution)
Contact: wenlongh@stanford.edu
Award nominations: Oral Presentation at ICLR 2026 Workshop on World Models, Outstanding Paper Award at ICLR 2026 Workshop on Lifelong Agents
Links: Paper | Website | Video
Keywords: embodied agents, vision language models, benchmarking, world modeling
From Tokens to Thoughts: How LLMs and Humans Trade Compression for Meaning
Authors: Chen Shani, Liron Soffer, Dan Jurafsky, Yann LeCun, Ravid Shwartz-Ziv
Contact: cshani@stanford.edu
Links: Paper
Keywords: human-centered NLP, machine cognition, information theory for LLMs
Front-Loading Reasoning: The Synergy between Pretraining and Post-Training Data
Authors: Syeda Nahida Akter, Shrimai Prabhumoye, Eric Nyberg, Mostofa Patwary, Mohammad Shoeybi, Yejin Choi, Bryan Catanzaro
Contact: yejinc@cs.stanford.edu
Links: Paper
Keywords: pretraining, supervised finetuning, reasoning, LLM
Geometry-aware 4D Video Generation for Robot Manipulation
Authors: Zeyi Liu, Shuang Li, Eric Cousineau, Siyuan Feng, Benjamin Burchfiel, Shuran Song
Contact: liuzeyi@stanford.edu
Links: Paper | Website | Video
Keywords: video generation, robot manipulation, 3D perception
Goedel-Prover-V2: Scaling Formal Theorem Proving with Scaffolded Data Synthesis and Self-Correction
Authors: Yong Lin, Shange Tang, Bohan Lyu, Ziran Yang, Jui-Hui Chung, Haoyu Zhao, Lai Jiang, Yihan Geng, Jiawei Ge, Jingruo Sun, Jiayun Wu, Jiri Gesi, Ximing Lu, David Acuna, Kaiyu Yang, Hongzhou Lin, Yejin Choi, Danqi Chen, Sanjeev Arora, Chi Jin
Contact: yejinc@cs.stanford.edu
Links: Paper | Website
Keywords: theorem proving, reasoning
Humanline: Online Alignment as Perceptual Loss
Authors: Sijia Liu, Niklas Muennighoff, Kawin Ethayarajh
Contact: kawin@uchicago.edu
Links: Paper
Keywords: humanline, KTO
HUME: Measuring the Human-Model Performance Gap in Text Embedding Tasks
Authors: Adnan El Assadi, Isaac Chung, Roman Solomatin, Niklas Muennighoff, Kenneth Enevoldsen
Contact: niklasm@stanford.edu
Links: Paper | Website
Keywords: embeddings, MTEB
In-the-Flow Agentic System Optimization for Effective Planning and Tool Use
Authors: Zhuofeng Li, Haoxiang Zhang, Seungju Han, Sheng Liu, Jianwen Xie, Yu Zhang, Yejin Choi, James Zou, Pan Lu
Contact: lupantech@gmail.com
Award nominations: ICLR Oral Presentation, Best Paper Runner-up, NeurIPS 2025 ER Workshop
Links: Paper | Website | Video
Keywords: LLM agents, reinforcement learning, tool use, multi-turn interaction
Kevin: Multi-Turn RL for Generating CUDA Kernels
Authors: Carlo Baronio*, Pietro Marsella*, Ben Pan*, Simon Guo, Silas Alberti
Contact: simonguo@stanford.edu
Links: Paper | Website | Blog Post
Keywords: multi-turn, RL, GPU kernel, code generation
Learning to Summarize User Information for Personalized Reinforcement Learning from Human Feedback
Authors: Hyunji (Alex) Nam, Yanming Wan, Mickel Liu, Peter F. Ahnn, Jianxun Lian, Natasha Jaques
Contact: hjnam@stanford.edu
Links: Paper | Website
Keywords: pluralistic alignment, LLM personalization, reward modeling for RLHF
LitmusValues: Will AI Tell Lies to Save Sick Children? Litmus-Testing AI Values Prioritization with AIRiskDilemmas
Authors: Yu Ying Chiu, Zhilin Wang, Sharan Maiya, Yejin Choi, Kyle Fish, Sydney Levine, Evan J Hubinger
Contact: yejinc@cs.stanford.edu
Links: Paper | Website
Keywords: AI values, value alignment, AI risk, dilemma
Markovian Transformers for Informative Language Modeling
Authors: Scott W Viteri, Max Lamparth, Peter Chatain, Clark Barrett
Contact: scottviteri@gmail.com
Links: Paper
Keywords: Markovian transformers, chain-of-thought reasoning, language model interpretability, causal reasoning, reinforcement learning, next-token prediction, GSM8K, large language models
Measuring LLM Novelty As The Frontier Of Original And High-Quality Output
Authors: Vishakh Padmakumar, Chen Yueh-Han, Jane Pan, Valerie Chen, He He
Contact: vishakhp@stanford.edu
Links: Paper | Blog Post
Keywords: generation, evaluation, memorization, novelty, benchmark, creativity
Mixture of Contexts for Long Video Generation
Authors: Shengqu Cai, Ceyuan Yang, Lvmin Zhang, Yuwei Guo, Junfei Xiao, Ziyan Yang, Yinghao Xu, Zhenheng Yang, Alan Yuille, Leonidas Guibas, Maneesh Agrawala, Lu Jiang, Gordon Wetzstein
Contact: shengqu@stanford.edu
Links: Paper | Website | Blog Post | Video
Keywords: video generation, generative models, sparse attention, movie generation
MoReBench: Evaluating Procedural and Pluralistic Moral Reasoning in Language Models, More than Outcomes
Authors: Yu Ying Chiu, Michael S. Lee, Rachel Calcott, Brandon Handoko, Paul de Font-Reaulx, Paula Rodriguez, Chen Bo Calvin Zhang, Ziwen Han, Udari Madhushani Sehwag, Yash Maurya, Christina Q Knight, Harry R. Lloyd, Florence Bacus, Mantas Mazeika, Bing Liu, Yejin Choi, Mitchell L Gordon, Sydney Levine
Contact: yejinc@cs.stanford.edu
Links: Paper | Website
Keywords: moral reasoning, reasoning evaluation, AI safety
Moving Beyond Medical Exams: A Clinician-Annotated Fairness Dataset of Real-World Tasks and Ambiguity in Mental Healthcare
Authors: Max Lamparth, Declan Grabb, Amy Franks, Scott Gershan, Kaitlyn N Kunstman, Aaron Lulla, Monika Drummond Roots, Manu Sharma, Aryan Shrivastava, Nina Vasan, Colleen Waickman
Contact: lamparth@stanford.edu
Links: Paper | Blog Post
Keywords: AI for healthcare, mental health, fairness, bias, dataset, language models, decision-making, uncertainty, expert annotation
mR3: Multilingual Rubric-Agnostic Reward Reasoning Models
Authors: David Anugraha, Shou-Yi Hung, Zilu Tang, Annie En-Shiun Lee, Derry Tanti Wijaya, Genta Indra Winata
Contact: davidanu@stanford.edu
Links: Paper | Website
Keywords: multilingual, reasoning, reward model, rubrics
Multiplayer Nash Preference Optimization
Authors: Fang Wu, Xu Huang, Weihao Xuan, Zhiwei Zhang, Yijia Xiao, Guancheng Wan, Xiaomin Li, Bing Hu, Peng Xia, Jure Leskovec, Yejin Choi
Contact: fangwu97@stanford.edu
Award nominations: Oral
Links: Paper | Website
Keywords: preference optimization, RLHF, LLM post-training
No, of Course I Can! Deeper Fine-Tuning Attacks That Bypass Token-Level Safety Mechanisms
Authors: Joshua Kazdan, Abhay Puri, Rylan Schaeffer, Lisa Yu, Chris Cundy, Jason Stanley, Sanmi Koyejo, Krishnamurthy Dj Dvijotham
Contact: jkazdan@stanford.edu
Links: Paper
Keywords: jailbreaking, attacks, AI safety, red-teaming, fine-tuning, fine-tuning attacks
ODESteer: A Unified ODE-Based Steering Framework for LLM Alignment
Authors: Hongjue Zhao, Haosen Sun, Jiangtao Kong, Xiaochang Li, Qineng Wang, Liwei Jiang, Qi Zhu, Tarek F. Abdelzaher, Yejin Choi, Manling Li, Huajie Shao
Contact: yejinc@cs.stanford.edu
Links: Paper | Website
Keywords: LLM alignment, representation engineering, activation steering, ODE-based framework, barrier functions
OpenThoughts: Data Recipes for Reasoning Models
Authors: Etash Kumar Guha, Ryan Marten, Sedrick Keh, Negin Raoof, Georgios Smyrnis, Hritik Bansal, Marianna Nezhurina, Jean Mercat, Trung Vu, Zayne Rea Sprague, Ashima Suvarna, Benjamin Feuer, Leon Liangyu Chen, Zaid Khan, Eric Frankel, Sachin Grover, Caroline Choi, Niklas Muennighoff, Shiye Su, Wanjia Zhao, et al.
Contact: yejinc@cs.stanford.edu
Links: Paper
Keywords: reasoning, data, LLM
Optimal Aggregation Mechanisms for AI Benchmarking and Platinum Benchmarks
Authors: Andreas Haupt, Anka Reuel, Mykel Kochenderfer, Sanmi Koyejo
Contact: h4upt@stanford.edu
Links: Paper
Keywords: mechanism design, AI evaluation, benchmarking, multi-tasking
pi-Flow: Policy-Based Few-Step Generation via Imitation Distillation
Authors: Hansheng Chen, Kai Zhang, Hao Tan, Leonidas Guibas, Gordon Wetzstein, Sai Bi
Contact: hshchen@stanford.edu
Links: Paper | Website
Keywords: diffusion models, flow models, few-step generation, distillation
Polychromic Objectives for Reinforcement Learning
Authors: Jubayer Ibn Hamid, Ifdita Hasan Orney, Ellen Xu, Chelsea Finn, Dorsa Sadigh
Contact: jubayer@stanford.edu
Links: Paper
Keywords: reinforcement learning, exploration
PoseX: AI Defeats Physics-based Methods on Protein Ligand Cross-Docking
Authors: Yize Jiang, Xinze Li, Yuanyuan Zhang, Jin Han, Youjun Xu, Ayush Pandit, Zaixi Zhang, Mengdi Wang, Mengyang Wang, Chong Liu, Guang Yang, Yejin Choi, Yingzhou Lu, Wu-Jun Li, Tianfan Fu, Fang Wu, Junhong Liu
Contact: fangwu97@stanford.edu
Links: Paper | Website
Keywords: AI docking, AI co-folding, protein-ligand interaction, cross docking
Pre-training under infinite compute
Authors: Konwoo Kim, Suhas Kotha, Percy Liang, Tatsunori Hashimoto
Contact: konwoo@stanford.edu, kotha@stanford.edu
Award nominations: Oral
Links: Paper
Keywords: pre-training, data efficiency, scaling laws
Pretraining Scaling Laws for Generative Evaluations of Language Models
Authors: Rylan Schaeffer, Noam Itzhak Levi, Brando Miranda, Sanmi Koyejo
Contact: rschaef@cs.stanford.edu
Links: Paper
Keywords: language models, large language models, scaling laws, evaluations, generative evaluations, sampling
ProfBench: Multi-Domain Rubrics requiring Professional Knowledge to Answer and Judge
Authors: Zhilin Wang, Jaehun Jung, Ximing Lu, Shizhe Diao, Ellie Evans, Jiaqi Zeng, Pavlo Molchanov, Yejin Choi, Jan Kautz, Yi Dong
Contact: yejinc@cs.stanford.edu
Links: Paper | Website
Keywords: expert-annotated, professional knowledge, LLM judge, rubric evaluation
Relational Transformer: Toward Zero-Shot Foundation Models for Relational Data
Authors: Rishabh Ranjan, Valter Hudovernik, Mark Znidar, Charilaos Kanatsoulis, Roshan Upendra, Mahmoud Mohammadi, Joe Meyer, Tom Palczewski, Carlos Guestrin, Jure Leskovec
Contact: ranjanr@stanford.edu
Award nominations: Oral at AI for Tabular Data Workshop, NeurIPS 2025
Links: Paper | Website
Keywords: foundation models, tabular data, relational databases, transformer
RLP: Reinforcement as a Pretraining Objective
Authors: Ali Hatamizadeh, Syeda Nahida Akter, Shrimai Prabhumoye, Jan Kautz, Mostofa Patwary, Mohammad Shoeybi, Bryan Catanzaro, Yejin Choi
Contact: yejinc@cs.stanford.edu
Links: Paper
Keywords: reinforcement learning, pretraining, reasoning, large language models
SCRIBES: Web-Scale Script-Based Semi-Structured Data Extraction with Reinforcement Learning
Authors: Shicheng Liu, Kai Sun, Lisheng Fu, Xilun Chen, Xinyuan Zhang, Zhaojiang Lin, Rulin Shao, Yue Liu, Anuj Kumar, Wen-tau Yih, Xin Luna Dong
Contact: shicheng@cs.stanford.edu
Links: Paper
Keywords: semi-structured data, reinforcement learning, information extraction
SimBench: Benchmarking the Ability of Large Language Models to Simulate Human Behaviors
Authors: Tiancheng Hu, Joachim Baumann, Lorenzo Lupo, Nigel Collier, Dirk Hovy, Paul Röttger
Contact: th656@cam.ac.uk
Links: Paper | Website
Keywords: human behavior simulation, large language models, benchmarking, computational social science, human-AI alignment, calibration, human-centered AI
SimpleToM: Exposing the Gap between Explicit ToM Inference and Implicit ToM Application in LLMs
Authors: Yuling Gu, Oyvind Tafjord, Hyunwoo Kim, Jared Moore, Ronan Le Bras, Peter Clark, Yejin Choi
Contact: yejinc@cs.stanford.edu
Links: Paper
Keywords: theory of mind, social reasoning, LLM benchmark, mental state, behavior, judgment, false belief
Spatial Mental Modeling from Limited Views
Authors: Qineng Wang, Baiqiao Yin, Pingyue Zhang, Jianshu Zhang, Kangrui Wang, Zihan Wang, Jieyu Zhang, Keshigeyan Chandrasegaran, Han Liu, Ranjay Krishna, Saining Xie, Jiajun Wu, Li Fei-Fei, Manling Li
Contact: qinengwang2029@u.northwestern.edu
Award nominations: Best Paper Honorable Mention at NeurIPS 2025 LAW Workshop, Best Paper Award at ICCV 2025 SP4V Workshop, The Best of ICCV featured by Voxel51
Links: Paper | Website
Keywords: vision language models, VLMs, multimodal language models, spatial intelligence, spatial reasoning
Spectrum Tuning: Post-Training for Distributional Coverage and In-Context Steerability
Authors: Taylor Sorensen, Benjamin Newman, Jared Moore, Chan Young Park, Jillian Fisher, Niloofar Mireshghallah, Liwei Jiang, Yejin Choi
Contact: yejinc@cs.stanford.edu
Links: Paper | Website
Keywords: post-training, language models, distributional learning, alignment, pluralistic alignment, uncertainty estimation
Speculative Speculative Decoding
Authors: Tanishq Kumar, Tri Dao, Avner May
Contact: tanishq@stanford.edu
Links: Paper
Keywords: inference, LLMs, systems
Theory of Space: Can Foundation Models Construct Spatial Beliefs through Active Exploration?
Authors: Pingyue Zhang, Zihan Huang, Yue Wang, Jieyu Zhang, Letian Xue, Zihan Wang, Qineng Wang, Keshigeyan Chandrasegaran, Ruohan Zhang, Yejin Choi, Ranjay Krishna, Jiajun Wu, Li Fei-Fei, Manling Li
Contact: pingyuezhang2029@u.northwestern.edu
Links: Paper | Website | Blog Post
Keywords: large language model, vision-language model, spatial reasoning, spatial agent, active exploration
ThinkMorph: Emergent Properties in Multimodal Interleaved Chain-of-Thought Reasoning
Authors: Jiawei Gu, Yunzhuo Hao, Huichen Will Wang, Linjie Li, Michael Qizhe Shieh, Yejin Choi, Ranjay Krishna, Yu Cheng
Contact: yejinc@cs.stanford.edu
Links: Paper | Website
Keywords: multimodal reasoning, interleaved chain-of-thought, unified model
TrustGen: A Platform of Dynamic Benchmarking on the Trustworthiness of Generative Foundation Models
Authors: Yue Huang, Chujie Gao, Siyuan Wu, Haoran Wang, Xiangqi Wang, Jiayi Ye, Yujun Zhou, Yanbo Wang, Jiawen Shi, Qihui Zhang, Han Bao, Zhaoyi Liu, Yuan Li, Tianrui Guan, Peiran Wang, Haomin Zhuang, Dongping Chen, Kehan Guo, Andy Zou, Bryan Hooi, et al.
Contact: yhuang37@nd.edu
Links: Paper | Website
Keywords: trustworthiness, generative model, large language model, vision-language model, dynamic evaluation, benchmark
Tversky Neural Networks: Psychologically Plausible Deep Learning with Differentiable Tversky Similarity
Authors: Moussa Koulako Bala Doumbouya, Dan Jurafsky, Christopher D Manning
Contact: moussa@stanford.edu
Links: Paper | Website | Blog Post
Keywords: machine learning, psychology, neural networks
Unfolding Spatial Cognition: Evaluating Multimodal Models on Visual Simulations
Authors: Linjie Li, Mahtab Bigverdi, Jiawei Gu, Zixian Ma, Yinuo Yang, Ziang Li, Yejin Choi, Ranjay Krishna
Contact: yejinc@cs.stanford.edu
Links: Paper
Keywords: spatial reasoning, visual reasoning
Unified 3D Scene Understanding Through Physical World Modeling
Authors: Wanhee Lee, Klemen Kotar, Rahul Mysore Venkatesh, Jared Watrous, Honglin Chen, Khai Loong Aw, Daniel L. K. Yamins
Contact: wanhee@stanford.edu
Links: Paper
Keywords: 3D scene understanding, visual world models
We look forward to seeing you at ICLR 2026!