
In recent years, there has been a growing trend toward training data-centric, large-scale foundation models that reduce reliance on structural priors. However, is simply scaling up Transformers truly the ultimate solution for computer vision? In this workshop, we aim to reintroduce structural priors and explore how they can further push the boundaries of foundation models.
Our workshop provides an interdisciplinary space for sharing ideas across domains. For example, scene-aware 2D perception can enhance 3D modeling and robotic manipulation, while geometric reasoning can enhance the visual grounding of 2D perception and multimodal models. Through these interactions, we aim to better define the role of priors in vision foundation models.
Our topics include but are not limited to:
- Scene-aware vision models for images and videos.
- Geometry and equivariance for 3D vision.
- Temporal and motion priors for videos.
- Behavioral priors for robotics and egocentric views.
- Physics priors for world models and interactions.

Georgia Tech & NVIDIA

Google DeepMind

Stanford

UT Austin

NYU

MIT
Opening Remarks and Welcome | 08:50-09:00 |
Keynote Talk: Speaker TBD
Title TBD
|
09:00-09:40 |
Keynote Talk: Speaker TBD
Title TBD
|
09:40-10:20 |
Coffee Break | 10:20-10:40 |
Keynote Talk: Speaker TBD
Title TBD
|
10:40-11:20 |
Spotlight Talk
Title TBD
|
11:20-11:35 |
Spotlight Talk
Title TBD
|
11:35-11:50 |
Lunch Break | 11:50-12:30 |
Accepted Paper Poster Session | 12:30-13:30 |
Keynote Talk: Speaker TBD
Title TBD
|
13:30-14:10 |
Keynote Talk: Speaker TBD
Title TBD
|
14:10-14:50 |
Coffee Break | 14:50-15:10 |
Keynote Talk: Speaker TBD
Title TBD
|
15:10-15:50 |
Spotlight Talk
Title TBD
|
15:50-16:05 |
Spotlight Talk
Title TBD
|
16:05-16:20 |
Closing Remarks | 16:20-16:30 |
Accepted Paper Poster Session | 16:30-17:30 |
Ground-Displacement Forecasting from Satellite Image Time Series via a Koopman-Prior Autoencoder | |
Spatial Mental Modeling from Limited Views | |
SEAL-Pose: Enhancing 3D Human Pose Estimation through Trainable Loss Function | |
StereoDiff: Stereo-Diffusion Synergy for Video Depth Estimation | |
VidMP3: Video Editing by Representing Motion with Pose and Position Priors | |
The Diashow Paradox: Stronger 3D-Aware Representations Emerge from Image Sets, Not Videos | |
Identity-Motion Trade-offs in Text-to-Video via Query-Guided Attention Priors | |
Axis-level Symmetry Detection with Group-Equivariant Representation | |
Combinative Matching for Geometric Shape Assembly | |
Perspective-Aware Reasoning in Vision-Language Models via Mental Imagery Simulation | |
Generic Event Boundary Detection via Denoising Diffusion | |
SHED Light on Segmentation for Depth Estimation | |
Few-Shot Pattern Detection via Template Matching and Regression | |
MultiViewPano: A Generalist Approach to 360-degree Panorama Generation | |
LACONIC: A 3D Layout Adapter for Controllable Image Creation | |
SuperDec: 3D Scene Decomposition with Superquadric Primitives | |
Injecting Geometric Scene Priors into Vision Transformers for Improved 2D-3D Understanding |

UMich

Stanford

Black Forest Labs

Google DeepMind

NVIDIA

Stanford

UMich