|
CRISP: Contact-guided Real2Sim from Monocular Video with Planar Scene Primitives
Zihan Wang*,
Jiashun Wang*,
Jeff Tan,
Yiwen Zhao,
Jessica Hodgins,
Shubham Tulsiani
Deva Ramanan
arXiv 2025
Website |
arXiv |
Github
We recover high-quality and simulatable scene geometry from Internet videos, for humanoid real-to-sim within complex terrains.
|
|
MonoFusion: Sparse-View 4D Reconstruction via Monocular Fusion
Zihan Wang,
Jeff Tan,
Tarasha Khurana,
Neehar Peri,
Deva Ramanan
ICCV 2025
Website |
arXiv |
Github
We reconstruct dynamic 4D human-centric scenes (e.g. playing the piano and bicycle repair) from sparse-view video.
|
|
DiffusionSfM: Predicting Structure and Motion via Ray Origin and Endpoint Diffusion
Qitao Zhao,
Amy Lin,
Jeff Tan,
Jason Y. Zhang,
Deva Ramanan,
Shubham Tulsiani
CVPR 2025
Website |
arXiv |
Github
From a set of multi-view images, we learn a denoising diffusion model that outputs scene geometry and cameras in global frame.
|
|
DressRecon: Freeform 4D Human Reconstruction from Monocular Video
Jeff Tan,
Donglai Xiang,
Shubham Tulsiani,
Deva Ramanan,
Gengshan Yang
3DV 2025 (Oral)
Website |
arXiv |
Github
From a single monocular video, we reconstruct humans in loose clothing and interacting with objects, using a hierarchical deformation field and image-based priors.
|
|
|
Distilling Neural Fields for Real-Time Articulated Shape Reconstruction
Jeff Tan,
Gengshan Yang,
Deva Ramanan
CVPR 2023
Website |
Paper |
Github
We learn real-time feed-forward pose and shape predictors, by distilling knowledge from offline differentiable rendering optimizers.
|
Credits to Jon Barron for this website's template.
|
|