Research

Paper

TESTING February 27, 2026

UFO-4D: Unposed Feedforward 4D Reconstruction from Two Images

Authors

Junhwa Hur, Charles Herrmann, Songyou Peng, Philipp Henzler, Zeyu Ma, Todd Zickler, Deqing Sun

Abstract

Dense 4D reconstruction from unposed images remains a critical challenge, with current methods relying on slow test-time optimization or fragmented, task-specific feedforward models. We introduce UFO-4D, a unified feedforward framework to reconstruct a dense, explicit 4D representation from just a pair of unposed images. UFO-4D directly estimates dynamic 3D Gaussian Splats, enabling the joint and consistent estimation of 3D geometry, 3D motion, and camera pose in a feedforward manner. Our core insight is that differentiably rendering multiple signals from a single Dynamic 3D Gaussian representation offers major training advantages. This approach enables a self-supervised image synthesis loss while tightly coupling appearance, depth, and motion. Since all modalities share the same geometric primitives, supervising one inherently regularizes and improves the others. This synergy overcomes data scarcity, allowing UFO-4D to outperform prior work by up to 3 times in joint geometry, motion, and camera pose estimation. Our representation also enables high-fidelity 4D interpolation across novel views and time. Please visit our project page for visual results: https://ufo-4d.github.io/

Metadata

arXiv ID: 2602.24290
Provider: ARXIV
Primary Category: cs.CV
Published: 2026-02-27
Fetched: 2026-03-02 06:04

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2602.24290v1</id>\n    <title>UFO-4D: Unposed Feedforward 4D Reconstruction from Two Images</title>\n    <updated>2026-02-27T18:59:54Z</updated>\n    <link href='https://arxiv.org/abs/2602.24290v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2602.24290v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>Dense 4D reconstruction from unposed images remains a critical challenge, with current methods relying on slow test-time optimization or fragmented, task-specific feedforward models. We introduce UFO-4D, a unified feedforward framework to reconstruct a dense, explicit 4D representation from just a pair of unposed images. UFO-4D directly estimates dynamic 3D Gaussian Splats, enabling the joint and consistent estimation of 3D geometry, 3D motion, and camera pose in a feedforward manner. Our core insight is that differentiably rendering multiple signals from a single Dynamic 3D Gaussian representation offers major training advantages. This approach enables a self-supervised image synthesis loss while tightly coupling appearance, depth, and motion. Since all modalities share the same geometric primitives, supervising one inherently regularizes and improves the others. This synergy overcomes data scarcity, allowing UFO-4D to outperform prior work by up to 3 times in joint geometry, motion, and camera pose estimation. Our representation also enables high-fidelity 4D interpolation across novel views and time. Please visit our project page for visual results: https://ufo-4d.github.io/</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.CV'/>\n    <published>2026-02-27T18:59:54Z</published>\n    <arxiv:comment>ICLR 2026, Project page: https://ufo-4d.github.io/</arxiv:comment>\n    <arxiv:primary_category term='cs.CV'/>\n    <author>\n      <name>Junhwa Hur</name>\n    </author>\n    <author>\n      <name>Charles Herrmann</name>\n    </author>\n    <author>\n      <name>Songyou Peng</name>\n    </author>\n    <author>\n      <name>Philipp Henzler</name>\n    </author>\n    <author>\n      <name>Zeyu Ma</name>\n    </author>\n    <author>\n      <name>Todd Zickler</name>\n    </author>\n    <author>\n      <name>Deqing Sun</name>\n    </author>\n  </entry>"
}