FreeVS: Generative View Synthesis on Free Driving Trajectory

1CASIA & UCAS, 2CUHK, 3CAIR, HKISI, CAS

The pipeline of FreeVS.

Abstract

Existing reconstruction-based novel view synthesis methods for driving scenes focus on synthesizing camera views along the recorded trajectory of the ego vehicle. Their image rendering performance will severely degrade on viewpoints falling out of the recorded trajectory, where camera rays are untrained. We propose FreeVS, a novel fully generative approach that can synthesize camera views on free new trajectories in real driving scenes. To control the generation results to be 3D consistent with the real scenes and accurate in viewpoint pose, we propose the pseudo-image representation of view priors to control the generation process. Viewpoint transformation simulation is applied on pseudo-images to simulate camera movement in each direction. Once trained, FreeVS can be applied to any validation sequences without reconstruction process and synthesis views on novel trajectories. Moreover, we propose two new challenging benchmarks tailored to driving scenes, which are novel camera synthesis and novel trajectory synthesis, emphasizing the freedom of viewpoints. Given that no ground truth images are available on novel trajectories, we also propose to evaluate the consistency of images synthesized on novel trajectories with 3D perception models. Experiments on the Waymo Open Dataset show that FreeVS has a strong image synthesis performance on both the recorded trajectories and novel trajectories. The code will be released.

Free-Trajectory View Generation Demos

Scene 1 (GT Video)

Scene 1 (Novel Trajectory)

Scene 2 (GT Video)

Scene 2 (Novel Trajectory)

Scene 3 (GT Video)

Scene 3 (Novel Trajectory)

Scene 4 (GT Video)

Scene 4 (Novel Trajectory)

Scene 5 (GT Video)

Scene 5 (Novel Trajectory)

Scene 6 (GT Video)

Scene 6 (Novel Trajectory)

Compare with SOTA NVS methods

Modifying Objets in Scene

Scene 1 (GT Video)

Scene 1 (Modifying Object Appearance)

Scene 2 (GT Video)

Scene 2 (Modifying Object Position)

Novel Camera Sythesis

We also propose the novel camera synthesis benchmark, in which we drop all images of certain camera views(e.g. the front-side views) and require NVS methods to synthesis images on those unseen camera position.

Compare with SOTA NVS methods on novel camera synthesis