SynSin: End-to-end View Synthesis from a Single Image: Difference between revisions

From David's Wiki
No edit summary
No edit summary
Line 1: Line 1:
[[Category:Papers]]
SynSin: End-to-end View Synthesis from a Single Image (CVPR 2020)
SynSin: End-to-end View Synthesis from a Single Image (CVPR 2020)



Revision as of 16:37, 31 January 2022


SynSin: End-to-end View Synthesis from a Single Image (CVPR 2020)

Authors: Olivia Wiles, Georgia Gkioxari, Richard Szeliski, Justin Johnson
Affiliations: University of Oxford, Facebook AI Research, Facebook, University of Michigan

Method

Figure 2 from SynSin Paper
  1. First a depth map and a set of features are generated for each pixel using depth network \(d\) and feature network \(f\).
  2. The depths are used to create a 3D point cloud of features \(P\).
  3. Features are repositioned using the transformation matrix T.
  4. Repositioned features are rendered using a neural point cloud renderer.
  5. Rendered features are passed through a refinement network \(g\).


Architecture

Feature Network

Depth Network

Neural Point Cloud Rendering

Refinement Network

Evaluation

References