RayNet:CNN+MRF

选中这篇文章的原因和目的是:

  • (1)与我之前的思路类似,结合传统方法对CNN的抽象学习任务进行简化。学习他的实现方案。
  • (2)从图像获得三维信息,与我现在准备做的工作也类似,学习相关国内外研究现状。
  • (3)看看人家故事怎么讲的。

讲故事

In this paper, we consider the problem of reconstructing a dense 3D model using images captured from different views. Recent methods based on convolutional neural networks (CNN) allow learning the entire task from data. However, they do not incorporate the physics of image formation such as perspective geometry and occlusion. Instead, classical approaches based on Markov Random Fields (MRF) with ray-potentials explicitly model these physical processes, but they cannot cope with large surface appearance variations across different viewpoints. In this paper, we propose RayNet, which combines the strengths of both frameworks. RayNet integrates a CNN that learns view-invariant feature representations with an MRF that explicitly encodes the physics of perspective projection and occlusion. We train RayNet end-to-end using empirical risk minimization. We thoroughly evaluate our approach on challenging real-world datasets and demonstrate its benefits over a piece-wise trained baseline, hand-crafted models as well as other learning-based approaches.

猜你喜欢

转载自blog.csdn.net/duyue3052/article/details/81304082
CNN