-
Notifications
You must be signed in to change notification settings - Fork 37
Description
Method
In this section, we describe our unsupervised framework for monocular depth estimation. We first review the self-supervised training pipeline for monocular depth estimation, and then introduce the co-attention module and pose graph consistency loss function.
Supervision from Image Reconstruction
Following the formulation in \cite{zhou_unsupervised_2017}, the whole framework includes a DispNet and a PoseNet, the DispNet produces depth map and the PoseNet produces the relative pose between two RGB frames.
Given a sequence of consecutive frames
Consider the adjacent frame pair
the function
Hence the problem is formulated to the minimization of a phtometric reprojection error
List
Here is a list:
- Xue Bai, Jue Wang, David Simons, and Guillermo Sapiro.Video SnapCut: robust video object cutout using localized classifiers. TOG, 28(3):70, 2009.
- Linchao Bao, Baoyuan Wu, and Wei Liu. CNN in MRF: Video object segmentation via inference in a CNN-based higher-order spatio-temporal MRF. In CVPR, 2018
Code
Here is some code:
def bi_search(arr:list, x:int):
l, r = 0, len(arr)
while l < r:
m = (l + r) >> 1
if arr[m] >= x: r = m
else: l = m + 1
return lImage
Table
| A | B | C |
|---|---|---|
| 123 | 456 | 789 |
