📝 Note: 1) Explanations of abbreviations and more datasets will be added in future updates. 2) GC=GeometryCrafter POM=POMATO
Dataset | S y n |
V i d |
D y n |
I n d |
O u t |
Resolution | C 3 R |
D P |
G C |
M o G |
P O M |
R D |
U D 2 |
V D A |
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1 | Spring | ✓ | ✓ | ✓ | ✗ | ✓ | 1920×1080 | T | E | T | T | - | - | - | - |
2 | MVS-Synth | ✓ | ✓ | ✗ | ✗ | ✓ | 1920×1080 | T | T | T | T | - | - | - | - |
3 | Mid-Air | ✓ | ✓ | ✗ | ✗ | ✓ | 1024×1024 | - | - | T | T | - | - | - | - |
4 | MatrixCity | ✓ | ✓ | ✗ | ✗ | ✓ | 1000×1000 | - | - | T | T | - | - | T | - |
5 | SAIL-VOS 3D | ✓ | ✓ | ✓ | ✓ | ✓ | 1280×800 | - | T | - | - | - | - | - | - |
6 | BEDLAM | ✓ | ✓ | ✓ | ✓ | ✓ | 1280×720 | T | T | - | - | - | - | T | - |
7 | Dynamic Replica | ✓ | ✓ | ✓ | ✓ | ✗ | 1280×720 | T | T | T | - | T | - | T | - |
8 | PointOdyssey | ✓ | ✓ | ✓ | ✓ | ✓ | 960×540 | T | - | - | - | T | E | T | T |
9 | DyDToF | ✓ | ✓ | ✓ | ✓ | ✗ | 960×540 | - | - | - | - | - | E | - | - |
10 | IRS | ✓ | ✓ | ✗ | ✓ | ✗ | 960×540 | T | T | T | T | - | - | - | T |
11 | Scene Flow | ✓ | ✓ | ✓ | ✓ | ✓ | 960×540 | - | - | E | - | - | - | - | - |
12 | TartanAir | ✓ | ✓ | ✓ | ✓ | ✓ | 640×480 | T | T | T | T | T | T | T | T |
13 | |||||||||||||||
14 | ParallelDomain-4D | ✓ | ✓ | ✓ | ✗ | ✓ | 640×480 | - | - | - | - | T | - | - | - |
15 | GTA-SfM | ✓ | ✓ | ✗ | ✗ | ✓ | 640×480 | - | - | T | T | - | - | - | - |
16 | MPI Sintel | ✓ | ✓ | ✓ | ✓ | ✓ | 1024×436 | E | E | E | E | E | - | E | E |
17 | Virtual KITTI 2 | ✓ | ✓ | ✓ | ✗ | ✓ | 1242×375 | T | T | T | - | - | - | - | T |
- Bonn RGB-D Dynamic (5 video clips with 110 frames each): AbsRel<=0.079
- NYU-Depth V2: AbsRel<=0.0424 (relative depth)
- NYU-Depth V2: AbsRel<=0.051 (metric depth)
- Appendix 1: Rules for qualifying models for the rankings (to do)
- Appendix 2: Metrics selection for the rankings (to do)
- Appendix 3: List of all research papers from the above rankings
📝 Note: There are no quantitative comparison results of StereoCrafter yet, so this ranking is based on my own perceptual judgement of the qualitative comparison results shown in Figure 7. One output frame (right view) is compared with one input frame (left view) from the video clip: 22_dogskateboarder and one output frame (right view) is compared with one input frame (left view) from the video clip: scooter-black
RK | Model Links: Venue Repository |
Rank ↓ (human perceptual judgment) |
---|---|---|
1 | StereoCrafter |
1 |
2-3 | Immersity AI | 2-3 |
2-3 | Owl3D | 2-3 |
4 | Deep3D |
4 |
📝 Note: 1) See Figure 4 2) The ranking order is determined in the first instance by a direct comparison of the scores of two models in the same paper. If there is no such direct comparison in any paper or there is a disagreement in different papers, the ranking order is determined by the best score of the compared two models in all papers that are shown in the columns as data sources. The DepthCrafter rank is based on the latest version 1.0.1.
📝 Note: The ranking order is determined in the first instance by a direct comparison of the scores of two models in the same paper. If there is no such direct comparison in any paper or there is a disagreement in different papers, the ranking order is determined by the best score of the compared two models in all papers that are shown in the columns as data sources. The Metric3D v2 ViT-Large rank is not based on a score of 0.134, which is probably just an anomaly.