You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Codebase for ACM MM'22 paper titled "Robust Multimodal Depth Estimation using Transformer based Generative Adversarial Networks"
2
+
3
+
This repo is the PyTorch implementation of our ACM Multimedia'22 paper on [Robust Multimodal Depth Estimation using Transformer based Generative Adversarial Networks](https://dl.acm.org/doi/abs/10.1145/3503161.3548418)
4
+
5
+
<palign="center">
6
+
<img src="misc/teapot.png" alt="photo not available">
7
+
</p>
8
+
9
+
## Requirements
10
+
The base environment(python 3.6) consists of:
11
+
```
12
+
pytorch == 1.10.2
13
+
torchvision == 0.11.3
14
+
tensorboard == 1.15
15
+
py-opencv == 4.5.5
16
+
pillow == 8.4.0
17
+
numpy == 1.17.4
18
+
typing == 3.6.4
19
+
```
20
+
21
+
## Dataset
22
+
Primarily two datasets were used [ShapeNet](https://shapenet.org/) and [NYU_v2](https://cs.nyu.edu/~silberman/datasets/nyu_depth_v2.html)
2. Modify the val dataloader path with appropriate data directory
30
+
3. Typically the directory has the following structure
31
+
----|->data.nyu_v2|
32
+
|->train|
33
+
|->sparse_depth
34
+
|->depth_gt
35
+
|->image_rgb
36
+
|->meta_info.txt
37
+
|->val|
38
+
|->sparse_depth
39
+
|->depth_gt
40
+
|->image_rgb
41
+
|->meta_info.txt
42
+
|->sample|
43
+
|->sparse_depth
44
+
|->depth_gt
45
+
|->image_rgb
46
+
|->meta_info.txt
47
+
48
+
4. The "depth_gt" and "sparse_depth" are the folders containing dense and sparse depth respectively
49
+
5. The meta_info.txt contains the file names of these folders. Refer to misc/ folder for sample meta_info file
50
+
6. The folder "sample" contains a few sparse samples. This is to track the model learning visually. This is optional.
51
+
52
+
53
+
## Validation
54
+
You can run standalone validation if you have a trained model. For that the checkpoint model path has to have 2 files named generator_best.pth and discriminator_best.pth. You can invoke the validation script by:
This work was supported in part by National Science Foundation (NSF) SOPHIA (CCF-1822923) and Center for Brain-inspired Computing (C-BRIC) & Center for Research in Intelligent Storage and Processing in Memory (CRISP), two of the six centers in JUMP, a Semiconductor Research Corporation (SRC) program sponsored by DARPA.
0 commit comments