add model links
This commit is contained in:
parent
b7b9cf2912
commit
cf9b118345
13
README.md
13
README.md
@ -20,6 +20,9 @@ We further share insights on our model of uncertainty in case of limited observa
|
|||||||
```
|
```
|
||||||
|
|
||||||
Add link paper
|
Add link paper
|
||||||
|
|
||||||
|
Add link video
|
||||||
|
|
||||||

|

|
||||||
# Setup
|
# Setup
|
||||||
|
|
||||||
@ -49,7 +52,8 @@ mkdir arrays models kitti nuscenes logs
|
|||||||
```
|
```
|
||||||
|
|
||||||
### Pre-trained Models
|
### Pre-trained Models
|
||||||
* Download a MonoLoco pre-trained model from Google Drive: ADD LINK and save it in `data/models`
|
* Download a MonoLoco pre-trained model from
|
||||||
|
[Google Drive](https://drive.google.com/open?id=1F7UG1HPXGlDD_qL-AN5cv2Eg-mhdQkwv) and save it in `data/models`
|
||||||
* Download a Pifpaf pre-trained model from [openpifpaf](https://github.com/vita-epfl/openpifpaf) project
|
* Download a Pifpaf pre-trained model from [openpifpaf](https://github.com/vita-epfl/openpifpaf) project
|
||||||
and save it into `data/models`
|
and save it into `data/models`
|
||||||
|
|
||||||
@ -81,7 +85,8 @@ Output options include json files and/or visualization of the predictions on the
|
|||||||
* In case you provide a ground-truth json file to compare the predictions of MonoLoco,
|
* In case you provide a ground-truth json file to compare the predictions of MonoLoco,
|
||||||
the script will match every detection using Intersection over Union metric.
|
the script will match every detection using Intersection over Union metric.
|
||||||
The ground truth file can be generated using the subparser `prep` and called with the command `--path_gt`.
|
The ground truth file can be generated using the subparser `prep` and called with the command `--path_gt`.
|
||||||
Check preprocess section for more details.
|
Check preprocess section for more details or download the file from
|
||||||
|
[here](https://drive.google.com/open?id=1F7UG1HPXGlDD_qL-AN5cv2Eg-mhdQkwv).
|
||||||
|
|
||||||
* In case you don't provide a ground-truth file, the script will look for a predefined path.
|
* In case you don't provide a ground-truth file, the script will look for a predefined path.
|
||||||
If it does not find the file, it will generate images
|
If it does not find the file, it will generate images
|
||||||
@ -119,13 +124,13 @@ To extract pifpaf joints, you also need to download training images, put it in a
|
|||||||
data/kitti/images`
|
data/kitti/images`
|
||||||
|
|
||||||
#### 2) nuScenes dataset
|
#### 2) nuScenes dataset
|
||||||
Download nuScenes dataset (any version: Mini, Teaser or TrainVal) from [nuScenes](https://www.nuscenes.org/download),
|
Download nuScenes dataset from [nuScenes](https://www.nuscenes.org/download) (either Mini or Full),
|
||||||
save it anywhere and soft link it in `data/nuscenes`
|
save it anywhere and soft link it in `data/nuscenes`
|
||||||
|
|
||||||
|
|
||||||
### Annotations to preprocess
|
### Annotations to preprocess
|
||||||
MonoLoco is trained using 2D human pose joints. To create them run pifaf over KITTI or nuScenes training images.
|
MonoLoco is trained using 2D human pose joints. To create them run pifaf over KITTI or nuScenes training images.
|
||||||
You can create them running the predict script and using `--network pifpaf`.
|
You can create them running the predict script and using `--networks pifpaf`.
|
||||||
|
|
||||||
### Inputs joints for training
|
### Inputs joints for training
|
||||||
MonoLoco is trained using 2D human pose joints matched with the ground truth location provided by
|
MonoLoco is trained using 2D human pose joints matched with the ground truth location provided by
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user