VRIS_vip / docs /data.md
dianecy's picture
Add files using upload-large-folder tool
9b855a7 verified
|
raw
history blame
4.64 kB

Data Preparation

Create a new directory data to store all the datasets.

Ref-COCO

Download the dataset from the official website COCO.
RefCOCO/+/g use the COCO2014 train split. Download the annotation files from github.

Convert the annotation files:

python3 tools/data/convert_refexp_to_coco.py

Finally, we expect the directory structure to be the following:

ReferFormer
β”œβ”€β”€ data
β”‚   β”œβ”€β”€ coco
β”‚   β”‚   β”œβ”€β”€ train2014
β”‚   β”‚   β”œβ”€β”€ refcoco
β”‚   β”‚   β”‚   β”œβ”€β”€ instances_refcoco_train.json
β”‚   β”‚   β”‚   β”œβ”€β”€ instances_refcoco_val.json
β”‚   β”‚   β”œβ”€β”€ refcoco+
β”‚   β”‚   β”‚   β”œβ”€β”€ instances_refcoco+_train.json
β”‚   β”‚   β”‚   β”œβ”€β”€ instances_refcoco+_val.json
β”‚   β”‚   β”œβ”€β”€ refcocog
β”‚   β”‚   β”‚   β”œβ”€β”€ instances_refcocog_train.json
β”‚   β”‚   β”‚   β”œβ”€β”€ instances_refcocog_val.json

Ref-Youtube-VOS

Download the dataset from the competition's website here. Then, extract and organize the file. We expect the directory structure to be the following:

ReferFormer
β”œβ”€β”€ data
β”‚   β”œβ”€β”€ ref-youtube-vos
β”‚   β”‚   β”œβ”€β”€ meta_expressions
β”‚   β”‚   β”œβ”€β”€ train
β”‚   β”‚   β”‚   β”œβ”€β”€ JPEGImages
β”‚   β”‚   β”‚   β”œβ”€β”€ Annotations
β”‚   β”‚   β”‚   β”œβ”€β”€ meta.json
β”‚   β”‚   β”œβ”€β”€ valid
β”‚   β”‚   β”‚   β”œβ”€β”€ JPEGImages

Ref-DAVIS17

Downlaod the DAVIS2017 dataset from the website. Note that you only need to download the two zip files DAVIS-2017-Unsupervised-trainval-480p.zip and DAVIS-2017_semantics-480p.zip. Download the text annotations from the website. Then, put the zip files in the directory as follows.

ReferFormer
β”œβ”€β”€ data
β”‚   β”œβ”€β”€ ref-davis
β”‚   β”‚   β”œβ”€β”€ DAVIS-2017_semantics-480p.zip
β”‚   β”‚   β”œβ”€β”€ DAVIS-2017-Unsupervised-trainval-480p.zip
β”‚   β”‚   β”œβ”€β”€ davis_text_annotations.zip

Unzip these zip files.

unzip -o davis_text_annotations.zip
unzip -o DAVIS-2017_semantics-480p.zip
unzip -o DAVIS-2017-Unsupervised-trainval-480p.zip

Preprocess the dataset to Ref-Youtube-VOS format. (Make sure you are in the main directory)

python tools/data/convert_davis_to_ytvos.py

Finally, unzip the file DAVIS-2017-Unsupervised-trainval-480p.zip again (since we use mv in preprocess for efficiency).

unzip -o DAVIS-2017-Unsupervised-trainval-480p.zip

A2D-Sentences

Follow the instructions and download the dataset from the website here. Then, extract the files. Additionally, we use the same json annotation files generated by MTTR. Please download these files from onedrive. We expect the directory structure to be the following:

ReferFormer
β”œβ”€β”€ data
β”‚   β”œβ”€β”€ a2d_sentences
β”‚   β”‚   β”œβ”€β”€ Release
β”‚   β”‚   β”œβ”€β”€ text_annotations
β”‚   β”‚   β”‚   β”œβ”€β”€ a2d_annotation_with_instances
β”‚   β”‚   β”‚   β”œβ”€β”€ a2d_annotation.txt
β”‚   β”‚   β”‚   β”œβ”€β”€ a2d_missed_videos.txt
β”‚   β”‚   β”œβ”€β”€ a2d_sentences_single_frame_test_annotations.json
β”‚   β”‚   β”œβ”€β”€ a2d_sentences_single_frame_train_annotations.json
β”‚   β”‚   β”œβ”€β”€ a2d_sentences_test_annotations_in_coco_format.json

JHMDB-Sentences

Follow the instructions and download the dataset from the website here. Then, extract the files. Additionally, we use the same json annotation files generated by MTTR. Please download these files from onedrive. We expect the directory structure to be the following:

ReferFormer
β”œβ”€β”€ data
β”‚   β”œβ”€β”€ jhmdb_sentences
β”‚   β”‚   β”œβ”€β”€ Rename_Images
β”‚   β”‚   β”œβ”€β”€ puppet_mask
β”‚   β”‚   β”œβ”€β”€ jhmdb_annotation.txt
β”‚   β”‚   β”œβ”€β”€ jhmdb_sentences_samples_metadata.json
β”‚   β”‚   β”œβ”€β”€ jhmdb_sentences_gt_annotations_in_coco_format.json