This repository is for ACL2020 paper Refer360: A Referring Expression Recognition Dataset in 360 Images. For annotation setup please take a look at the repo for frontend and the repo for backend.
Setup an anaconda environment first if you don't have one. Create conda environment refer360
and activate it.
conda create -n refer360 python=3.7 -y
source activate refer360
Run install.sh
to install packages to your environment.
Preprocess the json files for the simulator:
PYTHONPATH=.. python dump_data.py ../data/dumps all
We used SUN360 image database as our source of scenes. Please fill out this form to download the images. Move downloaded refer360images.tar.gz
to data/
extract via tar xvfz refer360images.tar.gz
where the folder structure should be as follows for scene types and categories under data/refer360images
:
$ tree -d
├── indoor
│ ├── bedroom
│ ├── expo_showroom
│ ├── living_room
│ ├── restaurant
│ ├── shop
└── outdoor
├── plaza_courtyard
└── street
Now you should be able to run the random agent under src
.
source activate refer360sim
PYTHONPATH=.. python random_agent.py
This command should generate an experiment folder src/exp-random
. To simulate what the random agent did pick a json file under that folder and run the simulate_history
.
PYTHONPATH=.. python simulate_history.py exp-random/samples/2100_randomagent.json 1 15
Till the agents actions are over use awsd
keys to observe what the agent did. After the agent's actions are over (there should be a blue prompt on the upper left frame), you can use awsd
to move the field of view to left,up,down,right. Press c
to close the window.
For supervised training use:
PYTHONPATH=.. python train_tf.py --help
For supervised training use:
PYTHONPATH=.. python train_rl.py --help
To see the list of command line options. This README will be updated with more instructions for different kinds of experiments soon.
- Update readme with
- Each the usage branch's