Skip to content

Commit 2bd1a66

Browse files
authored
Merge pull request #20 from OpenGVLab/release
Release
2 parents 7835690 + bfad364 commit 2bd1a66

File tree

3 files changed

+8
-7
lines changed

3 files changed

+8
-7
lines changed

VisionLLMv2/docs/data_det.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -155,7 +155,7 @@ data/reasonseg
155155

156156
### COCO
157157

158-
Follow the instructions below to prepare the data:
158+
Follow the instructions below to prepare the data (We follow the evaluation from [PSALM](https://github.com/zamling/PSALM)):
159159

160160
```
161161
# Step 1: Create the data directory
@@ -165,7 +165,7 @@ mkdir -p data/coco && cd data/coco
165165
wget http://images.cocodataset.org/zips/train2017.zip && unzip train2017.zip
166166
wget http://images.cocodataset.org/zips/val2017.zip && unzip val2017.zip
167167
168-
# Step 3: Download and place the annotation files
168+
# Step 3: Download and place the annotation files from PSALM
169169
# Download the annotation files from official website https://drive.google.com/file/d/1EcC1tl1OQRgIqqy7KFG7JZz2KHujAQB3/view
170170
171171
cd ../..

VisionLLMv2/docs/eval_region-vqa.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@ huggingface-cli download --resume-download --local-dir-use-symlinks False senten
4444
cd ..
4545
```
4646

47-
Specify the subset or full set (`lvis`, `paco`) you would like to evaluate in [visionllmv2/eval/eval_region_classification.py](https://github.com/OpenGVLab/VisionLLM/blob/7befe44a38f874fba6835445dbd0177f0b6b46d9/VisionLLMv2/visionllmv2/eval/eval_region_classification.py#L381).
47+
Specify the datasets (`lvis`, `paco`) you would like to evaluate in [visionllmv2/eval/eval_region_classification.py](https://github.com/OpenGVLab/VisionLLM/blob/7befe44a38f874fba6835445dbd0177f0b6b46d9/VisionLLMv2/visionllmv2/eval/eval_region_classification.py#L381).
4848

4949
```
5050
GPUS=8 bash scripts/vllmv2_7b/eval/dist_eval_region_classification.sh work_dirs/VisionLLMv2

VisionLLMv2/docs/install.md

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -37,12 +37,13 @@ conda activate vllmv2
3737
pip install torch==2.0.1 torchvision==0.15.2 torchaudio==2.0.2 --index-url https://download.pytorch.org/whl/cu118
3838
```
3939

40-
Then, please refer to [install.sh](https://github.com/OpenGVLab/VisionLLM/blob/release/VisionLLMv2/docs/install.md) to install the necessary packages step by step.
40+
Then, please refer to [install.sh](https://github.com/OpenGVLab/VisionLLM/blob/release/VisionLLMv2/install.sh) to install the necessary packages step by step.
4141

42-
- Additional:
42+
- Additionally:
4343

44-
`pycocoevalcap` is used to evaluate the metrics for image/region captioning. You can install it by yourself.
45-
For your convenience, you can directly download it and unzip the file.
44+
`pycocoevalcap` is used to evaluate the metrics for image/region captioning. You can install it by yourself.
45+
46+
For your convenience, we provide the full folder of pycocoevalcap. You can directly download and use it with the following commands.
4647
```
4748
wget https://huggingface.co/OpenGVLab/VisionLLMv2/resolve/main/data/pycocoevalcap.zip
4849
unzip -qq pycocoevalcap.zip

0 commit comments

Comments
 (0)