Skip to content
/ HTT Public

[AAAI 2024] Heterogeneous Test-time Training for Multi-modal Person Re-identification

License

Notifications You must be signed in to change notification settings

ziwang1121/HTT

Repository files navigation

This is the official codes of our paper Heterogeneous Test-Time Training for Multi-Modal Person Re-identification Paper Link accepted by AAAI 2024.

Overall Introduction

Inspired by the development of Test-time Training (TTT), we propose a novel method for the multi-modal person ReID task, termed Heterogeneous Test-time Training (HTT), to improve performance on unseen test data by utilizing the relationship between heterogeneous modalities and fine-tuning the network before inference.


(a) Traditional multi-modal training exclusively utilizes labeled training data. (b) The proposed heterogeneous test-time training additionally leverages unlabeled test data for optimization.

Environment

Please follow the previous works TransReID (ICCV 2021) and IEEE (AAAI 2022).

Datasets

Multi-modality Person ReID Dataset: RGBNT201 (Paper Link, Download Link)

Multi-modality Vehicle ReID Datasets: RGBNT100 & RGBN300 (GitHub)

Network

Overview of network

Highlight

  1. Cross-identity inter-modal loss (CIM loss) code

    CE loss separates features from two identities. Triplet loss controls the distance of features. 3M loss constrains the distance among modalities in each sample. CIM loss constrains the distance between inter-modal features from different identities.

  1. Multi-modal Test-time Training (MTT) code

Train

# train RGBN300
python train.py --config_file configs/RGBN300/vit_base.yml
# train RGBNT100
python train.py --config_file configs/RGBNT100/vit_base.yml
# train RGBNT201
python train.py --config_file configs/RGBNT201/vit_base.yml

Test-time Training

# test-time training RGBN300
python test_time_train.py --config_file configs/RGBN300/vit_base_ttt.yml
# test-time training RGBNT100
python test_time_train.py --config_file configs/RGBNT100/vit_base_ttt.yml
# test-time training RGBNT201
python test_time_train.py --config_file configs/RGBNT201/vit_base_ttt.yml

Test

At the end of each epoch of training and test-time training, a test will be performed, and the test results can be directly seen.

Or you can run the following code to test, and note that the test model path needs to be changed.

# test RGBN300
python test.py --config_file configs/RGBN300/vit_base_ttt.yml
# test RGBNT100
python test.py --config_file configs/RGBNT100/vit_base_ttt.yml
# test RGBNT201
python test.py --config_file configs/RGBNT201/vit_base_ttt.yml

Results

Citation and Contact

@inproceedings{wang2024heterogeneous,
  title={Heterogeneous Test-Time Training for Multi-Modal Person Re-identification},
  author={Wang, Zi and Huang, Huaibo and Zheng, Aihua and He, Ran},
  booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
  volume={38},
  number={6},
  pages={5850--5858},
  year={2024}
}

If you have any questions about the project or are interested in multi-modal ReID, please feel free to contact me (ziwang1121@foxmail.com).

About

[AAAI 2024] Heterogeneous Test-time Training for Multi-modal Person Re-identification

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages