S2-FPN: Scale-ware Strip Attention Guided Feature Pyramid Network for Real-time Semantic Segmentation
This project contains the Pytorch implementation for the proposed S2-FPN: [Arxiv].
Modern high-performance semantic segmentation methods employ a heavy backbone and dilated convolution to extract the relevant feature. Although extracting features with both contextual and semantic information is critical for the segmentation tasks, it brings a memory footprint and high computation cost for real-time applications. This paper presents a new model to achieve a trade-off between accuracy/speed for real-time road scene semantic segmentation. Specifially, we proposed a lightweight model named Scale-aware Strip Attention Guided Feature Pyramid Network (S2-FPN). Our network consists of three main modules: Attention Pyramid Fusion (APF) module, Scale-aware Strip Attention Module (SSAM), and Global Feature Upsample (GFU) module. APF adopts an attention mechanisms to learn discriminative multi-scale features and help close the semantic gap between different levels. APF uses the scale-aware attention to encode global context with vertical stripping operation and models the long-range dependencies, which helps relate pixels with similar semantic label. In addition, APF employs channel-wise reweighting block (CRB) to emphasize the channel features. Finally, the decoder of S2-FPN then adopts GFU, which is used to fuse features from APF and the encoder. Extensive experiments have been conducted on two challenging semantic segmentation benchmarks, which demonstrate that our approach achieves better accuracy/speed trade-off with different model settings. The proposed models have achieved a results of 76.2%mIoU/87.3FPS, 77.4%mIoU/67FPS, and 77.8%mIoU/30.5FPS on Cityscapes dataset, and 69.6%mIoU,71.0% mIoU,and 74.2% mIoU on Camvid dataset. The detailed architecture of S2-FPN
Attention Pyramid Fusion Module
-
S2FPN works with [ResNet18,ResNet34,ResNet50,ResNet101,ResNet152]. We tested it with ResNet18 and 34.
-
upload pretrained weights
-
18/10/2023 we updated the scale-aware attention block and updated with new model. Now you can use the new models as S2FPN and the old one is changed to S2FPNv1
-
Our paper was marked as state of the art in Papers with Code.
- Pyroch Environment
- Env: Python 3.6; PyTorch 1.0; CUDA 10.1; cuDNN V8
- Install some packages
pip install opencv-python pillow numpy matplotlib
- Clone this repository
git clone https://github.com/mohamedac29/S2-FPN
cd S2-FPN
- Dataset
You need to download the Cityscapes, and put the files in the dataset
folder with following structure.
├── cityscapes_test_list.txt
├── cityscapes_train_list.txt
├── cityscapes_trainval_list.txt
├── cityscapes_val_list.txt
├── cityscapes_val.txt
├── gtCoarse
│ ├── train
│ ├── train_extra
│ └── val
├── gtFine
│ ├── test
│ ├── train
│ └── val
├── leftImg8bit
│ ├── test
│ ├── train
│ └── val
-
Convert the dataset to 19 categories.
-
Download the Camvid dataset and put the files in the following structure.
├── camvid_test_list.txt
├── camvid_train_list.txt
├── camvid_trainval_list.txt
├── camvid_val_list.txt
├── test
├── testannot
├── train
├── trainannot
├── val
└── valannot
You can download the pretrained weights. There are some differences in the accuracy listed here
- Camvid and Cityscapes Datasets. FPS computed based on GTX1080Ti
Model | Dataset | Pretrained | Train type | test (mIoU) | FPS | model |
---|---|---|---|---|---|---|
S2FPN18 | Cityscapes | ResNet18 | train | 76.7% | 84.2 | ckpt |
S2FPN34 | Cityscapes | ResNet34 | train | 77.4% | 64.4 | ckpt |
S2FPN34M | Cityscapes | ResNet34M | train | 78.2% | 28.5 | |
S2FPN18 | CamVid | ResNet18 | trainval | 70.4% | 122.2 | ckpt |
S2FPN34 | CamVid | ResNet34 | trainval | 71.2% | 104.2 | ckpt |
S2FPN34M | CamVid | ResNet34M | trainval | 74.5% | 53.4 | ckpt |
You can download the pretrained weights. There are some differences in the accuracy listed here
- Camvid and Cityscapes Datasets. FPS computed based on GTX1080Ti
Dataset | Pretrained | Train type | mIoU | FPS | model |
---|---|---|---|---|---|
Cityscapes | ResNet18v1 | train | 76.3% | 87.3 | ckpt |
Cityscapes | ResNet34 | ||||
Cityscapes | ResNet34M | ||||
CamVid | ResNet18 | trainval | 70.1% | 124.2 | ckpt |
CamVid | ResNet34 | trainval | 71.0% | 107.2 | |
CamVid | ResNet34M | trainval | 74.76% | 55.5 | ckpt |
- Training on Camvid datsaset
python train.py --dataset camvid --model SSFPN --max_epochs 150 --train_type trainval --lr 3e-4 --batch_size 8
- Training on Camvid datsaset - train_type [trainval,trainval]
python train.py --dataset cityscapes --model SSFPN --max_epochs 500 --train_type trainval --lr 3e-4 --batch_size 8
- Testing on Camvid datsaset
python test.py --dataset camvid --model SSFPN --checkpoint ./checkpoint/camvid/SSFPNbs8gpu1_trainval/model_150.pth --gpus 0
- Testing on Cityscapes datsaset
python test.py --dataset cityscapes --model SSFPN --checkpoint ./checkpoint/cityscapes/SSFPNbs8gpu1_trainval/model_500.pth --gpus 0
- Inference speed with input resolution 512x1024
python eval_fps.py 512,1024
If you find this work useful in your research, please consider citing.
@article{elhassan2022s,
title={S2-FPN: Scale-ware Strip Attention Guided Feature Pyramid Network for Real-time Semantic Segmentation},
author={Elhassan, Mohammed AM and Yang, Chenhui and Huang, Chenxi and Legesse Munea, Tewodros and Hong, Xin},
journal={arXiv e-prints},
pages={arXiv--2206},
year={2022}
}