A shared task on Detecting Signs of Depression from Social Media Text at LT-EDI 2022, ACL 2022 Workshop. We won the 🔥second place🔥 and the paper is available at here. The brief introduction of this work can be referred to our blog.
Given social media postings in English, the system should classify the signs of depression into three labels namely “not depressed”, “moderately depressed”, and “severely depressed”.
-
Method 1: Gradient Boosting Models + VAD Score
- Add sentiment features by VADER (preprocessing/)
python add_feature.py --preprocessing {boolean}
- Train model (ml/)
python sentiment_features_classifier.py --embedding {name} --model {name}
- Add sentiment features by VADER (preprocessing/)
-
Method 2: Pre-trained Language Models
- Train model
python3 main.py --model_type [roberta/electra/deberta]
- Ensemble and evaluate (for dev and test)
python3 ensemble.py --path [file path] --mode [dev/test]
- Train model
-
Method 3: Pre-trained Language Models + VAD Score + Supervised Contrastive Learning (plm_scl/)
- Train model
python main.py {pre-trained name}
- Evaluate model
You need to modify {MODEL} and {MODEL_NAME} to your pre-trained model and corresponding path in
python evaluate.py
evaluate.py
.
- Train model
-
Power Weighted Sum
python ensemble.py
The dataset comprises training, development and test set. The data files are in Tab Separated Values (tsv) format with three columns namely posting_id (pid), text data and label.
Tran | Dev | Test | |
---|---|---|---|
Not depressed | 1,971 | 1,830 | |
Moderate | 6,019 | 2,306 | |
Severe | 901 | 360 | |
Total | 8,891 | 4,496 | 3,245 |
Performance will be measured in terms of macro averaged Precision, macro averaged Recall and macro averaged F1-Score across all the classes.
We report the hyper-parameters of each method as follows.
- Method 1: Gradient Boosting Models + VAD Score
- General
- Pretrained Sentence Embedding: MPNet
- LightGBM
LR num_leaves n_estimators max_depth 0.5 64 70 9 - XGBoost
LR gamma n_estimators max_depth subsample 0.1 0.02 100 6 0.98
- General
- Method 2: Pre-trained Language Models
- General
LR Epochs 2e-5 20 - RoBERTa
Seed Warm Up Batch Size 13 4 3 - DeBERTa
Seed Warm Up Batch Size 49 8 6 - ELECTRA
Seed Warm Up Batch Size 17 5 2
- General
- Method 3: Pre-trained Language Models + VAD Score + Supervised Contrastive Learning
Epochs LR Batch Size Seed Warmup Steps Hidden Dimension Dropout Lambda_{ce} Lambda_{scl} 20 4e-5 8 17 5 512 0.1 0.7 0.3 - Power Weighted Sum
- ensemble_weight: [1, 0.67, 0.69]
- power: 4
If you use our dataset or find our work is relevant to your research, please cite:
@inproceedings{wang-etal-2022-nycu,
title = "{NYCU}{\_}{TWD}@{LT}-{EDI}-{ACL}2022: Ensemble Models with {VADER} and Contrastive Learning for Detecting Signs of Depression from Social Media",
author = "Wang, Wei-Yao and
Tang, Yu-Chien and
Du, Wei-Wei and
Peng, Wen-Chih",
booktitle = "Proceedings of the Second Workshop on Language Technology for Equality, Diversity and Inclusion",
month = may,
year = "2022",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.ltedi-1.15",
doi = "10.18653/v1/2022.ltedi-1.15",
pages = "136--139",
}