Skip to content

Latest commit

 

History

History
31 lines (23 loc) · 1.22 KB

README.md

File metadata and controls

31 lines (23 loc) · 1.22 KB

Google quest challenge


solution to the competition https://www.kaggle.com/c/google-quest-challenge. 240th place with no post-processing. With post-processing, the output of the model could be in the silver medal zone.

Overview

Task

Computers are good at answering questions with single, verifiable answers. But humans are often still better at answering questions about opinions, recommendations, or personal experiences. The task is predictive modeling for natural language understanding about intents of questions, and how they answered, i.e. helpful, interesting, or well-written. Evaluated by Spearman's correlation coefficient.

Networks

  • BERT
  • Pooling and then concatenation of last few hidden states from BERT
  • Multi-sample dropout

Data

  • 5-fold group k-fold split
  • Augmentation with multiple truncations of the sequence (head, tail, mix)
  • TTA

Optimizer

  • AdamW with linear warmup and linear annealing learning rate schedule
  • BCE Loss for 30 multiple soft labels

Environment

nvidia tesla p100
nvidia apex
pytorch 1.2.0
transformers 2.3.0