You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: model_cards/deepset/roberta-base-squad2/README.md
+16-7Lines changed: 16 additions & 7 deletions
Original file line number
Diff line number
Diff line change
@@ -5,7 +5,7 @@ datasets:
5
5
6
6
# roberta-base for QA
7
7
8
-
NOTE: This model has been superseded by deepset/roberta-base-squad2-v2. For an explanation of why, see [this github issue](https://github.com/deepset-ai/FARM/issues/552) from the FARM repository.
8
+
NOTE: This is version 2 of the model. See [this github issue](https://github.com/deepset-ai/FARM/issues/552) from the FARM repository for an explanation of why we updated. If you'd like to use version 1, specify `revision="v1.0"` when loading the model in Transformers 3.5.
9
9
10
10
## Overview
11
11
**Language model:** roberta-base
@@ -19,10 +19,10 @@ NOTE: This model has been superseded by deepset/roberta-base-squad2-v2. For an e
19
19
## Hyperparameters
20
20
21
21
```
22
-
batch_size = 50
23
-
n_epochs = 3
22
+
batch_size = 96
23
+
n_epochs = 2
24
24
base_LM_model = "roberta-base"
25
-
max_seq_len = 384
25
+
max_seq_len = 386
26
26
learning_rate = 3e-5
27
27
lr_schedule = LinearWarmup
28
28
warmup_proportion = 0.2
@@ -32,9 +32,18 @@ max_query_length=64
32
32
33
33
## Performance
34
34
Evaluated on the SQuAD 2.0 dev set with the [official eval script](https://worksheets.codalab.org/rest/bundles/0x6b567e1cf2e041ec80d7098f031c5c9e/contents/blob/).
35
+
35
36
```
36
-
"exact": 78.49743114629833,
37
-
"f1": 81.73092721240889
37
+
"exact": 79.97136359807968
38
+
"f1": 83.00449234495325
39
+
40
+
"total": 11873
41
+
"HasAns_exact": 78.03643724696356
42
+
"HasAns_f1": 84.11139298441825
43
+
"HasAns_total": 5928
44
+
"NoAns_exact": 81.90075693860386
45
+
"NoAns_f1": 81.90075693860386
46
+
"NoAns_total": 5945
38
47
```
39
48
40
49
## Usage
@@ -85,7 +94,7 @@ For doing QA at scale (i.e. many docs instead of single paragraph), you can load
0 commit comments