@@ -9,10 +9,10 @@ We strictly follow the optimizer configuration as in HuggingFace and use batch s
9
9
| -------| ------- | :---------: | :------: | :--: |
10
10
| HuggingFace Default (bert-base-cased)| Test Set | 90.71 | 92.04| 91.37|
11
11
| HuggingFace Default (roberta-base)* | Test Set | 89.41 | 91.47| 90.43|
12
- | BERT-base-cased (this repo)| Test set | 91.69 | 92.05 | 91.87 |
13
- | BERT-large-cased (this repo)| Test Set | 92.03 | 92.17 | 92.10 |
14
- | Roberta-base (this repo)| Test Set | 91.88 | 93.01 | 92.44|
15
- | Roberta-large (this repo)| Test Set | ** 92.27** | ** 93.18** | ** 92.72** |
12
+ | BERT-base-cased + CRF (this repo)| Test set | 91.69 | 92.05 | 91.87 |
13
+ | BERT-large-cased + CRF (this repo)| Test Set | 92.03 | 92.17 | 92.10 |
14
+ | Roberta-base + CRF (this repo)| Test Set | 91.88 | 93.01 | 92.44|
15
+ | Roberta-large + CRF (this repo)| Test Set | ** 92.27** | ** 93.18** | ** 92.72** |
16
16
HuggingFace Default (roberta-base)* has an issue with tokenization (There is no leading space).
17
17
18
18
We didn't achieve 92.4 F1 as reported in the BERT paper.
@@ -25,8 +25,8 @@ I think one of the main reasons is they are using the document-level dataset ins
25
25
26
26
| Model| Dataset | Precision | Recall | F1 |
27
27
| -------| ------- | :---------: | :------: | :--: |
28
- | BERT-base-cased (this repo)| Test Set | 89.57 | 89.45 | 89.51 |
29
- | BERT-large-cased (this repo)* | Test Set | - | -| -|
30
- | Roberta-base (this repo)| Test Set | ** 90.12** | ** 91.25** | ** 90.68** |
28
+ | BERT-base-cased + CRF (this repo)| Test Set | 89.57 | 89.45 | 89.51 |
29
+ | BERT-large-cased + CRF (this repo)* | Test Set | - | -| -|
30
+ | Roberta-base + CRF (this repo)| Test Set | ** 90.12** | ** 91.25** | ** 90.68** |
31
31
32
32
Roberta-base (this repo)* is still running. The others are not finished yet.
0 commit comments