Update README.md
Browse files
README.md
CHANGED
|
@@ -161,6 +161,27 @@ python3 scripts/convert_bert_from_uer_to_huggingface.py --input_model_path model
|
|
| 161 |
### BibTeX entry and citation info
|
| 162 |
|
| 163 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 164 |
@article{zhao2019uer,
|
| 165 |
title={UER: An Open-Source Toolkit for Pre-training Models},
|
| 166 |
author={Zhao, Zhe and Chen, Hui and Zhang, Jinbin and Zhao, Xin and Liu, Tao and Lu, Wei and Chen, Xi and Deng, Haotang and Ju, Qi and Du, Xiaoyong},
|
|
|
|
| 161 |
### BibTeX entry and citation info
|
| 162 |
|
| 163 |
```
|
| 164 |
+
@article{devlin2018bert,
|
| 165 |
+
title={Bert: Pre-training of deep bidirectional transformers for language understanding},
|
| 166 |
+
author={Devlin, Jacob and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
|
| 167 |
+
journal={arXiv preprint arXiv:1810.04805},
|
| 168 |
+
year={2018}
|
| 169 |
+
}
|
| 170 |
+
|
| 171 |
+
@article{liu2019roberta,
|
| 172 |
+
title={Roberta: A robustly optimized bert pretraining approach},
|
| 173 |
+
author={Liu, Yinhan and Ott, Myle and Goyal, Naman and Du, Jingfei and Joshi, Mandar and Chen, Danqi and Levy, Omer and Lewis, Mike and Zettlemoyer, Luke and Stoyanov, Veselin},
|
| 174 |
+
journal={arXiv preprint arXiv:1907.11692},
|
| 175 |
+
year={2019}
|
| 176 |
+
}
|
| 177 |
+
|
| 178 |
+
@article{turc2019,
|
| 179 |
+
title={Well-Read Students Learn Better: On the Importance of Pre-training Compact Models},
|
| 180 |
+
author={Turc, Iulia and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
|
| 181 |
+
journal={arXiv preprint arXiv:1908.08962v2 },
|
| 182 |
+
year={2019}
|
| 183 |
+
}
|
| 184 |
+
|
| 185 |
@article{zhao2019uer,
|
| 186 |
title={UER: An Open-Source Toolkit for Pre-training Models},
|
| 187 |
author={Zhao, Zhe and Chen, Hui and Zhang, Jinbin and Zhao, Xin and Liu, Tao and Lu, Wei and Chen, Xi and Deng, Haotang and Ju, Qi and Du, Xiaoyong},
|