ai-forever's picture
Update README.md
46539f6
|
raw
history blame
553 Bytes
---
language:
- en
datasets:
- conll2003
- wnut_17
- jnlpba
- conll2012
- BTC
tags:
- PyTorch
---
# BERT base uncased model pre-trained on 5 NER datasets
Model was trained by [SberIDP](https://github.com)
* Task: `NER`
* Training Data is 5 datasets: CoNLL-2003, WNUT17, JNLPBA, CoNLL-2012 (OntoNotes), BTC
The model is described [in this article](https://habr.com/ru/company/sberbank/blog/).
It is pretrained for NER task using [Reptile](https://openai.com/blog/reptile/) and can be finetuned for new entities with only a small amount of samples.