Migrate model card from transformers-repo
Browse filesRead announcement at https://discuss.huggingface.co/t/announcement-all-model-cards-will-be-migrated-to-hf-co-model-repos/2755
Original file history: https://github.com/huggingface/transformers/commits/master/model_cards/deepset/gelectra-large/README.md
README.md
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
language: de
|
| 3 |
+
license: mit
|
| 4 |
+
datasets:
|
| 5 |
+
- wikipedia
|
| 6 |
+
- OPUS
|
| 7 |
+
- OpenLegalData
|
| 8 |
+
- oscar
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
# German ELECTRA large
|
| 12 |
+
|
| 13 |
+
Released, Oct 2020, this is a German ELECTRA language model trained collaboratively by the makers of the original German BERT (aka "bert-base-german-cased") and the dbmdz BERT (aka bert-base-german-dbmdz-cased). In our [paper](https://arxiv.org/pdf/2010.10906.pdf), we outline the steps taken to train our model and show that this is the state of the art German language model.
|
| 14 |
+
|
| 15 |
+
## Overview
|
| 16 |
+
**Paper:** [here](https://arxiv.org/pdf/2010.10906.pdf)
|
| 17 |
+
**Architecture:** ELECTRA large (discriminator)
|
| 18 |
+
**Language:** German
|
| 19 |
+
|
| 20 |
+
## Performance
|
| 21 |
+
```
|
| 22 |
+
GermEval18 Coarse: 80.70
|
| 23 |
+
GermEval18 Fine: 55.16
|
| 24 |
+
GermEval14: 88.95
|
| 25 |
+
```
|
| 26 |
+
|
| 27 |
+
See also:
|
| 28 |
+
deepset/gbert-base
|
| 29 |
+
deepset/gbert-large
|
| 30 |
+
deepset/gelectra-base
|
| 31 |
+
deepset/gelectra-large
|
| 32 |
+
deepset/gelectra-base-generator
|
| 33 |
+
deepset/gelectra-large-generator
|
| 34 |
+
|
| 35 |
+
## Authors
|
| 36 |
+
Branden Chan: `branden.chan [at] deepset.ai`
|
| 37 |
+
Stefan Schweter: `stefan [at] schweter.eu`
|
| 38 |
+
Timo Möller: `timo.moeller [at] deepset.ai`
|
| 39 |
+
|
| 40 |
+
## About us
|
| 41 |
+

|
| 42 |
+
|
| 43 |
+
We bring NLP to the industry via open source!
|
| 44 |
+
Our focus: Industry specific language models & large scale QA systems.
|
| 45 |
+
|
| 46 |
+
Some of our work:
|
| 47 |
+
- [German BERT (aka "bert-base-german-cased")](https://deepset.ai/german-bert)
|
| 48 |
+
- [FARM](https://github.com/deepset-ai/FARM)
|
| 49 |
+
- [Haystack](https://github.com/deepset-ai/haystack/)
|
| 50 |
+
|
| 51 |
+
Get in touch:
|
| 52 |
+
[Twitter](https://twitter.com/deepset_ai) | [LinkedIn](https://www.linkedin.com/company/deepset-ai/) | [Website](https://deepset.ai)
|