Update README.md
Browse files
README.md
CHANGED
@@ -40,6 +40,8 @@ inference:
|
|
40 |
- [Citing information](#citing-information)
|
41 |
- [Disclaimer](#disclaimer)
|
42 |
|
|
|
|
|
43 |
## Model description
|
44 |
RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019.
|
45 |
|
@@ -97,7 +99,7 @@ Intelligence (SEDIA) within the framework of the Plan-TL.},
|
|
97 |
}
|
98 |
```
|
99 |
|
100 |
-
|
101 |
|
102 |
The models published in this repository are intended for a generalist purpose and are available to third parties. These models may have bias and/or any other undesirable distortions.
|
103 |
|
|
|
40 |
- [Citing information](#citing-information)
|
41 |
- [Disclaimer](#disclaimer)
|
42 |
|
43 |
+
</details>
|
44 |
+
|
45 |
## Model description
|
46 |
RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019.
|
47 |
|
|
|
99 |
}
|
100 |
```
|
101 |
|
102 |
+
### Disclaimer
|
103 |
|
104 |
The models published in this repository are intended for a generalist purpose and are available to third parties. These models may have bias and/or any other undesirable distortions.
|
105 |
|