Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,23 @@
|
|
1 |
---
|
2 |
license: mit
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: mit
|
3 |
---
|
4 |
+
|
5 |
+
[joeddav/distilbert-base-uncased-go-emotions-student](https://huggingface.co/joeddav/distilbert-base-uncased-go-emotions-student) converted to ONNX and quantized using optimum.
|
6 |
+
|
7 |
+
---
|
8 |
+
|
9 |
+
# distilbert-base-uncased-go-emotions-student
|
10 |
+
|
11 |
+
## Model Description
|
12 |
+
|
13 |
+
This model is distilled from the zero-shot classification pipeline on the unlabeled GoEmotions dataset using [this
|
14 |
+
script](https://github.com/huggingface/transformers/tree/master/examples/research_projects/zero-shot-distillation).
|
15 |
+
It was trained with mixed precision for 10 epochs and otherwise used the default script arguments.
|
16 |
+
|
17 |
+
## Intended Usage
|
18 |
+
|
19 |
+
The model can be used like any other model trained on GoEmotions, but will likely not perform as well as a model
|
20 |
+
trained with full supervision. It is primarily intended as a demo of how an expensive NLI-based zero-shot model
|
21 |
+
can be distilled to a more efficient student, allowing a classifier to be trained with only unlabeled data. Note
|
22 |
+
that although the GoEmotions dataset allow multiple labels per instance, the teacher used single-label
|
23 |
+
classification to create psuedo-labels.
|