Update README.md
Browse files
README.md
CHANGED
@@ -16,11 +16,12 @@ The labels explanation:
|
|
16 |
- `relevance`: is the last message in the dialogue relevant in the context of the full dialogue
|
17 |
- `specificity`: is the last message in the dialogue interesting and promotes the continuation of the dialogue
|
18 |
|
19 |
-
The preferable length of the dialogue is 4 where the last message is needed to be estimated
|
20 |
-
|
21 |
It is pretrained on a large corpus of dialog data in unsupervised manner: the model is trained to predict whether last response was in a real dialog, or it was pulled from some other dialog at random.
|
22 |
|
23 |
Then it was finetuned on manually labelled examples (dataset will be posted soon).
|
|
|
|
|
|
|
24 |
It is pretrained on corpus of dialog data and finetuned on [tinkoff-ai/context_similarity](https://huggingface.co/tinkoff-ai/context_similarity).
|
25 |
The performance of the model on validation split (dataset will be posted soon)[tinkoff-ai/context_similarity](https://huggingface.co/tinkoff-ai/context_similarity) (with the best thresholds for validation samples):
|
26 |
|
@@ -47,5 +48,6 @@ with torch.inference_mode():
|
|
47 |
relevance, specificity = probas
|
48 |
```
|
49 |
|
50 |
-
The [app]
|
|
|
51 |
The work was done during internship at Tinkoff by [egoriyaa](https://github.com/egoriyaa)
|
|
|
16 |
- `relevance`: is the last message in the dialogue relevant in the context of the full dialogue
|
17 |
- `specificity`: is the last message in the dialogue interesting and promotes the continuation of the dialogue
|
18 |
|
|
|
|
|
19 |
It is pretrained on a large corpus of dialog data in unsupervised manner: the model is trained to predict whether last response was in a real dialog, or it was pulled from some other dialog at random.
|
20 |
|
21 |
Then it was finetuned on manually labelled examples (dataset will be posted soon).
|
22 |
+
|
23 |
+
The model was trained with the dialogue length 4 where the last message is needed to be estimated. Each message in the dialogue was tokenized separately with ``` max_length = max_seq_length // 4 ```.
|
24 |
+
|
25 |
It is pretrained on corpus of dialog data and finetuned on [tinkoff-ai/context_similarity](https://huggingface.co/tinkoff-ai/context_similarity).
|
26 |
The performance of the model on validation split (dataset will be posted soon)[tinkoff-ai/context_similarity](https://huggingface.co/tinkoff-ai/context_similarity) (with the best thresholds for validation samples):
|
27 |
|
|
|
48 |
relevance, specificity = probas
|
49 |
```
|
50 |
|
51 |
+
The [app](https://huggingface.co/spaces/tinkoff-ai/response-quality-classifiers) where you can easily evaluate this model
|
52 |
+
|
53 |
The work was done during internship at Tinkoff by [egoriyaa](https://github.com/egoriyaa)
|