egoriya commited on
Commit
799292a
1 Parent(s): 84a19e3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -3
README.md CHANGED
@@ -16,11 +16,12 @@ The labels explanation:
16
  - `relevance`: is the last message in the dialogue relevant in the context of the full dialogue
17
  - `specificity`: is the last message in the dialogue interesting and promotes the continuation of the dialogue
18
 
19
- The preferable length of the dialogue is 4 where the last message is needed to be estimated
20
-
21
  It is pretrained on a large corpus of dialog data in unsupervised manner: the model is trained to predict whether last response was in a real dialog, or it was pulled from some other dialog at random.
22
 
23
  Then it was finetuned on manually labelled examples (dataset will be posted soon).
 
 
 
24
  It is pretrained on corpus of dialog data and finetuned on [tinkoff-ai/context_similarity](https://huggingface.co/tinkoff-ai/context_similarity).
25
  The performance of the model on validation split (dataset will be posted soon)[tinkoff-ai/context_similarity](https://huggingface.co/tinkoff-ai/context_similarity) (with the best thresholds for validation samples):
26
 
@@ -47,5 +48,6 @@ with torch.inference_mode():
47
  relevance, specificity = probas
48
  ```
49
 
50
- The [app] (https://huggingface.co/spaces/tinkoff-ai/response-quality-classifiers) where you can easily evaluate this model
 
51
  The work was done during internship at Tinkoff by [egoriyaa](https://github.com/egoriyaa)
 
16
  - `relevance`: is the last message in the dialogue relevant in the context of the full dialogue
17
  - `specificity`: is the last message in the dialogue interesting and promotes the continuation of the dialogue
18
 
 
 
19
  It is pretrained on a large corpus of dialog data in unsupervised manner: the model is trained to predict whether last response was in a real dialog, or it was pulled from some other dialog at random.
20
 
21
  Then it was finetuned on manually labelled examples (dataset will be posted soon).
22
+
23
+ The model was trained with the dialogue length 4 where the last message is needed to be estimated. Each message in the dialogue was tokenized separately with ``` max_length = max_seq_length // 4 ```.
24
+
25
  It is pretrained on corpus of dialog data and finetuned on [tinkoff-ai/context_similarity](https://huggingface.co/tinkoff-ai/context_similarity).
26
  The performance of the model on validation split (dataset will be posted soon)[tinkoff-ai/context_similarity](https://huggingface.co/tinkoff-ai/context_similarity) (with the best thresholds for validation samples):
27
 
 
48
  relevance, specificity = probas
49
  ```
50
 
51
+ The [app](https://huggingface.co/spaces/tinkoff-ai/response-quality-classifiers) where you can easily evaluate this model
52
+
53
  The work was done during internship at Tinkoff by [egoriyaa](https://github.com/egoriyaa)