egoriya commited on
Commit
ea35c2a
1 Parent(s): a8f8dcb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -8
README.md CHANGED
@@ -17,10 +17,9 @@ The labels explanation:
17
  - `specificity`: is the last message in the dialogue interesting and promotes the continuation of the dialogue.
18
 
19
  It is pretrained on a large corpus of dialog data in unsupervised manner: the model is trained to predict whether last response was in a real dialog, or it was pulled from some other dialog at random.
20
-
21
  Then it was finetuned on manually labelled examples (dataset will be posted soon).
22
 
23
- The model was trained with the dialogue length 4 where the last message is needed to be estimated. Each message in the dialogue was tokenized separately with ``` max_length = max_seq_length // 4 ```.
24
 
25
  The performance of the model on validation split (dataset will be posted soon) (with the best thresholds for validation samples):
26
 
@@ -34,12 +33,12 @@ The performance of the model on validation split (dataset will be posted soon) (
34
  How to use:
35
 
36
  ```python
37
- # pip install transformers
38
  from transformers import AutoTokenizer, AutoModelForSequenceClassification
39
  import torch
40
- tokenizer = AutoTokenizer.from_pretrained("tinkoff-ai/response-quality-classifier-tiny")
41
- model = AutoModelForSequenceClassification.from_pretrained("tinkoff-ai/response-quality-classifier-tiny")
42
- # model.cuda()
43
  inputs = tokenizer('[CLS]привет[SEP]привет![SEP]как дела?[RESPONSE_TOKEN]норм, у тя как?', max_length=128, add_special_tokens=False, return_tensors='pt')
44
  with torch.inference_mode():
45
  logits = model(**inputs).logits
@@ -47,6 +46,6 @@ with torch.inference_mode():
47
  relevance, specificity = probas
48
  ```
49
 
50
- The [app](https://huggingface.co/spaces/tinkoff-ai/response-quality-classifiers) where you can easily evaluate this model.
51
 
52
- The work was done during internship at Tinkoff by [egoriyaa](https://github.com/egoriyaa).
 
17
  - `specificity`: is the last message in the dialogue interesting and promotes the continuation of the dialogue.
18
 
19
  It is pretrained on a large corpus of dialog data in unsupervised manner: the model is trained to predict whether last response was in a real dialog, or it was pulled from some other dialog at random.
 
20
  Then it was finetuned on manually labelled examples (dataset will be posted soon).
21
 
22
+ The model was trained with three messages in the context and one response. Each message was tokenized separately with ``` max_length = 32 ```.
23
 
24
  The performance of the model on validation split (dataset will be posted soon) (with the best thresholds for validation samples):
25
 
 
33
  How to use:
34
 
35
  ```python
36
+ pip install transformers
37
  from transformers import AutoTokenizer, AutoModelForSequenceClassification
38
  import torch
39
+ tokenizer = AutoTokenizer.from_pretrained('tinkoff-ai/response-quality-classifier-tiny')
40
+ model = AutoModelForSequenceClassification.from_pretrained('tinkoff-ai/response-quality-classifier-tiny')
41
+ model.cuda()
42
  inputs = tokenizer('[CLS]привет[SEP]привет![SEP]как дела?[RESPONSE_TOKEN]норм, у тя как?', max_length=128, add_special_tokens=False, return_tensors='pt')
43
  with torch.inference_mode():
44
  logits = model(**inputs).logits
 
46
  relevance, specificity = probas
47
  ```
48
 
49
+ The [app](https://huggingface.co/spaces/tinkoff-ai/response-quality-classifiers) where you can easily interact with this model.
50
 
51
+ The work was done during internship at Tinkoff by [egoriyaa](https://github.com/egoriyaa), mentored by [solemn-leader](https://huggingface.co/solemn-leader).