Update README.md
Browse files
README.md
CHANGED
@@ -17,10 +17,9 @@ The labels explanation:
|
|
17 |
- `specificity`: is the last message in the dialogue interesting and promotes the continuation of the dialogue.
|
18 |
|
19 |
It is pretrained on a large corpus of dialog data in unsupervised manner: the model is trained to predict whether last response was in a real dialog, or it was pulled from some other dialog at random.
|
20 |
-
|
21 |
Then it was finetuned on manually labelled examples (dataset will be posted soon).
|
22 |
|
23 |
-
The model was trained with
|
24 |
|
25 |
The performance of the model on validation split (dataset will be posted soon) (with the best thresholds for validation samples):
|
26 |
|
@@ -34,12 +33,12 @@ The performance of the model on validation split (dataset will be posted soon) (
|
|
34 |
How to use:
|
35 |
|
36 |
```python
|
37 |
-
|
38 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
39 |
import torch
|
40 |
-
tokenizer = AutoTokenizer.from_pretrained(
|
41 |
-
model = AutoModelForSequenceClassification.from_pretrained(
|
42 |
-
|
43 |
inputs = tokenizer('[CLS]привет[SEP]привет![SEP]как дела?[RESPONSE_TOKEN]норм, у тя как?', max_length=128, add_special_tokens=False, return_tensors='pt')
|
44 |
with torch.inference_mode():
|
45 |
logits = model(**inputs).logits
|
@@ -47,6 +46,6 @@ with torch.inference_mode():
|
|
47 |
relevance, specificity = probas
|
48 |
```
|
49 |
|
50 |
-
The [app](https://huggingface.co/spaces/tinkoff-ai/response-quality-classifiers) where you can easily
|
51 |
|
52 |
-
The work was done during internship at Tinkoff by [egoriyaa](https://github.com/egoriyaa).
|
|
|
17 |
- `specificity`: is the last message in the dialogue interesting and promotes the continuation of the dialogue.
|
18 |
|
19 |
It is pretrained on a large corpus of dialog data in unsupervised manner: the model is trained to predict whether last response was in a real dialog, or it was pulled from some other dialog at random.
|
|
|
20 |
Then it was finetuned on manually labelled examples (dataset will be posted soon).
|
21 |
|
22 |
+
The model was trained with three messages in the context and one response. Each message was tokenized separately with ``` max_length = 32 ```.
|
23 |
|
24 |
The performance of the model on validation split (dataset will be posted soon) (with the best thresholds for validation samples):
|
25 |
|
|
|
33 |
How to use:
|
34 |
|
35 |
```python
|
36 |
+
pip install transformers
|
37 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
38 |
import torch
|
39 |
+
tokenizer = AutoTokenizer.from_pretrained('tinkoff-ai/response-quality-classifier-tiny')
|
40 |
+
model = AutoModelForSequenceClassification.from_pretrained('tinkoff-ai/response-quality-classifier-tiny')
|
41 |
+
model.cuda()
|
42 |
inputs = tokenizer('[CLS]привет[SEP]привет![SEP]как дела?[RESPONSE_TOKEN]норм, у тя как?', max_length=128, add_special_tokens=False, return_tensors='pt')
|
43 |
with torch.inference_mode():
|
44 |
logits = model(**inputs).logits
|
|
|
46 |
relevance, specificity = probas
|
47 |
```
|
48 |
|
49 |
+
The [app](https://huggingface.co/spaces/tinkoff-ai/response-quality-classifiers) where you can easily interact with this model.
|
50 |
|
51 |
+
The work was done during internship at Tinkoff by [egoriyaa](https://github.com/egoriyaa), mentored by [solemn-leader](https://huggingface.co/solemn-leader).
|