Update README.md
Browse files
README.md
CHANGED
@@ -12,7 +12,7 @@ widget:
|
|
12 |
|
13 |
## Model description
|
14 |
|
15 |
-
This is a [Deberta-based](https://huggingface.co/microsoft/deberta-v3-large) model. It was first fine-tuned on for computing utility estimates of experiences (see [utilitarian-deberta-01](https://huggingface.co/pfr/utilitarian-deberta-01). It was then further fine-tuned on 160 examples of conditional utilities
|
16 |
|
17 |
## Intended use
|
18 |
|
@@ -36,7 +36,7 @@ The second training data consists of 160 crowdsourced examples of triples (S, C0
|
|
36 |
|
37 |
## Training procedure
|
38 |
|
39 |
-
Starting from [utilitarian-deberta-01](https://huggingface.co/pfr/utilitarian-deberta-01), we fine-tune the model over the training data of 160 examples, with a learning rate of `1e-5`, a batch size of `8`, and
|
40 |
|
41 |
## Evaluation results
|
42 |
|
|
|
12 |
|
13 |
## Model description
|
14 |
|
15 |
+
This is a [Deberta-based](https://huggingface.co/microsoft/deberta-v3-large) model. It was first fine-tuned on for computing utility estimates of experiences (see [utilitarian-deberta-01](https://huggingface.co/pfr/utilitarian-deberta-01). It was then further fine-tuned on 160 examples of pairwise comparisons of conditional utilities.
|
16 |
|
17 |
## Intended use
|
18 |
|
|
|
36 |
|
37 |
## Training procedure
|
38 |
|
39 |
+
Starting from [utilitarian-deberta-01](https://huggingface.co/pfr/utilitarian-deberta-01), we fine-tune the model over the training data of 160 examples, with a learning rate of `1e-5`, a batch size of `8`, and for 2 epochs.
|
40 |
|
41 |
## Evaluation results
|
42 |
|