Blog models
Collection
This is a set of models that come from the trainings I do in my personal blog
•
12 items
•
Updated
This model is a fine-tuned version of openai-community/gpt2 on Maximofn/short-jokes-dataset dataset. It achieves the following results on the evaluation set:
It is the result of the post Fine tunning SML
This model generated english jokes
Text generation, english jokes
It is training on Maximofn/short-jokes-dataset dataset.
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
3.3866 | 1.0 | 7447 | 3.2590 |
3.2599 | 2.0 | 14894 | 3.1997 |
3.2126 | 3.0 | 22341 | 3.1920 |
Base model
openai-community/gpt2