File size: 2,490 Bytes
77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad 77f55a0 b1712ad |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 |
---
library_name: peft
base_model: core42/jais-13b
license: mit
datasets:
- arbml/Ashaar_dataset
language:
- ar
metrics:
- perplexity
- bertscore
---
# Model Card for Model ID
Fine-tuned using QLoRA for poem generation task.
### Model Description
We utilize Ashaar dataset and fine-tune the model to generate poems.
The input to the model is structred as follows:
'''
\#\#\# Instruction: Generate a poem based on the following title, and the given era:
\#\#\# Input: \{Title of a poem + poet era\}
\#\#\# Response: \{Poem verses\}
'''
- **Developed by:** Abdelrahman ’Boda’ Sadallah, Anastasiia Demidova, Daria Kotova
- **Model type:** Causal LM
- **Language(s) (NLP):** Arabic
- **Finetuned from model [optional]:** core42/jais-13b
### Model Sources
- **Repository:** https://github.com/BodaSadalla98/llm-optimized-fintuning
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
The model is the result of our AI project. If you intend to use it, please, refer to the repo.
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
For improving stories generation, you can play parameters: temeperature, top_p/top_k, repetition_penalty, etc.
## Training Details
### Training Data
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
Link to the dataset on huggungface: https://huggingface.co/datasets/arbml/ashaar.
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
Test split of the same dataset.
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
We are using perplexity and BERTScore.
### Results
Perplexity: 48.3125
BERTScore: 59.33
## Training procedure
The following `bitsandbytes` quantization config was used during training:
- quant_method: bitsandbytes
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: fp4
- bnb_4bit_use_double_quant: False
- bnb_4bit_compute_dtype: float32
### Framework versions
- PEFT 0.6.0.dev0 |