Update README.md
Browse files
README.md
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
datasets:
|
3 |
- jondurbin/airoboros-gpt4-1.4.1
|
4 |
---
|
5 |
-
# RoPE Scaled QLoRA
|
6 |
|
7 |
LoRA Weights can be found here: https://huggingface.co/bhenrym14/airoboros-13b-gpt4-1.4.1-PI-8192-LoRA
|
8 |
|
@@ -14,16 +14,17 @@ This is [Jon Durbin's Airoboros 13B GPT4 1.4](https://huggingface.co/jondurbin/a
|
|
14 |
- Context length extended to 8192 by RoPE Scaled Embeddings, but NOT via the superHOT LoRA. I started with base Llama-13b.
|
15 |
- Training sequences beyond 2048 have the target truncated to equal 2048.
|
16 |
- Used airoboros-gpt4-1.4.1 dataset instead of airoboros-gpt4-1.4
|
17 |
-
- This is a QLoRA
|
18 |
|
19 |
-
Otherwise, I emulated the training process as closely as possible (rank 64 QLoRA) It was trained on 1x RTX 6000 Ada for ~18 hours.
|
20 |
|
21 |
## How to Use
|
22 |
The easiest way is to use [oobabooga text-generation-webui](https://github.com/oobabooga/text-generation-webui) with ExLlama. You'll need to set max_seq_len to 8192 and compress_pos_emb to 4.
|
23 |
|
|
|
24 |
|
25 |
## Motivation
|
26 |
-
Recent advancements in extending context by RoPE scaling ([kaiokendev](https://kaiokendev.github.io/til#extending-context-to-8k) and [meta AI)](https://arxiv.org/abs/2306.15595)) demonstrate the ability to extend the context window without (total) retraining. Finetuning has shown to be necessary to properly leverage the longer context. The superHOT LoRA is an adapter that has been
|
27 |
|
28 |
## Relative Performance (perplexity)
|
29 |
| Model | Context (tokens) | Perplexity |
|
|
|
2 |
datasets:
|
3 |
- jondurbin/airoboros-gpt4-1.4.1
|
4 |
---
|
5 |
+
# RoPE Scaled QLoRA Fine-tune of Llama-13b on airoboros-gpt4-1.4.1 (GPTQ)
|
6 |
|
7 |
LoRA Weights can be found here: https://huggingface.co/bhenrym14/airoboros-13b-gpt4-1.4.1-PI-8192-LoRA
|
8 |
|
|
|
14 |
- Context length extended to 8192 by RoPE Scaled Embeddings, but NOT via the superHOT LoRA. I started with base Llama-13b.
|
15 |
- Training sequences beyond 2048 have the target truncated to equal 2048.
|
16 |
- Used airoboros-gpt4-1.4.1 dataset instead of airoboros-gpt4-1.4
|
17 |
+
- **This is a QLoRA fine-tune**. The original 13b model is a full fine-tune.
|
18 |
|
19 |
+
Otherwise, I emulated the training process as closely as possible (rank 64 QLoRA). It was trained on 1x RTX 6000 Ada for ~18 hours.
|
20 |
|
21 |
## How to Use
|
22 |
The easiest way is to use [oobabooga text-generation-webui](https://github.com/oobabooga/text-generation-webui) with ExLlama. You'll need to set max_seq_len to 8192 and compress_pos_emb to 4.
|
23 |
|
24 |
+
If you wish to use AutoGPTQ/GPTQ-for-Llama instead, you'll need to patch in the appropriate RoPE scaling module. see: [replace_llama_rope_with_scaled_rope](https://github.com/bhenrym14/qlora-airoboros-longcontext/blob/main/scaledllama/llama_rope_scaled_monkey_patch.py)
|
25 |
|
26 |
## Motivation
|
27 |
+
Recent advancements in extending context by RoPE scaling ([kaiokendev](https://kaiokendev.github.io/til#extending-context-to-8k) and [meta AI)](https://arxiv.org/abs/2306.15595)) demonstrate the ability to extend the context window without (total) retraining. Finetuning has shown to be necessary to properly leverage the longer context. The superHOT LoRA is an adapter that has been fine-tuned on longer context (8192 tokens); even when applied to models trained on dissimilar datasets, it successfully extends the context window to which the model can attend. While it's impressive this adapter is so flexible, how much does performance suffer relative to a model that has been fine-tuned with the scaled embeddings from the start? This is an experiment to explore this.
|
28 |
|
29 |
## Relative Performance (perplexity)
|
30 |
| Model | Context (tokens) | Perplexity |
|