bhenrym14 commited on
Commit
5db64ea
1 Parent(s): dc39e0f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -2,7 +2,7 @@
2
  datasets:
3
  - jondurbin/airoboros-gpt4-1.4.1
4
  ---
5
- # RoPE Scaled QLoRA Finetune of airoboros-13b-gpt4-1.4.1 (GPTQ)
6
 
7
  LoRA Weights can be found here: https://huggingface.co/bhenrym14/airoboros-13b-gpt4-1.4.1-PI-8192-LoRA
8
 
@@ -14,16 +14,17 @@ This is [Jon Durbin's Airoboros 13B GPT4 1.4](https://huggingface.co/jondurbin/a
14
  - Context length extended to 8192 by RoPE Scaled Embeddings, but NOT via the superHOT LoRA. I started with base Llama-13b.
15
  - Training sequences beyond 2048 have the target truncated to equal 2048.
16
  - Used airoboros-gpt4-1.4.1 dataset instead of airoboros-gpt4-1.4
17
- - This is a QLoRA finetune. The original 13b model is a full finetune.
18
 
19
- Otherwise, I emulated the training process as closely as possible (rank 64 QLoRA) It was trained on 1x RTX 6000 Ada for ~18 hours.
20
 
21
  ## How to Use
22
  The easiest way is to use [oobabooga text-generation-webui](https://github.com/oobabooga/text-generation-webui) with ExLlama. You'll need to set max_seq_len to 8192 and compress_pos_emb to 4.
23
 
 
24
 
25
  ## Motivation
26
- Recent advancements in extending context by RoPE scaling ([kaiokendev](https://kaiokendev.github.io/til#extending-context-to-8k) and [meta AI)](https://arxiv.org/abs/2306.15595)) demonstrate the ability to extend the context window without (total) retraining. Finetuning has shown to be necessary to properly leverage the longer context. The superHOT LoRA is an adapter that has been finetuned on longer context (8192 tokens); even when applied to models trained on dissimilar datasets, it successfully extends the context window to which the model can attend. While it's impressive this adapter is so flexible, how much does performance suffer relative to a model that has been finetuned with the scaled embeddings from the start? This is an experiment to explore this.
27
 
28
  ## Relative Performance (perplexity)
29
  | Model | Context (tokens) | Perplexity |
 
2
  datasets:
3
  - jondurbin/airoboros-gpt4-1.4.1
4
  ---
5
+ # RoPE Scaled QLoRA Fine-tune of Llama-13b on airoboros-gpt4-1.4.1 (GPTQ)
6
 
7
  LoRA Weights can be found here: https://huggingface.co/bhenrym14/airoboros-13b-gpt4-1.4.1-PI-8192-LoRA
8
 
 
14
  - Context length extended to 8192 by RoPE Scaled Embeddings, but NOT via the superHOT LoRA. I started with base Llama-13b.
15
  - Training sequences beyond 2048 have the target truncated to equal 2048.
16
  - Used airoboros-gpt4-1.4.1 dataset instead of airoboros-gpt4-1.4
17
+ - **This is a QLoRA fine-tune**. The original 13b model is a full fine-tune.
18
 
19
+ Otherwise, I emulated the training process as closely as possible (rank 64 QLoRA). It was trained on 1x RTX 6000 Ada for ~18 hours.
20
 
21
  ## How to Use
22
  The easiest way is to use [oobabooga text-generation-webui](https://github.com/oobabooga/text-generation-webui) with ExLlama. You'll need to set max_seq_len to 8192 and compress_pos_emb to 4.
23
 
24
+ If you wish to use AutoGPTQ/GPTQ-for-Llama instead, you'll need to patch in the appropriate RoPE scaling module. see: [replace_llama_rope_with_scaled_rope](https://github.com/bhenrym14/qlora-airoboros-longcontext/blob/main/scaledllama/llama_rope_scaled_monkey_patch.py)
25
 
26
  ## Motivation
27
+ Recent advancements in extending context by RoPE scaling ([kaiokendev](https://kaiokendev.github.io/til#extending-context-to-8k) and [meta AI)](https://arxiv.org/abs/2306.15595)) demonstrate the ability to extend the context window without (total) retraining. Finetuning has shown to be necessary to properly leverage the longer context. The superHOT LoRA is an adapter that has been fine-tuned on longer context (8192 tokens); even when applied to models trained on dissimilar datasets, it successfully extends the context window to which the model can attend. While it's impressive this adapter is so flexible, how much does performance suffer relative to a model that has been fine-tuned with the scaled embeddings from the start? This is an experiment to explore this.
28
 
29
  ## Relative Performance (perplexity)
30
  | Model | Context (tokens) | Perplexity |