TheBloke commited on
Commit
c02d334
1 Parent(s): 0cda6bc

Updating model files

Browse files
Files changed (1) hide show
  1. README.md +23 -1
README.md CHANGED
@@ -4,6 +4,17 @@ inference: false
4
  datasets:
5
  - gozfarb/ShareGPT_Vicuna_unfiltered
6
  ---
 
 
 
 
 
 
 
 
 
 
 
7
 
8
  # VicUnlocked-30B-LoRA GPTQ
9
 
@@ -41,7 +52,7 @@ In the `main` branch - the default one - you will find `VicUnlocked-30B-LoRA-GPT
41
 
42
  This will work with all versions of GPTQ-for-LLaMa. It has maximum compatibility
43
 
44
- It was created without groupsize so as to minimise VRAM requirements. It is created with the `--act-order` parameter to improve inference quality.
45
 
46
  * `VicUnlocked-30B-LoRA-GPTQ-4bit-128g.compat.no-act-order.safetensors`
47
  * Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
@@ -54,6 +65,17 @@ It was created without groupsize so as to minimise VRAM requirements. It is crea
54
  ```
55
 
56
 
 
 
 
 
 
 
 
 
 
 
 
57
  # Original model card
58
 
59
  # Convert tools
 
4
  datasets:
5
  - gozfarb/ShareGPT_Vicuna_unfiltered
6
  ---
7
+ <div style="width: 100%;">
8
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
9
+ </div>
10
+ <div style="display: flex; justify-content: space-between; width: 100%;">
11
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
12
+ <p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
13
+ </div>
14
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
15
+ <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
16
+ </div>
17
+ </div>
18
 
19
  # VicUnlocked-30B-LoRA GPTQ
20
 
 
52
 
53
  This will work with all versions of GPTQ-for-LLaMa. It has maximum compatibility
54
 
55
+ It was created without groupsize so as to minimise VRAM requirements. It is created with the `--act-order` parameter to improve inference quality.
56
 
57
  * `VicUnlocked-30B-LoRA-GPTQ-4bit-128g.compat.no-act-order.safetensors`
58
  * Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
 
65
  ```
66
 
67
 
68
+ ## Want to support my work?
69
+
70
+ I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
71
+
72
+ So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
73
+
74
+ Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
75
+
76
+ * Patreon: coming soon! (just awaiting approval)
77
+ * Ko-Fi: https://ko-fi.com/TheBlokeAI
78
+ * Discord: https://discord.gg/UBgz4VXf
79
  # Original model card
80
 
81
  # Convert tools