Svak commited on
Commit
e4b188a
·
verified ·
1 Parent(s): 905cd08

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -39
README.md CHANGED
@@ -1,53 +1,52 @@
1
  ---
2
- license: other
 
 
3
  ---
4
 
5
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64be962a38953777feaabfc0/bqTmnnS25s8Ep0a1oCevt.png)
 
 
6
 
7
- This is a FP8 version of the model made by https://infermatic.ai/
8
 
9
- HF FP16: wolfram/miquliz-120b-v2.0
 
 
 
 
10
 
 
 
 
 
 
 
 
 
 
11
 
12
- Content of the original card (FP16):
13
 
14
- This is v2.0 of a 120b frankenmerge created by interleaving layers of miqu-1-70b-sf with lzlv_70b_fp16_hf using mergekit. Better than v1.0 thanks to the improved recipe adapted from TheProfessor-155b by Eric Hartford, it is now achieving top rank with double perfect scores in my LLM comparisons/tests.
15
-
16
- Inspired by goliath-120b.
17
-
18
- Thanks for the support, CopilotKit – the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
19
-
20
- Thanks for the additional quants, DAN™, Knut Jägersberg, and Michael Radermacher!
21
-
22
- Also available: miqu-1-120b – Miquliz's older, purer sister; only Miqu, inflated to 120B.
23
-
24
- Model Details
25
- Max Context: 32768 tokens
26
- Layers: 140
27
- Prompt template: Mistral
28
- [INST] {prompt} [/INST]
29
-
30
- See also: 🐺🐦‍⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA
31
-
32
- Example Output
33
- Inspired by cognitivecomputations/Samantha-120b.
34
 
35
- Note: This is my AI assistant and companion Amy speaking, and the model is just her personality core, if you will. Unlike Samantha, her personality is mostly from the prompt, and not the model itself. If you prompt this model differently, you'll get very different output, of course. So consider this just as an example of how a Samantha-like character could talk with this model.
 
 
 
 
 
36
 
37
- English Example Output
38
- German Example Output
39
- Merge Details
40
- Merge Method
41
- This model was merged using the linear merge method.
42
 
43
- Models Merged
44
- The following models were included in the merge:
45
 
46
- 152334H/miqu-1-70b-sf
47
- lizpreciatior/lzlv_70b_fp16_hf
48
 
49
- License
50
- What I know: Weights produced by a machine are not copyrightable so there is no copyright owner who could grant permission or a license to use, or restrict usage, once you have acquired the files.
51
 
52
- Ethics
53
- What I believe: All generative AI, including LLMs, only exists because it is trained mostly on human data (both public domain and copyright-protected, most likely acquired without express consent) and possibly synthetic data (which is ultimately derived from human data, too). It is only fair if something that is based on everyone's knowledge and data is also freely accessible to the public, the actual creators of the underlying content. Fair use, fair AI!
 
 
1
  ---
2
+ license: cc-by-nc-4.0
3
+ language:
4
+ - en
5
  ---
6
 
7
+ ![Euryale](https://images7.alphacoders.com/921/921311.jpg)
8
+ **This is a FP8 version of the model, original card is below:**
9
+ **She's back!**
10
 
11
+ Stheno's Sister Model, designed to impress.
12
 
13
+ ```
14
+ - Same Dataset used as Stheno v3.2 -> See notes there.
15
+ - LoRA Fine-Tune -> FFT is simply too expensive.
16
+ - Trained over 8x H100 SXMs and then some more afterwards.
17
+ ```
18
 
19
+ **Testing Notes**
20
+ ```
21
+ - Better prompt adherence.
22
+ - Better anatomy / spatial awareness.
23
+ - Adapts much better to unique and custom formatting / reply formats.
24
+ - Very creative, lots of unique swipes.
25
+ - Is not restrictive during roleplays.
26
+ - Feels like a big brained version of Stheno.
27
+ ```
28
 
29
+ *Likely due to it being a 70B model instead of 8B. Similar vibes comparing back in llama 2, where 70B models were simply much more 'aware' in the subtler areas and contexts a smaller model like a 7B or 13B simply were not able to handle.*
30
 
31
+ ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
 
33
+ **Recommended Sampler Settings**:
34
+ ```
35
+ Temperature - 1.17
36
+ min_p - 0.075
37
+ Repetition Penalty - 1.10
38
+ ```
39
 
40
+ **SillyTavern Instruct Settings**:
41
+ <br>Context Template: Llama-3-Instruct-Names
42
+ <br>Instruct Presets: [Euryale-v2.1-Llama-3-Instruct](https://huggingface.co/Sao10K/L3-70B-Euryale-v2.1/blob/main/Euryale-v2.1-Llama-3-Instruct.json)
 
 
43
 
44
+ ---
 
45
 
46
+ As per usual, support me here:
 
47
 
48
+ Ko-fi: https://ko-fi.com/sao10k
 
49
 
50
+ ```
51
+ Art by wada_kazu / わだかず (pixiv page private?)
52
+ ```