sometimesanotion commited on
Commit
6b8b883
·
verified ·
1 Parent(s): 1f73368

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -7
README.md CHANGED
@@ -17,10 +17,14 @@ metrics:
17
  - accuracy
18
  pipeline_tag: text-generation
19
  ---
20
- ![Lamarck.webp](https://huggingface.co/sometimesanotion/Lamarck-14B-v0.6/resolve/main/Lamarck.webp)
21
  ---
22
 
23
- Lamarck 14B v0.6:
 
 
 
 
24
 
25
  ## Merge Details
26
 
@@ -32,16 +36,13 @@ The model_stock, breadcrumbs, and della_linear all use the following models:
32
 
33
  [sometimesanotion/Qwen2.5-14B-Vimarckoso-v3](https://huggingface.co/sometimesanotion/Qwen2.5-14B-Vimarckoso-v3)
34
  [sometimesanotion/Lamarck-14B-v0.3](https://huggingface.co/sometimesanotion/Lamarck-14B-v0.3)
35
- [sometimesanotion/Qwenvergence-14B-v3-Prose](https://huggingface.co/sometimesanotion/Qwenvergence-14B-v3-Prose)
36
- [Krystalan/DRT-o1-14B](https://huggingface.co/Krystalan/DRT-o1-14B)
37
  [underwoods/medius-erebus-magnum-14b](https://huggingface.co/underwoods/medius-erebus-magnum-14b)
38
  [sometimesanotion/Abliterate-Qwenvergence](https://huggingface.co/sometimesanotion/Abliterate-Qwenvergence) - A custom version of [huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2](https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2)
39
 
40
  ### Configuration
41
 
42
- The following YAML configurations were used to initialize and finalize this model:
43
-
44
-
45
  This model was made in two branches: a della_linear merge, and a sequence of model_stock and then breadcrumbs SLERP-merged below.
46
 
47
  ```yaml
 
17
  - accuracy
18
  pipeline_tag: text-generation
19
  ---
20
+ ![Lamarck.webp](https://huggingface.co/sometimesanotion/Lamarck-14B-v0.6-rc4/resolve/main/Lamarck.webp)
21
  ---
22
 
23
+ Lamarck 14B v0.6: A generalist merge focused on multi-step reasoning, prose, multi-language ability, and code. It is based on components that have punched above their weight in the 14 billion parameter class.
24
+
25
+ Previous releases were based on a SLERP merge of model_stock->della branches focused on reasoning and prose. The prose branch got surprisingly good at reasoning, and the reasoning branch being the base for IFEVAL became an all-around generalist. Some of you have already downloaded the reasoning branch, released as [sometimesanotion/Qwen2.5-14B-Vimarckoso-v3](https://huggingface.co/sometimesanotion/Qwen2.5-14B-Vimarckoso-v3).
26
+
27
+ Lamarck 0.6 aims to build upon Vimarckoso v3's all-around strength with strong buffs to prose and translation quality, and strong reasoning for its class. Updates to come as leaderboards become available to evaluate it in-depth.
28
 
29
  ## Merge Details
30
 
 
36
 
37
  [sometimesanotion/Qwen2.5-14B-Vimarckoso-v3](https://huggingface.co/sometimesanotion/Qwen2.5-14B-Vimarckoso-v3)
38
  [sometimesanotion/Lamarck-14B-v0.3](https://huggingface.co/sometimesanotion/Lamarck-14B-v0.3)
39
+ [sometimesanotion/Qwenvergence-14B-v3-Prose](https://huggingface.co/sometimesanotion/Qwenvergence-14B-v3-Prose) - a model_stock merge of multiple prose-oriented models which posts surprisingly high MATH, GPQA, and MUSR scores.
40
+ [Krystalan/DRT-o1-14B](https://huggingface.co/Krystalan/DRT-o1-14B) - A particularly interesting model which applies extra reasoning to language translation. Check out their fascinating research paper at [arxiv.org/abs/2412.17498](https://arxiv.org/abs/2412.17498).
41
  [underwoods/medius-erebus-magnum-14b](https://huggingface.co/underwoods/medius-erebus-magnum-14b)
42
  [sometimesanotion/Abliterate-Qwenvergence](https://huggingface.co/sometimesanotion/Abliterate-Qwenvergence) - A custom version of [huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2](https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2)
43
 
44
  ### Configuration
45
 
 
 
 
46
  This model was made in two branches: a della_linear merge, and a sequence of model_stock and then breadcrumbs SLERP-merged below.
47
 
48
  ```yaml