some1nostr
commited on
Commit
•
2d1a819
1
Parent(s):
012e89d
Update README.md
Browse files
README.md
CHANGED
@@ -12,9 +12,9 @@ license: apache-2.0
|
|
12 |
|
13 |
# Pre Trained With
|
14 |
|
15 |
-
- Nostr notes: This makes the model know more about bitcoin and other topics discussed on Nostr.
|
16 |
- Health related topics
|
17 |
- Faith related topics
|
|
|
18 |
- Nutrition related topics
|
19 |
- Medicinal herbs
|
20 |
|
@@ -27,22 +27,20 @@ license: apache-2.0
|
|
27 |
|
28 |
# Uses
|
29 |
|
30 |
-
You can read more about the "LLM curation" via fine tuning with Nostr notes:
|
31 |
-
[here](https://habla.news/a/naddr1qvzqqqr4gupzp8lvwt2hnw42wu40nec7vw949ys4wgdvums0svs8yhktl8mhlpd3qqgx6cnjvfhhzet20pkhqdn2wenkvu6gy4y) and
|
32 |
-
[here](https://habla.news/a/naddr1qvzqqqr4gupzp8lvwt2hnw42wu40nec7vw949ys4wgdvums0svs8yhktl8mhlpd3qqxnzde3xumrswfjx56rjwf4kkqhsx).
|
33 |
-
|
34 |
Compared to other models this may know more about Nostr, Bitcoin and healthy living.
|
35 |
Closer aligment to opinions of people on Nostr. It may have alternative ideas to mainstream because of Nostr and my own curation.
|
36 |
So it is basically somewhere in between base llama 3.0 plus Nostr plus my values.
|
37 |
|
|
|
|
|
|
|
|
|
38 |
I am using the model here as a ground truth for Nostr related questions: https://wikifreedia.xyz/based-llm-leaderboard/npub1nlk894teh248w2heuu0x8z6jjg2hyxkwdc8cxgrjtm9lnamlskcsghjm9c
|
39 |
|
40 |
Use repeat penalty of 1.05 or more to avoid repetitions.
|
41 |
|
42 |
I hope you like it. Let me know about your experience. You can DM me on Nostr.
|
43 |
|
44 |
-
The number in the filenames (like 21345) means the version. I take the training steps and add those as the version. Each source, book, note adds to the version.
|
45 |
-
|
46 |
|
47 |
# Warning
|
48 |
|
@@ -54,12 +52,17 @@ There is no guarantee that the model will be of any use. It may hallucinate ofte
|
|
54 |
|
55 |
## Training Data
|
56 |
|
57 |
-
The sources mentioned above are converted to TXT files and used as pre training.
|
|
|
|
|
58 |
|
59 |
## Training Procedure
|
60 |
|
61 |
LLaMa-Factory is used to train on 2* RTX 3090! fsdp_qlora is the technique.
|
62 |
|
|
|
|
|
|
|
63 |
# Contributions
|
64 |
|
65 |
You can tip me on Nostr https://primal.net/p/npub1nlk894teh248w2heuu0x8z6jjg2hyxkwdc8cxgrjtm9lnamlskcsghjm9c
|
|
|
12 |
|
13 |
# Pre Trained With
|
14 |
|
|
|
15 |
- Health related topics
|
16 |
- Faith related topics
|
17 |
+
- Nostr notes: This makes the model know more about bitcoin and other topics discussed on Nostr.
|
18 |
- Nutrition related topics
|
19 |
- Medicinal herbs
|
20 |
|
|
|
27 |
|
28 |
# Uses
|
29 |
|
|
|
|
|
|
|
|
|
30 |
Compared to other models this may know more about Nostr, Bitcoin and healthy living.
|
31 |
Closer aligment to opinions of people on Nostr. It may have alternative ideas to mainstream because of Nostr and my own curation.
|
32 |
So it is basically somewhere in between base llama 3.0 plus Nostr plus my values.
|
33 |
|
34 |
+
You can read more about the "LLM curation" via fine tuning with Nostr notes:
|
35 |
+
[here](https://habla.news/a/naddr1qvzqqqr4gupzp8lvwt2hnw42wu40nec7vw949ys4wgdvums0svs8yhktl8mhlpd3qqgx6cnjvfhhzet20pkhqdn2wenkvu6gy4y) and
|
36 |
+
[here](https://habla.news/a/naddr1qvzqqqr4gupzp8lvwt2hnw42wu40nec7vw949ys4wgdvums0svs8yhktl8mhlpd3qqxnzde3xumrswfjx56rjwf4kkqhsx).
|
37 |
+
|
38 |
I am using the model here as a ground truth for Nostr related questions: https://wikifreedia.xyz/based-llm-leaderboard/npub1nlk894teh248w2heuu0x8z6jjg2hyxkwdc8cxgrjtm9lnamlskcsghjm9c
|
39 |
|
40 |
Use repeat penalty of 1.05 or more to avoid repetitions.
|
41 |
|
42 |
I hope you like it. Let me know about your experience. You can DM me on Nostr.
|
43 |
|
|
|
|
|
44 |
|
45 |
# Warning
|
46 |
|
|
|
52 |
|
53 |
## Training Data
|
54 |
|
55 |
+
The sources mentioned above are converted to TXT files and used as pre training. A few supervised fine tunings that helped with controlling length of the output.
|
56 |
+
|
57 |
+
Sources include like banned videos from youtube.
|
58 |
|
59 |
## Training Procedure
|
60 |
|
61 |
LLaMa-Factory is used to train on 2* RTX 3090! fsdp_qlora is the technique.
|
62 |
|
63 |
+
The number in the filenames (like 21345) means the version. I take the training steps and add those as the version. Each source, book, note adds to the version.
|
64 |
+
|
65 |
+
|
66 |
# Contributions
|
67 |
|
68 |
You can tip me on Nostr https://primal.net/p/npub1nlk894teh248w2heuu0x8z6jjg2hyxkwdc8cxgrjtm9lnamlskcsghjm9c
|