some1nostr commited on
Commit
431bdca
·
verified ·
1 Parent(s): a655eca

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -7
README.md CHANGED
@@ -9,7 +9,7 @@ license: apache-2.0
9
  # Model Card for Ostrich
10
 
11
 
12
- - **Trained with some of the Nostr notes**
13
  - **Aligned a bit in these domains:**
14
  - Bitcoin
15
  - Health
@@ -18,12 +18,11 @@ license: apache-2.0
18
  - Alternative medicine
19
  - Herbs
20
  - Nutrition
 
21
 
22
- Read more about it here:
23
- https://habla.news/a/naddr1qvzqqqr4gupzp8lvwt2hnw42wu40nec7vw949ys4wgdvums0svs8yhktl8mhlpd3qqxnzde3xsunjwfkxcunwv3jvtnjyc
24
 
25
- A running model is here:
26
- https://njump.me/npub1chadadwep45t4l7xx9z45p72xsxv7833zyy4tctdgh44lpc50nvsrjex2m (Though it may be down for maintenance time to time. You need to DM the bot for it to answer.)
27
 
28
 
29
  ## Model Details
@@ -31,7 +30,7 @@ https://njump.me/npub1chadadwep45t4l7xx9z45p72xsxv7833zyy4tctdgh44lpc50nvsrjex2m
31
 
32
  - **Finetuned from model:** https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct
33
 
34
- The number in the filenames like 4750 means the version. Higher numbers are newer versions.
35
 
36
 
37
  ## Uses
@@ -62,4 +61,4 @@ Information that aligns well with humanity is preferred.
62
 
63
  LLaMa-Factory is used to train on 2x3090! fsdp_qlora is the technique.
64
 
65
- The Nostr training took ~140 hours for a dataset of ~67MB.
 
9
  # Model Card for Ostrich
10
 
11
 
12
+ - **Trained with Nostr notes**
13
  - **Aligned a bit in these domains:**
14
  - Bitcoin
15
  - Health
 
18
  - Alternative medicine
19
  - Herbs
20
  - Nutrition
21
+ - Faith
22
 
23
+ You can read more about it here:
24
+ https://habla.news/a/naddr1qvzqqqr4gupzp8lvwt2hnw42wu40nec7vw949ys4wgdvums0svs8yhktl8mhlpd3qqxnzde3xumrswfjx56rjwf4kkqhsx
25
 
 
 
26
 
27
 
28
  ## Model Details
 
30
 
31
  - **Finetuned from model:** https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct
32
 
33
+ The number in the filenames like 7314 means the version.
34
 
35
 
36
  ## Uses
 
61
 
62
  LLaMa-Factory is used to train on 2x3090! fsdp_qlora is the technique.
63
 
64
+ The Nostr training took ~200 hours.