some1nostr
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -9,23 +9,23 @@ license: apache-2.0
|
|
9 |
# Model Card for Ostrich
|
10 |
|
11 |
|
12 |
-
-
|
13 |
-
- Trained a bit about nostr
|
14 |
- Trained a bit about bitcoin
|
15 |
-
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
|
|
|
|
20 |
|
|
|
|
|
21 |
|
22 |
## Model Details
|
23 |
|
24 |
-
Based on https://huggingface.co/crestf411/daybreak-miqu-1-70b-v1.0-hf because it is one of the most uncensored according to https://huggingface.co/spaces/DontPlanToEnd/UGI-Leaderboard.
|
25 |
-
|
26 |
|
27 |
-
- **
|
28 |
-
- **Finetuned from model:** https://huggingface.co/crestf411/daybreak-miqu-1-70b-v1.0-hf
|
29 |
|
30 |
|
31 |
## Uses
|
@@ -34,6 +34,10 @@ Ask any question, compared to other models this may know more about Nostr and Bi
|
|
34 |
You can use llama.cpp to chat with it.
|
35 |
You can also use llama-cpp-python package to use it in a Python script.
|
36 |
|
|
|
|
|
|
|
|
|
37 |
|
38 |
## Warning
|
39 |
|
@@ -45,10 +49,11 @@ The trainer, developer or uploader of this model does not assume any liability.
|
|
45 |
|
46 |
### Training Data
|
47 |
|
48 |
-
Nostr related info from web and nostr itself, bitcoin related info.
|
|
|
49 |
|
50 |
### Training Procedure
|
51 |
|
52 |
LLaMa-Factory is used to train on 2x3090! fsdp_qlora is the technique.
|
53 |
|
54 |
-
|
|
|
9 |
# Model Card for Ostrich
|
10 |
|
11 |
|
12 |
+
- Trained with some of the Nostr notes
|
|
|
13 |
- Trained a bit about bitcoin
|
14 |
+
- Aligned a bit in these domains:
|
15 |
+
- Health
|
16 |
+
- Permaculture
|
17 |
+
- Phytochemicals
|
18 |
+
- Alternative medicine
|
19 |
+
- Herbs
|
20 |
+
- Nutrition
|
21 |
|
22 |
+
Read more about it here:
|
23 |
+
https://habla.news/a/naddr1qvzqqqr4gupzp8lvwt2hnw42wu40nec7vw949ys4wgdvums0svs8yhktl8mhlpd3qqxnzde3xsunjwfkxcunwv3jvtnjyc
|
24 |
|
25 |
## Model Details
|
26 |
|
|
|
|
|
27 |
|
28 |
+
- **Finetuned from model:** https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct
|
|
|
29 |
|
30 |
|
31 |
## Uses
|
|
|
34 |
You can use llama.cpp to chat with it.
|
35 |
You can also use llama-cpp-python package to use it in a Python script.
|
36 |
|
37 |
+
Llama3 chat template can be used. <|begin_of_text|><|start_header_id|> ...
|
38 |
+
|
39 |
+
Use repeat penalty of 1.05 or more to avoid repetitions.
|
40 |
+
|
41 |
|
42 |
## Warning
|
43 |
|
|
|
49 |
|
50 |
### Training Data
|
51 |
|
52 |
+
Nostr related info from web and nostr itself, bitcoin related info. Info on health domain.
|
53 |
+
Information that aligns well with humanity is preferred.
|
54 |
|
55 |
### Training Procedure
|
56 |
|
57 |
LLaMa-Factory is used to train on 2x3090! fsdp_qlora is the technique.
|
58 |
|
59 |
+
The Nostr training took ~30 hours for a dataset of about 20MB.
|