cstr commited on
Commit
332de66
1 Parent(s): 181b4be

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +64 -0
README.md ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: transformers
4
+ pipeline_tag: text-generation
5
+ language:
6
+ - bg
7
+ - ca
8
+ - code
9
+ - cs
10
+ - cy
11
+ - da
12
+ - de
13
+ - el
14
+ - en
15
+ - es
16
+ - et
17
+ - eu
18
+ - fi
19
+ - fr
20
+ - ga
21
+ - gl
22
+ - hr
23
+ - hu
24
+ - it
25
+ - lt
26
+ - lv
27
+ - mt
28
+ - nl
29
+ - nn
30
+ - no
31
+ - oc
32
+ - pl
33
+ - pt
34
+ - ro
35
+ - ru
36
+ - sh
37
+ - sk
38
+ - sl
39
+ - sr
40
+ - sv
41
+ - uk
42
+ ---
43
+
44
+ # GGUF quants
45
+
46
+ Experimental GGUF quantization of [BSC-LT/salamandra-7b-instruct](https://huggingface.co/BSC-LT/salamandra-7b-instruct/) from llama.cpp (older version b2750).
47
+
48
+ ![](./images/salamandra_header.png)
49
+
50
+ # Salamandra Model Card
51
+
52
+ Salamandra comes in three different sizes — 2B, 7B and 40B parameters — with their respective base and instruction-tuned variants.
53
+ This model card corresponds to the 7B instructed version.
54
+
55
+ To visit the model cards of other Salamandra versions, please refer to the [Model Index](#model-index).
56
+
57
+ The entire Salamandra family is released under a permissive [Apache 2.0 license]((https://www.apache.org/licenses/LICENSE-2.0)).
58
+ Along with the open weights, all training scripts and configuration files are made publicly available in [this GitHub repository](https://github.com/langtech-bsc/salamandra).
59
+
60
+ > [!WARNING]
61
+ > **DISCLAIMER:** This model is a first proof-of-concept designed to demonstrate the instruction-following capabilities of recently released base models.
62
+ > It has been optimized to engage in conversation but has *NOT* been aligned through RLHF to filter or avoid sensitive topics.
63
+ > As a result, it may generate harmful or inappropriate content.
64
+ > The team is actively working to enhance its performance through further instruction and alignment with RL techniques.