DeepMount00
commited on
Commit
โข
b512732
1
Parent(s):
7a360e1
Update README.md
Browse files
README.md
CHANGED
@@ -1,35 +1,41 @@
|
|
1 |
-
|
|
|
|
|
|
|
2 |
|
3 |
-
## Model Description
|
4 |
Alireo-400M is a lightweight yet powerful Italian language model with 400M parameters, designed to provide efficient natural language processing capabilities while maintaining a smaller footprint compared to larger models.
|
5 |
|
6 |
-
|
|
|
7 |
* **Architecture**: Transformer-based language model ๐๏ธ
|
8 |
* **Parameters**: 400M ๐
|
9 |
* **Context Window**: 8K tokens ๐ช
|
10 |
* **Training Data**: Curated Italian text corpus (books, articles, web content) ๐
|
11 |
* **Model Size**: ~800MB ๐พ
|
12 |
|
13 |
-
|
|
|
14 |
Despite its compact size, Alireo-400M demonstrates impressive performance:
|
15 |
|
16 |
* **Benchmark Results**: Outperforms Qwen 0.5B across multiple benchmarks ๐
|
17 |
* **Language Understanding**: Maintains high accuracy in Italian language understanding tasks ๐ฏ
|
18 |
* **Speed**: Efficient inference speed due to optimized architecture โก
|
19 |
|
20 |
-
|
|
|
21 |
* Limited context window compared to larger models
|
22 |
* May struggle with highly specialized technical content
|
23 |
* Performance may vary on dialectal variations
|
24 |
* Not suitable for multilingual tasks
|
25 |
|
26 |
-
|
|
|
27 |
* **Minimum RAM**: 2GB
|
28 |
* **Recommended RAM**: 4GB
|
29 |
* **GPU**: Optional, but recommended for faster inference
|
30 |
* **Disk Space**: ~1GB (including model and dependencies)
|
31 |
|
32 |
-
|
33 |
|
34 |
```python
|
35 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
@@ -48,10 +54,12 @@ result = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
|
48 |
print(result)
|
49 |
```
|
50 |
|
51 |
-
|
|
|
52 |
Apache 2.0
|
53 |
|
54 |
-
|
|
|
55 |
```bibtex
|
56 |
@software{alireo2024,
|
57 |
author = {[Michele Montebovi]},
|
|
|
1 |
+
<h1 style="font-size: 48px; text-align: center;">Alireo-400M ๐ค ๐ฎ๐น</h1>
|
2 |
+
<p style="font-size: 24px; text-align: center;">A Lightweight Italian Language Model</p>
|
3 |
+
|
4 |
+
<h2 style="font-size: 32px; color: #2980b9;">Model Description ๐</h2>
|
5 |
|
|
|
6 |
Alireo-400M is a lightweight yet powerful Italian language model with 400M parameters, designed to provide efficient natural language processing capabilities while maintaining a smaller footprint compared to larger models.
|
7 |
|
8 |
+
<h2 style="font-size: 32px; color: #2980b9;">Key Features โจ</h2>
|
9 |
+
|
10 |
* **Architecture**: Transformer-based language model ๐๏ธ
|
11 |
* **Parameters**: 400M ๐
|
12 |
* **Context Window**: 8K tokens ๐ช
|
13 |
* **Training Data**: Curated Italian text corpus (books, articles, web content) ๐
|
14 |
* **Model Size**: ~800MB ๐พ
|
15 |
|
16 |
+
<h2 style="font-size: 32px; color: #2980b9;">Performance ๐</h2>
|
17 |
+
|
18 |
Despite its compact size, Alireo-400M demonstrates impressive performance:
|
19 |
|
20 |
* **Benchmark Results**: Outperforms Qwen 0.5B across multiple benchmarks ๐
|
21 |
* **Language Understanding**: Maintains high accuracy in Italian language understanding tasks ๐ฏ
|
22 |
* **Speed**: Efficient inference speed due to optimized architecture โก
|
23 |
|
24 |
+
<h2 style="font-size: 32px; color: #2980b9;">Limitations โ ๏ธ</h2>
|
25 |
+
|
26 |
* Limited context window compared to larger models
|
27 |
* May struggle with highly specialized technical content
|
28 |
* Performance may vary on dialectal variations
|
29 |
* Not suitable for multilingual tasks
|
30 |
|
31 |
+
<h2 style="font-size: 32px; color: #2980b9;">Hardware Requirements ๐ป</h2>
|
32 |
+
|
33 |
* **Minimum RAM**: 2GB
|
34 |
* **Recommended RAM**: 4GB
|
35 |
* **GPU**: Optional, but recommended for faster inference
|
36 |
* **Disk Space**: ~1GB (including model and dependencies)
|
37 |
|
38 |
+
<h2 style="font-size: 32px; color: #2980b9;">Usage Example ๐ก</h2>
|
39 |
|
40 |
```python
|
41 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
|
54 |
print(result)
|
55 |
```
|
56 |
|
57 |
+
<h2 style="font-size: 32px; color: #2980b9;">License ๐</h2>
|
58 |
+
|
59 |
Apache 2.0
|
60 |
|
61 |
+
<h2 style="font-size: 32px; color: #2980b9;">Citation ๐</h2>
|
62 |
+
|
63 |
```bibtex
|
64 |
@software{alireo2024,
|
65 |
author = {[Michele Montebovi]},
|