prithivMLmods
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -40,7 +40,7 @@ _/ |_ _______ |__|_____ ____ ____ __ __ | | __ __ _____
|
|
40 |
|
41 |
Triangulum 1B is a collection of pretrained and instruction-tuned generative models, designed for multilingual applications. These models are trained using synthetic datasets based on long chains of thought, enabling them to perform complex reasoning tasks effectively.
|
42 |
|
43 |
-
# **Key Features**
|
44 |
|
45 |
- **Foundation Model**: Built upon LLaMA's autoregressive language model, leveraging an optimized transformer architecture for enhanced performance.
|
46 |
|
@@ -48,6 +48,10 @@ Triangulum 1B is a collection of pretrained and instruction-tuned generative mod
|
|
48 |
|
49 |
- **Multilingual Support**: Designed to handle multiple languages, ensuring broad applicability across diverse linguistic contexts.
|
50 |
|
|
|
|
|
|
|
|
|
51 |
# **Training Approach**
|
52 |
|
53 |
1. **Synthetic Datasets**: Utilizes long chain-of-thought synthetic data to enhance reasoning capabilities.
|
|
|
40 |
|
41 |
Triangulum 1B is a collection of pretrained and instruction-tuned generative models, designed for multilingual applications. These models are trained using synthetic datasets based on long chains of thought, enabling them to perform complex reasoning tasks effectively.
|
42 |
|
43 |
+
# **Key Features & Model Architecture**
|
44 |
|
45 |
- **Foundation Model**: Built upon LLaMA's autoregressive language model, leveraging an optimized transformer architecture for enhanced performance.
|
46 |
|
|
|
48 |
|
49 |
- **Multilingual Support**: Designed to handle multiple languages, ensuring broad applicability across diverse linguistic contexts.
|
50 |
|
51 |
+
---
|
52 |
+
|
53 |
+
- Llama 3.2 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
|
54 |
+
|
55 |
# **Training Approach**
|
56 |
|
57 |
1. **Synthetic Datasets**: Utilizes long chain-of-thought synthetic data to enhance reasoning capabilities.
|