prithivMLmods commited on
Commit
baa031f
·
verified ·
1 Parent(s): 09f6781

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -1
README.md CHANGED
@@ -40,7 +40,7 @@ _/ |_ _______ |__|_____ ____ ____ __ __ | | __ __ _____
40
 
41
  Triangulum 1B is a collection of pretrained and instruction-tuned generative models, designed for multilingual applications. These models are trained using synthetic datasets based on long chains of thought, enabling them to perform complex reasoning tasks effectively.
42
 
43
- # **Key Features**
44
 
45
  - **Foundation Model**: Built upon LLaMA's autoregressive language model, leveraging an optimized transformer architecture for enhanced performance.
46
 
@@ -48,6 +48,10 @@ Triangulum 1B is a collection of pretrained and instruction-tuned generative mod
48
 
49
  - **Multilingual Support**: Designed to handle multiple languages, ensuring broad applicability across diverse linguistic contexts.
50
 
 
 
 
 
51
  # **Training Approach**
52
 
53
  1. **Synthetic Datasets**: Utilizes long chain-of-thought synthetic data to enhance reasoning capabilities.
 
40
 
41
  Triangulum 1B is a collection of pretrained and instruction-tuned generative models, designed for multilingual applications. These models are trained using synthetic datasets based on long chains of thought, enabling them to perform complex reasoning tasks effectively.
42
 
43
+ # **Key Features & Model Architecture**
44
 
45
  - **Foundation Model**: Built upon LLaMA's autoregressive language model, leveraging an optimized transformer architecture for enhanced performance.
46
 
 
48
 
49
  - **Multilingual Support**: Designed to handle multiple languages, ensuring broad applicability across diverse linguistic contexts.
50
 
51
+ ---
52
+
53
+ - Llama 3.2 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
54
+
55
  # **Training Approach**
56
 
57
  1. **Synthetic Datasets**: Utilizes long chain-of-thought synthetic data to enhance reasoning capabilities.