parambharat commited on
Commit
5709163
·
1 Parent(s): f9cf95c

chore: remove footnotes header

Browse files
Files changed (1) hide show
  1. rag/rag.py +0 -2
rag/rag.py CHANGED
@@ -65,8 +65,6 @@ The main mix of the training data for the Llama 3 405 billion parameter model is
65
 
66
  Regarding the amount of data used to train the model, the snippets do not provide a specific total volume of data in terms of tokens or bytes. However, they do mention that the model was pre-trained on a large dataset containing knowledge until the end of 2023[^2^]. Additionally, the training process involved pre-training on 2.87 trillion tokens before further adjustments[^3^].
67
 
68
- ## Footnotes
69
-
70
  [^1^]: "Scaling Laws for Data Mix," page 6.
71
  [^2^]: "Pre-Training Data," page 4.
72
  [^3^]: "Initial Pre-Training," page 14.
 
65
 
66
  Regarding the amount of data used to train the model, the snippets do not provide a specific total volume of data in terms of tokens or bytes. However, they do mention that the model was pre-trained on a large dataset containing knowledge until the end of 2023[^2^]. Additionally, the training process involved pre-training on 2.87 trillion tokens before further adjustments[^3^].
67
 
 
 
68
  [^1^]: "Scaling Laws for Data Mix," page 6.
69
  [^2^]: "Pre-Training Data," page 4.
70
  [^3^]: "Initial Pre-Training," page 14.