GGUF
Inference Endpoints
conversational
siacus commited on
Commit
4f74ff0
1 Parent(s): 194e051

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -0
README.md ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - siacus/dv_subject
5
+ base_model:
6
+ - meta-llama/Llama-2-7b-chat-hf
7
+ new_version: siacus/llama-2-7b-dv
8
+ ---
9
+ The data used to train the model are on Huggingface under [siacus/dv_subject](https://huggingface.co/datasets/siacus/dv_subject)
10
+
11
+ The small-dv version of the fine-tuned model works on a training-set of 5,000 randomly sampled data.
12
+
13
+ The large version works on the whole 76.1K training records.
14
+
15
+ The test set is of size 32.6K rows.
16
+
17
+ F16 version from merged weights created with llama.cpp on a CUDA GPU and the 4bit quantized version created on a Mac M2 Ultra Metal architecture. If you want to use the 4bit quantized version on CUDA, please quantize it directly from the F16 version.
18
+
19
+ For more information about this model refer the main repository for the supplementary material of the manuscript
20
+ [Rethinking Scale: The Efficacy of Fine-Tuned Open-Source LLMs in Large-Scale Reproducible Social Science Research](https://arxiv.org/abs/2411.00890).