prithivMLmods
commited on
Commit
•
03eda0f
1
Parent(s):
b2f50f6
Upload 6 files
Browse files- .gitattributes +4 -0
- Llama-Doctor-3.2-3B-Instruct.F16.gguf +3 -0
- Llama-Doctor-3.2-3B-Instruct.Q4_K_M.gguf +3 -0
- Llama-Doctor-3.2-3B-Instruct.Q5_K_M.gguf +3 -0
- Llama-Doctor-3.2-3B-Instruct.Q8_0.gguf +3 -0
- Modelfile +57 -0
- config.json +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
Llama-Doctor-3.2-3B-Instruct.F16.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
Llama-Doctor-3.2-3B-Instruct.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
Llama-Doctor-3.2-3B-Instruct.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
Llama-Doctor-3.2-3B-Instruct.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
Llama-Doctor-3.2-3B-Instruct.F16.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c01d5118ff997799f1c8d2552381eab47a21bca12b88274c3b7e58a9e76e8b29
|
3 |
+
size 6433688384
|
Llama-Doctor-3.2-3B-Instruct.Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fedc9b6eae2f732cf464326f69d2bdcad22c6a599009cc176d1816548c0e9dfa
|
3 |
+
size 2019377984
|
Llama-Doctor-3.2-3B-Instruct.Q5_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e1b0182648ea29589ccf2dc8f7865892b8b287087f2f64445b1b6f30d58cea80
|
3 |
+
size 2322154304
|
Llama-Doctor-3.2-3B-Instruct.Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fe07d624c022aebd0a84098725792aac8384eeaa7fa4b6ad296469e4c3b3b526
|
3 |
+
size 3421899584
|
Modelfile
ADDED
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
FROM /content/prithivMLmods/Llama-Doctor-3.2-3B-Instruct-GGUF/Llama-Doctor-3.2-3B-Instruct.F16.gguf
|
3 |
+
TEMPLATE """{{ if .Messages }}
|
4 |
+
{{- if or .System .Tools }}<|start_header_id|>system<|end_header_id|>
|
5 |
+
{{- if .System }}
|
6 |
+
|
7 |
+
{{ .System }}
|
8 |
+
{{- end }}
|
9 |
+
{{- if .Tools }}
|
10 |
+
|
11 |
+
You are a helpful assistant with tool calling capabilities. When you receive a tool call response, use the output to format an answer to the original use question.
|
12 |
+
{{- end }}
|
13 |
+
{{- end }}<|eot_id|>
|
14 |
+
{{- range $i, $_ := .Messages }}
|
15 |
+
{{- $last := eq (len (slice $.Messages $i)) 1 }}
|
16 |
+
{{- if eq .Role "user" }}<|start_header_id|>user<|end_header_id|>
|
17 |
+
{{- if and $.Tools $last }}
|
18 |
+
|
19 |
+
Given the following functions, please respond with a JSON for a function call with its proper arguments that best answers the given prompt.
|
20 |
+
|
21 |
+
Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}. Do not use variables.
|
22 |
+
|
23 |
+
{{ $.Tools }}
|
24 |
+
{{- end }}
|
25 |
+
|
26 |
+
{{ .Content }}<|eot_id|>{{ if $last }}<|start_header_id|>assistant<|end_header_id|>
|
27 |
+
|
28 |
+
{{ end }}
|
29 |
+
{{- else if eq .Role "assistant" }}<|start_header_id|>assistant<|end_header_id|>
|
30 |
+
{{- if .ToolCalls }}
|
31 |
+
|
32 |
+
{{- range .ToolCalls }}{"name": "{{ .Function.Name }}", "parameters": {{ .Function.Arguments }}}{{ end }}
|
33 |
+
{{- else }}
|
34 |
+
|
35 |
+
{{ .Content }}{{ if not $last }}<|eot_id|>{{ end }}
|
36 |
+
{{- end }}
|
37 |
+
{{- else if eq .Role "tool" }}<|start_header_id|>ipython<|end_header_id|>
|
38 |
+
|
39 |
+
{{ .Content }}<|eot_id|>{{ if $last }}<|start_header_id|>assistant<|end_header_id|>
|
40 |
+
|
41 |
+
{{ end }}
|
42 |
+
{{- end }}
|
43 |
+
{{- end }}
|
44 |
+
{{- else }}
|
45 |
+
{{- if .System }}<|start_header_id|>system<|end_header_id|>
|
46 |
+
|
47 |
+
{{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|>
|
48 |
+
|
49 |
+
{{ .Prompt }}<|eot_id|>{{ end }}<|start_header_id|>assistant<|end_header_id|>
|
50 |
+
|
51 |
+
{{ end }}{{ .Response }}{{ if .Response }}<|eot_id|>{{ end }}"""
|
52 |
+
PARAMETER stop "<|start_header_id|>"
|
53 |
+
PARAMETER stop "<|end_header_id|>"
|
54 |
+
PARAMETER stop "<|eot_id|>"
|
55 |
+
PARAMETER stop "<|eom_id|>"
|
56 |
+
PARAMETER temperature 1.5
|
57 |
+
PARAMETER min_p 0.1
|
config.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"model_type": "llama"
|
3 |
+
}
|