prithivMLmods commited on
Commit
03eda0f
1 Parent(s): b2f50f6

Upload 6 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Llama-Doctor-3.2-3B-Instruct.F16.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Llama-Doctor-3.2-3B-Instruct.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Llama-Doctor-3.2-3B-Instruct.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Llama-Doctor-3.2-3B-Instruct.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
Llama-Doctor-3.2-3B-Instruct.F16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c01d5118ff997799f1c8d2552381eab47a21bca12b88274c3b7e58a9e76e8b29
3
+ size 6433688384
Llama-Doctor-3.2-3B-Instruct.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fedc9b6eae2f732cf464326f69d2bdcad22c6a599009cc176d1816548c0e9dfa
3
+ size 2019377984
Llama-Doctor-3.2-3B-Instruct.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1b0182648ea29589ccf2dc8f7865892b8b287087f2f64445b1b6f30d58cea80
3
+ size 2322154304
Llama-Doctor-3.2-3B-Instruct.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe07d624c022aebd0a84098725792aac8384eeaa7fa4b6ad296469e4c3b3b526
3
+ size 3421899584
Modelfile ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ FROM /content/prithivMLmods/Llama-Doctor-3.2-3B-Instruct-GGUF/Llama-Doctor-3.2-3B-Instruct.F16.gguf
3
+ TEMPLATE """{{ if .Messages }}
4
+ {{- if or .System .Tools }}<|start_header_id|>system<|end_header_id|>
5
+ {{- if .System }}
6
+
7
+ {{ .System }}
8
+ {{- end }}
9
+ {{- if .Tools }}
10
+
11
+ You are a helpful assistant with tool calling capabilities. When you receive a tool call response, use the output to format an answer to the original use question.
12
+ {{- end }}
13
+ {{- end }}<|eot_id|>
14
+ {{- range $i, $_ := .Messages }}
15
+ {{- $last := eq (len (slice $.Messages $i)) 1 }}
16
+ {{- if eq .Role "user" }}<|start_header_id|>user<|end_header_id|>
17
+ {{- if and $.Tools $last }}
18
+
19
+ Given the following functions, please respond with a JSON for a function call with its proper arguments that best answers the given prompt.
20
+
21
+ Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}. Do not use variables.
22
+
23
+ {{ $.Tools }}
24
+ {{- end }}
25
+
26
+ {{ .Content }}<|eot_id|>{{ if $last }}<|start_header_id|>assistant<|end_header_id|>
27
+
28
+ {{ end }}
29
+ {{- else if eq .Role "assistant" }}<|start_header_id|>assistant<|end_header_id|>
30
+ {{- if .ToolCalls }}
31
+
32
+ {{- range .ToolCalls }}{"name": "{{ .Function.Name }}", "parameters": {{ .Function.Arguments }}}{{ end }}
33
+ {{- else }}
34
+
35
+ {{ .Content }}{{ if not $last }}<|eot_id|>{{ end }}
36
+ {{- end }}
37
+ {{- else if eq .Role "tool" }}<|start_header_id|>ipython<|end_header_id|>
38
+
39
+ {{ .Content }}<|eot_id|>{{ if $last }}<|start_header_id|>assistant<|end_header_id|>
40
+
41
+ {{ end }}
42
+ {{- end }}
43
+ {{- end }}
44
+ {{- else }}
45
+ {{- if .System }}<|start_header_id|>system<|end_header_id|>
46
+
47
+ {{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|>
48
+
49
+ {{ .Prompt }}<|eot_id|>{{ end }}<|start_header_id|>assistant<|end_header_id|>
50
+
51
+ {{ end }}{{ .Response }}{{ if .Response }}<|eot_id|>{{ end }}"""
52
+ PARAMETER stop "<|start_header_id|>"
53
+ PARAMETER stop "<|end_header_id|>"
54
+ PARAMETER stop "<|eot_id|>"
55
+ PARAMETER stop "<|eom_id|>"
56
+ PARAMETER temperature 1.5
57
+ PARAMETER min_p 0.1
config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "model_type": "llama"
3
+ }