benhaotang commited on
Commit
b7e43a3
·
verified ·
1 Parent(s): 2194f0c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +110 -1
README.md CHANGED
@@ -3,12 +3,121 @@ base_model: rombodawg/Rombos-Coder-V2.5-Qwen-7b
3
  tags:
4
  - llama-cpp
5
  - gguf-my-repo
 
6
  ---
7
 
8
  # benhaotang/Rombos-Coder-V2.5-Qwen-7b-Q8_0-GGUF
9
  This model was converted to GGUF format from [`rombodawg/Rombos-Coder-V2.5-Qwen-7b`](https://huggingface.co/rombodawg/Rombos-Coder-V2.5-Qwen-7b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
10
  Refer to the [original model card](https://huggingface.co/rombodawg/Rombos-Coder-V2.5-Qwen-7b) for more details on the model.
11
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  ## Use with llama.cpp
13
  Install llama.cpp through brew (works on Mac and Linux)
14
 
@@ -47,4 +156,4 @@ Step 3: Run inference through the main binary.
47
  or
48
  ```
49
  ./llama-server --hf-repo benhaotang/Rombos-Coder-V2.5-Qwen-7b-Q8_0-GGUF --hf-file rombos-coder-v2.5-qwen-7b-q8_0.gguf -c 2048
50
- ```
 
3
  tags:
4
  - llama-cpp
5
  - gguf-my-repo
6
+ - cline
7
  ---
8
 
9
  # benhaotang/Rombos-Coder-V2.5-Qwen-7b-Q8_0-GGUF
10
  This model was converted to GGUF format from [`rombodawg/Rombos-Coder-V2.5-Qwen-7b`](https://huggingface.co/rombodawg/Rombos-Coder-V2.5-Qwen-7b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
11
  Refer to the [original model card](https://huggingface.co/rombodawg/Rombos-Coder-V2.5-Qwen-7b) for more details on the model.
12
 
13
+ ## Use with Cline and Ollama
14
+
15
+ use this template file from https://github.com/maryasov/ollama-models-instruct-for-cline
16
+
17
+ ```
18
+ FROM rombos-coder-v2.5-qwen-7b-q6_k.gguf
19
+ TEMPLATE """{{- /* Initial system message with core instructions */ -}}
20
+ {{- if .Messages }}
21
+ {{- if or .System .Tools }}
22
+ <|im_start|>system
23
+ {{- if .System }}
24
+ {{ .System }}
25
+ {{- end }} {{- if .Tools }}
26
+ # Tools and XML Schema
27
+ You have access to the following tools. Each tool must be used according to this XML schema:
28
+
29
+ <tools>
30
+ {{- range .Tools }}
31
+ {{ .Function }}
32
+ {{- end }}
33
+ </tools>
34
+
35
+ ## Tool Use Format
36
+ 1. Think about the approach in <thinking> tags
37
+ 2. Call tool using XML format:
38
+ <tool_name>
39
+ <param_name>value</param_name>
40
+ </tool_name>
41
+ 3. Process tool response from:
42
+ <tool_response>result</tool_response>
43
+ {{- end }}
44
+ <|im_end|>
45
+ {{- end }}
46
+
47
+ {{- /* Message handling loop */ -}}
48
+ {{- range $i, $_ := .Messages }}
49
+ {{- $last := eq (len (slice $.Messages $i)) 1 }}
50
+
51
+ {{- /* User messages */ -}}
52
+ {{- if eq .Role "user" }}
53
+ <|im_start|>user
54
+ {{ .Content }}
55
+ <|im_end|>
56
+
57
+ {{- /* Assistant messages */ -}}
58
+ {{- else if eq .Role "assistant" }}
59
+ <|im_start|>assistant
60
+ {{- if .Content }}
61
+ {{ .Content }}
62
+ {{- else if .ToolCalls }}
63
+ {{- range .ToolCalls }}
64
+ <thinking>
65
+ [Analysis of current state and next steps]
66
+ </thinking>
67
+
68
+ <{{ .Function.Name }}>
69
+ {{- range $key, $value := .Function.Arguments }}
70
+ <{{ $key }}>{{ $value }}</{{ $key }}>
71
+ {{- end }}
72
+ </{{ .Function.Name }}>
73
+ {{- end }}
74
+ {{- end }}
75
+ {{- if not $last }}<|im_end|>{{- end }}
76
+
77
+ {{- /* Tool response handling */ -}}
78
+ {{- else if eq .Role "tool" }}
79
+ <|im_start|>user
80
+ <tool_response>
81
+ {{ .Content }}
82
+ </tool_response>
83
+ <|im_end|>
84
+ {{- end }}
85
+
86
+ {{- /* Prepare for next assistant response if needed */ -}}
87
+ {{- if and (ne .Role "assistant") $last }}
88
+ <|im_start|>assistant
89
+ {{- end }}
90
+ {{- end }}
91
+
92
+ {{- /* Handle single message case */ -}}
93
+ {{- else }}
94
+ {{- if .System }}
95
+ <|im_start|>system
96
+ {{ .System }}
97
+ <|im_end|>
98
+ {{- end }}
99
+
100
+ {{- if .Prompt }}
101
+ <|im_start|>user
102
+ {{ .Prompt }}
103
+ <|im_end|>
104
+ {{- end }}
105
+
106
+ <|im_start|>assistant
107
+ {{- end }}
108
+ {{ .Response }}
109
+ {{- if .Response }}<|im_end|>{{- end }}
110
+ """
111
+ PARAMETER repeat_last_n 64
112
+ PARAMETER repeat_penalty 1.1
113
+ PARAMETER stop "<|im_start|>"
114
+ PARAMETER stop "<|im_end|>"
115
+ PARAMETER stop "<|endoftext|>"
116
+ PARAMETER temperature 0.1
117
+ PARAMETER top_k 40
118
+ PARAMETER top_p 0.9
119
+ ```
120
+
121
  ## Use with llama.cpp
122
  Install llama.cpp through brew (works on Mac and Linux)
123
 
 
156
  or
157
  ```
158
  ./llama-server --hf-repo benhaotang/Rombos-Coder-V2.5-Qwen-7b-Q8_0-GGUF --hf-file rombos-coder-v2.5-qwen-7b-q8_0.gguf -c 2048
159
+ ```