metadata
license: other
license_name: glm-4
license_link: LICENSE
language:
- zh
- en
pipeline_tag: text-generation
tags:
- glm
- edge
inference: false
Glm-Edge-Chat-4B-GGUF
中文阅读, 点击这里
Inference with Ollama
Installation
The code for adapting this model is actively being integrated into the official llama.cpp
. You can test it using the
following adapted version:
git clone https://github.com/piDack/llama.cpp -b support_glm_edge_model
cmake -B build -DGGML_CUDA=ON # Or enable other acceleration hardware
cmake --build build -- -j
Inference
After installation, you can start the GLM-Edge Chat model using the following command:
llama-cli -m <path>/model.gguf -p "<|user|>\nhi<|assistant|>\n" -ngl 999
In the command-line interface, you can interact with the model by entering your requests, and the model will provide the corresponding responses.
License
The usage of this model’s weights is subject to the terms outlined in the LICENSE.