zR
1
e168d9e
metadata
license: other
license_name: glm-4
license_link: LICENSE
language:
  - zh
  - en
pipeline_tag: text-generation
tags:
  - glm
  - edge
inference: false

Glm-Edge-Chat-4B-GGUF

中文阅读, 点击这里

Inference with Ollama

Installation

The code for adapting this model is actively being integrated into the official llama.cpp. You can test it using the following adapted version:

git clone https://github.com/piDack/llama.cpp -b support_glm_edge_model
cmake -B build -DGGML_CUDA=ON # Or enable other acceleration hardware
cmake --build build -- -j 

Inference

After installation, you can start the GLM-Edge Chat model using the following command:

llama-cli -m <path>/model.gguf -p "<|user|>\nhi<|assistant|>\n" -ngl 999

In the command-line interface, you can interact with the model by entering your requests, and the model will provide the corresponding responses.

License

The usage of this model’s weights is subject to the terms outlined in the LICENSE.