dahara1's picture
Update README.md
3c4ac90 verified
|
raw
history blame
5.95 kB
metadata
language:
  - ja

本モデルについて about this model.

Qwen/Qwen2.5-3B-Instruct日本語が多く含まれる重要度行列(iMatrix)を使って量子化し、超長文(32K以上)要約を可能にしたgguf版です。日本語対応能力が多めに保持されている事を期待しています。
This is a gguf version of Qwen/Qwen2.5-3B-Instruct that has been quantized using importance matrix (iMatrix) that contains a lot of Japanese to enable summarization of long texts (over 32K). We hope that it retains a large amount of Japanese support.

少なくともQwen2.5-3B-Instruct-gguf-japanese-imatrix-128K/Qwen2.5-3B-Instruct-Q8_0-f16.ggufが32Kトークンを超える超長文を正しく要約できる事を確認済です。
It has been confirmed that at least Qwen2.5-3B-Instruct-gguf-japanese-imatrix-128K/Qwen2.5-3B-Instruct-Q8_0-f16.gguf can correctly summarize extremely long texts exceeding 32K tokens.

128Kコンテキスト延長についてはunsloth/Qwen2.5-Coder-32B-Instruct-128K-GGUFの指摘を参考にしています。ありがとう。
Regarding the 128K context extension, I have taken note of the suggestion made by unsloth/Qwen2.5-Coder-32B-Instruct-128K-GGUF. Thank you.

For ollama users

ollama ユーザーはFAQを参考にしてcontext window sizeパラメーターを修正してください。
If you use ollama, check FAQ and set context window size parameter like below.

/set parameter num_ctx 40960

or API

curl http://..../api/generate -d '{
  "model": ".....",
  "prompt": "......",
  "options": {
    "num_ctx": 40960
  }
}'

あなたが他のツールを使っている場合、同様にあなたの使っているツールのマニュアルを調べて、コンテキストウインドウサイズを延長する事を忘れないでください
ただし、コンテキストサイズを必要以上に大きくするとモデルの実行速度が低下するので注意してください
本モデルは理論上、最大値128K(131072)に設定できますが、実行速度と品質に影響が出る事が考えられます

If you are using other tools, be sure to extend the context window size as well, by consulting the manual of your tool.
But please note that increasing the context window size more than necessary will slow down the model's execution speed.
In theory, this model can be set to the maximum value of 128K(131072), but this may affect execution speed and quality.

Sample llama.cpp script

以下は、Wikipediaの約50,000文字(34.8Kトークン)の記事を取得して内容を要約するサンプルです
Below is a sample that retrieves a Wikipedia article of about 50,000 Japanese characters(34.8K tokens) and summarizes its contents.

llama.cpp server command sample.

./llama.cpp/build/bin/Release/llama-server.exe -m ./Qwen2.5-3B-Instruct-Q8_0-f16.gguf -c 40960

llama.cpp client script sample.

import transformers
import requests
import json
from transformers import AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-3B-Instruct")

url = "https://ja.wikipedia.org/wiki/%E7%94%B7%E3%81%AE%E5%A8%98"

def get_wikipedia_text(url):
    response = requests.get(url)
    if response.status_code == 200:
        from bs4 import BeautifulSoup
        soup = BeautifulSoup(response.text, 'html.parser')
        paragraphs = soup.find_all('p')
        text = "\n".join([p.get_text() for p in paragraphs])
        return text
    else:
        raise Exception(f"Failed to fetch the article. Status code: {response.status_code}")
        
if __name__ == "__main__":

    html_text = get_wikipedia_text(url)
    #html_text = html_text[:40000]

    instruct = "### 指示\n\n上記の文章を日本語で3行で要約してください"

    # instruct first version
    messages  = [
        {"role": "system", "content": "You are a helpful assistant."},
        {"role": "user", "content": instruct + "\n\n" + html_text},
    ]

    # instruct last version
    messages  = [
        {"role": "system", "content": "You are a helpful assistant."},
        {"role": "user", "content": html_text  + "\n\n" + instruct},
    ]
    
    prompt = tokenizer.apply_chat_template(
            messages,
            add_generation_prompt=True,
            tokenize=False
    )
    print(prompt)

    payload = {
            "prompt": prompt,
            "n_predict": 512
    }

    url = "http://localhost:8080/completion"
    headers = {
        "Content-Type": "application/json"
    }

    response = requests.post(url, headers=headers, data=json.dumps(payload))
    if response.status_code != 200:
        print(f"Error: {response.text}")

    response_data = response.json()

    response_content = response_data.get('content', '').strip()
    print(response_content)

出力結果(output sample)

This 128K model

128K instruct first version
128K instruct first version

128K instruct last version
128K instruct last version

Standard 32K model

32K instruct first version
32K instruct first version

32K instruct last version
32K instruct last version

32K instruct first versionでは要約指示がコンテキスト外になっており、指示が無視されている事に注目してください
Notice that in the 32K instruct first version the summary instruction is out of context and the instruction is ignored.