Qwen 0.6B Connections GRPO

A fine-tune of Qwen/Qwen3-0.6B to be really good at NYT Connections.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

CONNECTIONS_PROMPT = """
You are an expert puzzle solving model.
Find groups of words that are related to each other. Each group is four words long. There are exactly four groups in total.
You may only use each word in one group.
Respond in the following format:
<answer>
<group>
...
</group>
<group>
...
</group>
</answer>
# Example

User: candle, crayon, honeycomb, seal, defense, excuse, out, reason, kettles, mittens, raindrops, whiskers, canine, fang, molar, tusk
<answer>
<group> candle, crayon, honeycomb, seal</group>
<group> kettles, mittens, raindrops, whiskers</group>
<group> canine, fang, molar, tusk</group>
<group> defense, excuse, out, reason</group>
</answer>

"""


chat = [
    {"role": "system", "content": CONNECTIONS_PROMPT},
    {"role": "user", "content": "candle, crayon, honeycomb, seal, defense, excuse, out, reason, kettles, mittens, raindrops, whiskers, canine, fang, molar, tusk"}
]


pipeline = pipeline(task="text-generation", model="roborovski/connections-qwen-3-0.6B", torch_dtype=torch.bfloat16, device_map="auto")
response = pipeline(chat, max_new_tokens=512)

Downloads last month
7
Safetensors
Model size
596M params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for roborovski/connections-qwen-3-0.6B

Finetuned
Qwen/Qwen3-0.6B
Finetuned
(272)
this model