🧠 Qwen2.5 + GRPO — Structured Reasoning Model

A fine-tuned version of Qwen2.5 trained with GRPO (Group Relative Policy Optimization) to reason before it answers — not just pattern-match.


Overview

Most LLMs simulate reasoning by mimicking patterns seen during training. This model is different: it builds a real cognitive path on every response by following a strict, verifiable reasoning protocol enforced through reinforcement learning.

Every response goes through three mandatory stages:

Stage Tag Purpose
📌 Plan <planning> Understand the task and define an approach
🔍 Monitor <monitoring> Reason step by step, show calculations and logic
✅ Evaluate <evaluation> Verify the answer before committing

This isn't chain-of-thought bolted on top — the reasoning protocol is baked in via RL.

System Prompt

SYSTEM_PROMPT = """
You are an AI assistant that MUST produce structured reasoning.
Your response MUST EXACTLY follow this format:
<think>
<planning>
...
</planning>
<monitoring>
...
</monitoring>
<evaluation>
...
</evaluation>
</think>
<output>
...
</output>
FORMAT RULES:
1. The <think> block must contain exactly three sections in this order:
   <planning>, <monitoring>, <evaluation>
2. Each section must contain detailed reasoning in full sentences.
3. Minimum reasoning length:
   - <planning>: at least 40 tokens
   - <monitoring>: at least 80 tokens
   - <evaluation>: at least 40 tokens
4. The <monitoring> section MUST show explicit reasoning steps,
   including calculations, derivations, or logical deductions.
5. Generic placeholder phrases are forbidden, including:
   - "analyze the problem"
   - "determine the strategy"
   - "verify the solution"
   - "check correctness"
6. The reasoning must explicitly reference values, equations,
   or logical relationships from the problem.
7. The <output> section must contain ONLY the final answer.
INVALID RESPONSES:
Responses will be rejected if they contain:
- Empty sections
- Bullet point placeholders
- Generic reasoning
- Missing calculations when required
- Incorrect tag order
The format must always be strictly respected.
"""

Usage

from vllm import SamplingParams

def generate_response(question, choices):
    messages = [
        {
            "role": "system",
            "content": SYSTEM_PROMPT
        },
        {
            "role": "user",
            "content": (
                f"Examine the following question and select the right answer from given options.\n"
                f"The output must be only the number of the option.\n"
                f"Question: {question}\n"
                f"Provided options: {choices}\n"
            )
        }
    ]

    inputs = tokenizer.apply_chat_template(
        messages,
        tokenize=False,
        add_generation_prompt=True,
        return_tensors="pt",
    )

    sampling_params = SamplingParams(
        temperature=0.8,
        top_p=0.95,
        max_tokens=1024,
    )

    output = model.fast_generate(
        [inputs],
        sampling_params=sampling_params,
        lora_request=None,
    )[0].outputs[0].text

    return output

What Makes This Different

Feature Standard LLM This Model
Reasoning method Pattern matching Structured cognitive protocol
Reasoning enforcement None RL-baked (GRPO)
Output format Free-form Strictly validated
Self-verification No Yes — invalid structure = rejected response
Final answer Mixed with reasoning Isolated in <output>

MMLU Benchmark Results

We selected random 100 samples from each subsets of MMLU dataset. Performance across a range of MMLU subject categories:

🎓 College Courses

Subject Accuracy
College Mathematics 50%
College Computer Science 57%
Medicine 67%

🧑‍💼 Professional

Subject Accuracy
Professional Psychology 63%

🏫 High School Courses

Subject Accuracy
Psychology 83%
Computer Science 78%
Management 70%
Mathematics 68%
Statistics 66%
Biology 67%
Chemistry 62%
European History 64%

Results reflect accuracy on MMLU multiple-choice questions using the structured reasoning protocol described above.


Training

  • Base model: Qwen2.5
  • Training method: GRPO (Group Relative Policy Optimization)
  • Objective: Enforce structured reasoning as a non-negotiable output constraint, not a post-hoc addition
Downloads last month
364
Safetensors
Model size
3B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for alibidaran/Qwen_COG_Thinker_Merged

Base model

Qwen/Qwen2.5-3B
Finetuned
(28)
this model

Dataset used to train alibidaran/Qwen_COG_Thinker_Merged