syngen reasoning logo

πŸ‘‹ SYNGEN REASONING 0.6B

This is a 0.6B parameter LLM designed for synthetic grounded reasoning generation between final model outputs, specifically for dataset modifications, but can be used for multiple use cases which require reasoning.

For example, this model allows you to turn any chat dataset into a reasoning dataset as if it was generated by DeepSeek R1 or Openai's GPT OSS!

πŸ‘€ EXAMPLE DATASET GENERATED WITH IT

https://huggingface.co/datasets/Pinkstack/syngen-reasoning-example-80-smoltalk1

πŸ€” HOW TO USE

You should set the following:

  1. MIN P = 0.05
  2. TOP P = 0.85
  3. TEMPERATURE = 0.55
  4. MAX_NEW_TOKENS = 16384 (can be higher or lower, up to 32k)

Syngen uses the CHATML format, but it is not a chat model. This is how you do interference with it:

<|im_start|>system


THINKING GENERATION ONLY = TRUE

THINKING GENERATION STYLE = (DEEPSEEK_R1 OR GPT_OSS)<|im_end|>
<|im_start|>user
Generate thinking for the following:

The user message was: {human_msg}

The assistant response was: {assistant_msg}

Original system prompt was:{system_msg} (for GPT OSS please put "Reasoning effort: high/medium/low")<|im_end|>
<|im_start|>assistant

here are some examples:

Deepseek r1 possible reasoning

<|im_start|>system


THINKING GENERATION ONLY = TRUE

THINKING GENERATION STYLE = DEEPSEEK_R1<|im_end|>
<|im_start|>user
Generate thinking for the following:

The user message was: Your response should contain at least 3 sentences. Your answer must contain exactly 3 bullet points. Use the markdown bullet points such as: * This is point 1. Include keywords [food, water, shelter] in the response.

The assistant response was:* To ensure survival, one must have access to adequate food which provides the necessary nutrients and energy.
* Water is essential for hydration and maintaining bodily functions, and it is crucial to have a clean and reliable source.
* Shelter protects individuals from the elements and provides a safe environment, which is fundamental for overall well-being.

Original system prompt was:You are an Ai<|im_end|>
<|im_start|>assistant
<generated_thinking_dpsk>
Okay, the user wants me to generate a response with three bullet points using markdown. The keywords they provided are food, water, and shelter. Let's think about how to structure each bullet point.

First, maybe start with something related to food. Need to make sure it includes all three words. Maybe mention needing food to sustain life or staying hydrated? That covers both water and food indirectly but directly links them together. Then talk about shelter because without protection, you can't eat properly or stay alive. 

Wait, do I need to include water again here? If shelter means avoiding dehydration, then yes. But let's check if the first bullet starts with "To ensure..." that ties everything together. Also, keep it concise but cover all required keywords. Let me draft:

* To ensure survival, one must have access to adequate food which provides the necessary nutrients and energy.
* Water is essential for hydration and maintaining bodily functions; it’s critical to have a clean and reliable source.
* Shelter protects individuals from the elements and offers a safe environmentβ€”key factors for overall well-being.

That works. Each bullet has two key points (food and water) and shelters. Check keyword usage: food, water, shelter. Yes. Three bullets. Good. Make sure each sentence ends with period. Alright.
</generated_thinking_dpsk><|im_end|>

GPT OSS possible reasoning

<|im_start|>system


THINKING GENERATION ONLY = TRUE

THINKING GENERATION STYLE = GPT_OSS<|im_end|>
<|im_start|>user
Generate thinking for the following:

The user message was: Hello!

The assistant response was: Hi! How may I assist you today?

Original system prompt was:You are an Ai
Reasoning effort: low<|im_end|>
<|im_start|>assistant
<generated_thinking_gpt>
Must produce a simple greeting.
</generated_thinking_gpt><|im_end|>

Liked the model? Need help with it? Do you know how to improve it further? Please make a post in the community tab.

Downloads last month
85
Safetensors
Model size
0.6B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Pinkstack/syngen-reasoning-0.6b

Finetuned
(344)
this model
Quantizations
1 model

Dataset used to train Pinkstack/syngen-reasoning-0.6b