Llama 3.2 3B Reasoning Model

Model Details

Base Model: Meta Llama 3.2 3B Fine-tuning: Full-weight training on 13.9k DeepSeek R1 reasoning examples Training Infrastructure: A100 GPU with bf16 precision Context Length: 131,072 tokens Reasoning Format: Structured thinking with <think></think> tags

Usage

I recommend you use the GGUF version of this model it's easier that way and because this repo has multiple checkpoints.

Sampling Parameters

./llama-cli -m model-q4_0.gguf \
  --temp 0.3 \
  --top-p 0.9 \
  --top-k 40 \
  --repeat-penalty 1.15 \
  -p "Your prompt here" \
  -n 1024

Expected Output Format

The model will structure its responses with reasoning tags:

<think>
Let me solve this step by step...
Speed = Distance / Time
Speed = 300km / 4 hours = 75 km/h
</think>
The average speed of the train is 75 km/h (kilometers per hour).

Model Capabilities

Strengths:

  • Mathematical reasoning and calculations
  • Step-by-step problem solving
  • Logical analysis and deduction
  • Code reasoning and debugging
  • Scientific problem solving

Limitations:

  • May generate verbose reasoning for simple questions
  • Occasional repetition in thinking process
  • Not trained for specific domain knowledge beyond general reasoning

License

This model is based on Llama 3.2 and follows Meta's licensing terms.

Downloads last month
20
Safetensors
Model size
3B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for x1nx3r/Llama-3.2-3B-thinking-v3-13.9K

Finetuned
(370)
this model

Dataset used to train x1nx3r/Llama-3.2-3B-thinking-v3-13.9K