Llama 3.2 3B Reasoning Model
Model Details
Base Model: Meta Llama 3.2 3B
Fine-tuning: Full-weight training on 13.9k DeepSeek R1 reasoning examples
Training Infrastructure: A100 GPU with bf16 precision
Context Length: 131,072 tokens
Reasoning Format: Structured thinking with <think></think> tags
Usage
I recommend you use the GGUF version of this model it's easier that way and because this repo has multiple checkpoints.
Sampling Parameters
./llama-cli -m model-q4_0.gguf \
--temp 0.3 \
--top-p 0.9 \
--top-k 40 \
--repeat-penalty 1.15 \
-p "Your prompt here" \
-n 1024
Expected Output Format
The model will structure its responses with reasoning tags:
<think>
Let me solve this step by step...
Speed = Distance / Time
Speed = 300km / 4 hours = 75 km/h
</think>
The average speed of the train is 75 km/h (kilometers per hour).
Model Capabilities
Strengths:
- Mathematical reasoning and calculations
- Step-by-step problem solving
- Logical analysis and deduction
- Code reasoning and debugging
- Scientific problem solving
Limitations:
- May generate verbose reasoning for simple questions
- Occasional repetition in thinking process
- Not trained for specific domain knowledge beyond general reasoning
License
This model is based on Llama 3.2 and follows Meta's licensing terms.
- Downloads last month
- 20
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for x1nx3r/Llama-3.2-3B-thinking-v3-13.9K
Base model
meta-llama/Llama-3.2-3B