This project showcases fine-tuning of the Qwen2.5-3B model using the Chain-of-Thought (CoT) prompting technique to enhance mathematical reasoning capabilities. With efficient parameter-efficient fine-tuning (PEFT) via LoRA and the high-speed Unsloth framework, the model is optimized for clarity, depth, and accuracy in problem-solving.
-
Updated
Apr 14, 2025 - Jupyter Notebook