nvidia/Nemotron-Research-Reasoning-Qwen-1.5B
The Nemotron-Research-Reasoning-Qwen-1.5B is a 1.5 billion parameter language model developed by NVIDIA, built on the Qwen architecture with a 131072 token context length. It is specifically optimized for complex reasoning tasks, including mathematical problems, coding challenges, scientific questions, and logic puzzles. Trained using the ProRL algorithm, this model significantly outperforms other 1.5B models and matches or exceeds 7B models in reasoning benchmarks, making it suitable for advanced research and development in AI reasoning.