{"id":523,"date":"2025-03-10T08:26:16","date_gmt":"2025-03-10T08:26:16","guid":{"rendered":"https:\/\/violethoward.com\/new\/new-open-source-math-model-light-r1-32b-surpasses-equivalent-deepseek-performance-with-only-1000-in-training-costs\/"},"modified":"2025-03-10T08:26:16","modified_gmt":"2025-03-10T08:26:16","slug":"new-open-source-math-model-light-r1-32b-surpasses-equivalent-deepseek-performance-with-only-1000-in-training-costs","status":"publish","type":"post","link":"https:\/\/violethoward.com\/new\/new-open-source-math-model-light-r1-32b-surpasses-equivalent-deepseek-performance-with-only-1000-in-training-costs\/","title":{"rendered":"New open-source math model Light-R1-32B surpasses equivalent DeepSeek performance with only $1000 in training costs"},"content":{"rendered":" \r\n<br><div>\n\t\t\t\t<div id=\"boilerplate_2682874\" class=\"post-boilerplate boilerplate-before\">\n<p><em>Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More<\/em><\/p>\n\n\n\n<hr class=\"wp-block-separator has-css-opacity is-style-wide\"\/>\n<\/div><p>Researchers have introduced Light-R1-32B, a new open-source AI model optimized to solve advanced math problems. It is now available on Hugging Face under a permissive Apache 2.0 license \u2014 free for enterprises and researchers to take, deploy, fine-tune or modify as they wish, even for commercial purposes.<\/p>\n\n\n\n<p>The 32-billion parameter (number of model settings) model surpasses the performance of similarly sized (and even larger) open-source models such as DeepSeek-R1-Distill-Llama-70B and DeepSeek-R1-Distill-Qwen-32B on the third-party American Invitational Mathematics Examination (AIME) benchmark that contains 15 math problems designed for extremely advanced students and has an allotted time limit of 3 hours.<\/p>\n\n\n\n<figure class=\"wp-block-image size-full is-resized\"><img fetchpriority=\"high\" decoding=\"async\" width=\"599\" height=\"325\" src=\"https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/03\/Screenshot-2025-03-05-at-2.17.52%E2%80%AFPM.png\" alt=\"\" class=\"wp-image-2998844\" style=\"width:840px;height:auto\" srcset=\"https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/03\/Screenshot-2025-03-05-at-2.17.52\u202fPM.png 599w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/03\/Screenshot-2025-03-05-at-2.17.52\u202fPM.png?resize=300,163 300w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/03\/Screenshot-2025-03-05-at-2.17.52\u202fPM.png?resize=400,217 400w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/03\/Screenshot-2025-03-05-at-2.17.52\u202fPM.png?resize=578,314 578w\" sizes=\"(max-width: 599px) 100vw, 599px\"\/><\/figure>\n\n\n\n<p>Developed by Liang Wen, Fenrui Xiao, Xin He, Yunke Cai, Qi An, Zhenyu Duan, Yimin Du, Junchen Liu, Lifu Tang, Xiaowei Lv, Haosheng Zou, Yongchao Deng, Shousheng Jia and Xiangzheng Zhang, the model surpasses previous open-source alternatives on competitive math benchmarks.<\/p>\n\n\n\n<p>Incredibly, the researchers completed the model\u2019s training in fewer than six hours on 12 Nvidia H800 GPUs at an estimated total cost of $1,000. This makes Light-R1-32B one of the most accessible and practical approaches for developing high-performing math-specialized AI models. However, it\u2019s important to remember that the model was trained on a variant of Alibaba\u2019s open-source Qwen 2.5-32B-Instruct, which itself is presumed to have had much higher upfront training costs.<\/p>\n\n\n\n<p>Alongside the model, the team has released its training datasets and scripts and evaluation tools, providing a transparent and accessible framework for building math-focused AI models.<\/p>\n\n\n\n<p>The arrival of Light-R1-32B follows similar efforts from rivals, such as Microsoft Orca-Math.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"h-a-new-math-king-emerges\">A new math king emerges<\/h2>\n\n\n\n<p>To help Light-R1-32B tackle complex mathematical reasoning, the researchers trained on a model that wasn\u2019t equipped with long-chain-of-thought (COT) reasoning. They applied curriculum-based supervised fine-tuning (SFT) and direct preference otptimization (DPO) to refine its problem-solving capabilities.<\/p>\n\n\n\n<p>When evaluated, Light-R1-32B achieved 76.6 on AIME24 and 64.6 on AIME25, surpassing DeepSeek-R1-Distill-Qwen-32B, which scored 72.6 and 54.9, respectively. <\/p>\n\n\n\n<p>This improvement suggests that the curriculum-based training approach effectively enhances mathematical reasoning, even when training from models that initially lack long COT.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"h-fair-benchmarking\">Fair benchmarking<\/h2>\n\n\n\n<p>To ensure fair benchmarking, the researchers decontaminated training data against common reasoning benchmarks, including AIME24\/25, MATH-500 and GPQA Diamond, preventing data leakage.<\/p>\n\n\n\n<p>They also implemented difficulty-based response filtering using DeepScaleR-1.5B-preview, ultimately forming a 76,000-example dataset for the first stage of supervised fine-tuning. A second, more challenging dataset of 3,000 examples further improved performance.<\/p>\n\n\n\n<p>After training, the team merged multiple trained versions of Light-R1-32B, leading to additional gains. Notably, the model maintains strong generalization abilities on scientific reasoning tasks (GPQA), despite being math-specialized.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"h-how-enterprises-can-benefit\">How enterprises can benefit<\/h2>\n\n\n\n<p>Light-R1-32B is released under the Apache License 2.0, a permissive open-source license that allows free use, modification and commercial deployment without requiring derivative works to be open-sourced. This makes it an attractive option for enterprises, AI developers and software engineers looking to integrate or customize the model for proprietary applications.<\/p>\n\n\n\n<p>The license also includes a royalty-free, worldwide patent grant, reducing legal risks for businesses while discouraging patent disputes. Companies can freely deploy Light-R1-32B in commercial products, maintaining full control over their innovations while benefiting from an open and transparent AI ecosystem.<\/p>\n\n\n\n<p>For CEOs, CTOs and IT leaders, Apache 2.0 ensures cost efficiency and vendor independence, eliminating licensing fees and restrictive dependencies on proprietary AI solutions. AI developers and engineers gain the flexibility to fine-tune, integrate and extend the model without limitations, making it ideal for specialized math reasoning, research and enterprise AI applications. <\/p>\n\n\n\n<p>However, as the license provides no warranty or liability coverage, organizations should conduct their own security, compliance and performance assessments before deploying Light-R1-32B in critical environments.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"h-transparency-in-low-cost-training-and-optimization-for-math-problem-solving\">Transparency in low-cost training and optimization for math problem solving<\/h2>\n\n\n\n<p>The researchers emphasize that Light-R1-32B provides a validated, cost-effective way to train strong long CoT models in specialized domains.<\/p>\n\n\n\n<p>By sharing their methodology, training data and code, they aim to lower cost barriers for high-performance AI development. Looking ahead, they plan to explore reinforcement learning (RL) to further enhance the model\u2019s reasoning capabilities.<\/p>\n\n\n\n\n<div id=\"boilerplate_2660155\" class=\"post-boilerplate boilerplate-after\"><div class=\"Boilerplate__newsletter-container vb\">\n<div class=\"Boilerplate__newsletter-main\">\n<p><strong>Daily insights on business use cases with VB Daily<\/strong><\/p>\n<p class=\"copy\">If you want to impress your boss, VB Daily has you covered. We give you the inside scoop on what companies are doing with generative AI, from regulatory shifts to practical deployments, so you can share insights for maximum ROI.<\/p>\n<p class=\"Form__newsletter-legal\">Read our Privacy Policy<\/p>\n<p class=\"Form__success\" id=\"boilerplateNewsletterConfirmation\">\n\t\t\t\t\tThanks for subscribing. Check out more VB newsletters here.\n\t\t\t\t<\/p>\n<p class=\"Form__error\">An error occured.<\/p>\n<\/p><\/div>\n<div class=\"image-container\">\n\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/venturebeat.com\/wp-content\/themes\/vb-news\/brand\/img\/vb-daily-phone.png\" alt=\"\"\/>\n\t\t\t\t<\/div>\n<\/p><\/div>\n<\/div>\t\t\t<\/div>\r\n<br>\r\n<br><a href=\"https:\/\/venturebeat.com\/ai\/new-open-source-math-model-light-r1-32b-surpasses-equivalent-deepseek-performance-with-only-1000-in-training-costs\/\">Source link <\/a>","protected":false},"excerpt":{"rendered":"<p>Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Researchers have introduced Light-R1-32B, a new open-source AI model optimized to solve advanced math problems. It is now available on Hugging Face under a permissive Apache 2.0 license \u2014 free for enterprises and researchers to take, [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":524,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[33],"tags":[],"class_list":["post-523","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ai-automation"],"aioseo_notices":[],"jetpack_featured_media_url":"https:\/\/violethoward.com\/new\/wp-content\/uploads\/2025\/03\/cfr0z3n_humanoid_robot_wearing_crown_teaches_advanced_math_wi_99a27404-b5e3-4e99-8f76-dd1865a3282f_1.png","jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/posts\/523","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/comments?post=523"}],"version-history":[{"count":0,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/posts\/523\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/media\/524"}],"wp:attachment":[{"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/media?parent=523"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/categories?post=523"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/tags?post=523"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}<!-- This website is optimized by Airlift. Learn more: https://airlift.net. Template:. Learn more: https://airlift.net. Template: 69b0ea1f46fa5c3231e56837. Config Timestamp: 2026-03-11 04:05:51 UTC, Cached Timestamp: 2026-04-08 05:06:05 UTC -->