{"id":4373,"date":"2025-11-13T00:32:31","date_gmt":"2025-11-13T00:32:31","guid":{"rendered":"https:\/\/violethoward.com\/new\/weibos-new-open-source-ai-model-vibethinker-1-5b-outperforms-deepseek-r1-on-7800-post-training-budget\/"},"modified":"2025-11-13T00:32:31","modified_gmt":"2025-11-13T00:32:31","slug":"weibos-new-open-source-ai-model-vibethinker-1-5b-outperforms-deepseek-r1-on-7800-post-training-budget","status":"publish","type":"post","link":"https:\/\/violethoward.com\/new\/weibos-new-open-source-ai-model-vibethinker-1-5b-outperforms-deepseek-r1-on-7800-post-training-budget\/","title":{"rendered":"Weibo's new open source AI model VibeThinker-1.5B outperforms DeepSeek-R1 on $7,800 post-training budget"},"content":{"rendered":"


\n
<\/p>\n

Another day in late 2025, another impressive result from a Chinese company in open source artificial intelligence.<\/p>\n

Chinese social networking company Weibo's AI division recently released its open source VibeThinker-1.5B\u2014a 1.5 billion parameter large language model (LLM) that is a fine-tuned variant of rival Chinese tech firm Alibaba's Qwen2.5-Math-1.5B. <\/p>\n

It's available now for free download and usage by researchers and enterprise developers\u2014even for commercial purposes\u2014under a permissive MIT License on Hugging Face, GitHub and ModelScope, with a technical report on open access science publishing site arxiv.org.<\/p>\n

And yet, despite its compact size, VibeThinker-1.5B achieves benchmark-topping reasoning performance on math and code tasks, rivaling or surpassing models hundreds of times its size, even outperforming Chinese rival DeepSeek's famed R1 that went viral at the start of this year\u2014a 671-billion parameter model\u2014on formal reasoning benchmark.<\/p>\n

It further eclipses Mistral AI's Magistral Medium and holds its own against Anthropic's Claude Opus 4 and OpenAI's gpt-oss-20B Medium, all while requiring a fraction of the infrastructure and investment.<\/p>\n

It also does so having been post-trained on a budget of merely $7800 USD for compute resources (3900 GPU hours on Nvidia H800s) \u2014 far less than the tens, or even hundreds, of thousands of dollars typically required to fine-tune models of similar or larger scale.<\/p>\n

Recall this is not the total cost of the model's development, however: LLMs are trained in stages. First comes pre-training, when the model learns basic language structure and general knowledge by predicting the next word across enormous amounts of text from the internet, books, and articles. This gives it fluency but not much sense of how to follow instructions or hold a conversation<\/p>\n

Post-training comes next, using much smaller, higher-quality datasets\u2014typically collections of example questions, prompts, and expert-written answers\u2014to teach the model how to respond helpfully, reason through problems, and align with human expectations. Still, Weibo's post-training cost effectiveness on VibeThinker-1.5B is noteworthy and should be commended.<\/p>\n

The open-source release upends assumptions about parameter scale, compute intensity, and the minimum viable size for high-performance LLMs.<\/p>\n

A Different Training Approach: Spectrum-to-Signal<\/b><\/h3>\n

VibeThinker-1.5B owes its performance not to scale, but to the training framework behind it: the Spectrum-to-Signal Principle (SSP).<\/p>\n

Instead of optimizing a model purely for single-answer correctness (Pass@1), the SSP framework decouples supervised fine-tuning (SFT) and reinforcement learning (RL) into two distinct phases with different goals:<\/p>\n