7 min read

Weibo's VibeThinker‑1.5B: Tiny LLM, Big Wins on Reasoning

AI

ThinkTools Team

AI Research Lead

Introduction

In the fast‑moving world of large language models, a new entrant from China’s social‑media giant Weibo has already begun to rewrite the playbook. The company’s AI division has released VibeThinker‑1.5B, a 1.5‑billion‑parameter model that, according to its creators, eclipses the reasoning prowess of models that are hundreds of times larger. What makes the achievement even more striking is the modest post‑training budget of just $7,800, a fraction of the millions typically required to fine‑tune comparable systems. The model is freely available under an MIT license on Hugging Face, GitHub, and ModelScope, and is accompanied by a technical report on arXiv. This release challenges the entrenched belief that sheer scale is the only path to high‑quality reasoning and opens the door for smaller, more accessible models to tackle logic‑heavy tasks.

Weibo, often dubbed China’s version of X (formerly Twitter), has long been a platform for microblogging, multimedia, and trending topics. With 600 million monthly active users, it sits at the heart of China’s digital public square. Yet the platform’s revenue prospects have been under scrutiny, and the company has pivoted toward creator‑economy monetization, live streaming, and e‑commerce. The launch of VibeThinker‑1.5B signals a strategic shift: Weibo is positioning itself as a serious player in AI research and development, leveraging its user data and in‑house expertise to push the boundaries of what a compact model can achieve.

The model’s performance is not limited to a single domain. On the AIME‑25 math benchmark it scores 74.4, surpassing the 1.09‑trillion‑parameter Kimi K2 by more than ten points. In LiveCodeBench v6, it achieves 51.1, outpacing Claude Opus 4. Even on the GPQA‑Diamond benchmark, where it scores 46.7, it has doubled the performance of its base architecture. These results demonstrate that VibeThinker‑1.5B can compete with, and in some cases outperform, the most powerful closed‑source models, all while consuming a tenth of the compute budget.

The key to this success lies in a novel training framework called the Spectrum‑to‑Signal Principle (SSP). By decoupling supervised fine‑tuning (SFT) and reinforcement learning (RL) into distinct phases—each with its own objective—Weibo’s researchers have managed to amplify the model’s reasoning signal without relying on massive parameter counts. The following sections unpack the mechanics of SSP, the benchmark performance, and the implications for enterprise deployment.

Main Content

The Spectrum‑to‑Signal Principle

The SSP framework rethinks how we train models for reasoning. Traditional pipelines often conflate SFT and RL, optimizing for single‑answer correctness (Pass@1) in a single stage. SSP, by contrast, first runs a “Spectrum Phase” of SFT that encourages diversity across potential correct answers. This phase is designed to maximize Pass@K, building a wide array of plausible solution paths. The second stage, the “Signal Phase,” employs a MaxEnt‑Guided Policy Optimization (MGPO) system to identify and amplify the most accurate paths from the diverse pool. MGPO prioritizes problems where the model is most uncertain, using entropy‑based weighting to focus learning on the most informative examples.

This separation allows a small model to explore reasoning space more effectively. Instead of relying on the sheer breadth of a gigantic parameter set, VibeThinker‑1.5B learns to generate multiple candidate solutions and then selectively reinforces the best ones. The result is a model that can reason through complex math and code problems with a level of precision that rivals, and sometimes surpasses, much larger systems.

Benchmark Performance Across Domains

VibeThinker‑1.5B’s strengths are most evident in structured reasoning benchmarks. On AIME‑25, it achieves 74.4, beating the 1.09‑trillion‑parameter Kimi K2 by over ten points. In LiveCodeBench v6, it scores 51.1, outpacing Claude Opus 4’s 47.4. While its GPQA‑Diamond score of 46.7 falls short of GPT‑4.1 and Claude, it still represents a dramatic improvement over its base model, jumping from 16.4 to 46.7.

The model also holds its own against other reasoning‑centric open‑source models such as Mistral’s Magistral Medium and Anthropic’s Claude Opus 4. Across these benchmarks, VibeThinker‑1.5B consistently outperforms non‑reasoning LLMs, regardless of size. This pattern underscores the claim that parameter count alone is not the sole determinant of reasoning capability; training design and objective alignment play a pivotal role.

Enterprise Deployment Implications

For engineering leaders and AI teams, the practical ramifications are significant. A 1.5‑billion‑parameter model that can be deployed on edge devices—mobile phones, automotive systems, or embedded hardware—offers a low‑latency, low‑cost alternative to large, cloud‑centric models. Weibo recommends inference settings of temperature = 0.6, top_p = 0.95, and max tokens = 40960, which strike a balance between creativity and determinism.

Inference costs are estimated to be 20–70× cheaper than with large models, making VibeThinker‑1.5B an attractive foundation for reasoning‑capable agents that need to run locally. The model’s entropy‑targeted RL approach also provides a roadmap for teams that wish to refine smaller checkpoints rather than investing in large‑scale pre‑training. Moreover, the transparency of its benchmark results and data decontamination steps address emerging enterprise priorities around auditability and compliance.

Weibo’s Strategic Position

Weibo’s foray into AI research reflects a broader strategy to diversify beyond its core social‑media business. By leveraging its vast user base and data streams, the company can train models that understand Chinese language nuances and cultural contexts. The release of VibeThinker‑1.5B positions Weibo as a competitor in the AI ecosystem, potentially attracting developers, researchers, and enterprises looking for high‑quality, open‑source solutions.

The company’s move also signals a response to regulatory pressures. In September 2025, Weibo was cited in official warnings for content violations, highlighting the platform’s exposure to policy risks. By investing in AI, Weibo can offer tools that help moderate content, analyze user behavior, and provide safer, more engaging experiences, thereby aligning with government expectations while maintaining growth.

Conclusion

VibeThinker‑1.5B demonstrates that a well‑engineered training pipeline can unlock reasoning performance that rivals, and in some cases surpasses, the most powerful closed‑source models, all while operating on a fraction of the compute budget. The Spectrum‑to‑Signal Principle’s dual‑phase approach—diversity first, then signal amplification—offers a compelling alternative to traditional single‑stage fine‑tuning. For enterprises, the model’s low inference cost, edge‑deployability, and transparent audit trail make it a practical choice for reasoning‑intensive applications that previously required expensive, cloud‑bound solutions.

Beyond the immediate technical achievements, Weibo’s release signals a shift in the Chinese AI landscape. The company is no longer just a social‑media platform; it is becoming a research hub that can harness its user data and in‑house talent to produce cutting‑edge AI tools. As the industry continues to grapple with the trade‑offs between scale, cost, and performance, VibeThinker‑1.5B offers a blueprint for building powerful, efficient, and open‑source language models.

Call to Action

If you’re an AI researcher, developer, or enterprise architect looking to experiment with high‑quality reasoning without the overhead of massive models, VibeThinker‑1.5B is ready for you. Download the model from Hugging Face, explore the GitHub repository, and try the recommended inference settings to see how it performs on your own workloads. For businesses, consider integrating VibeThinker into edge‑based applications—chatbots, code assistants, or educational tools—to reduce latency and cost while delivering reliable reasoning. Finally, engage with the Weibo AI community: share your findings, contribute to the open‑source project, and help shape the next generation of compact, reasoning‑optimized language models.

We value your privacy

We use cookies, including Google Analytics, to improve your experience on our site. By accepting, you agree to our use of these cookies. Learn more