{"id":3246,"date":"2025-08-21T07:26:08","date_gmt":"2025-08-21T07:26:08","guid":{"rendered":"https:\/\/violethoward.com\/new\/bytedance-releases-new-open-source-seed-oss-36b-model\/"},"modified":"2025-08-21T07:26:08","modified_gmt":"2025-08-21T07:26:08","slug":"bytedance-releases-new-open-source-seed-oss-36b-model","status":"publish","type":"post","link":"https:\/\/violethoward.com\/new\/bytedance-releases-new-open-source-seed-oss-36b-model\/","title":{"rendered":"ByteDance releases new open source Seed-OSS-36B model"},"content":{"rendered":" \r\n
Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders.<\/em> Subscribe Now<\/em><\/p>\n\n\n\n TikTok is making headlines again today after the White House joined the popular social media application \u2014 but its parent company ByteDance, a Chinese web giant, also had a surprise announcement up its sleeve.<\/p>\n\n\n\n The company\u2019s Seed Team of AI researchers today released Seed-OSS-36B on AI code sharing website Hugging Face.<\/strong><\/p>\n\n\n\n Seed-OSS-36B is new line of open source, large language models (LLM) designed for advanced reasoning, and developer-focused usability with a longer token context<\/strong> \u2014 that is, how much information the models can accept as inputs and then output in a single exchange \u2014 than many competing LLMs from U.S. tech companies<\/strong>, even leaders such as OpenAI and Anthropic.<\/p>\n\n\n\n The collection introduces three main variants: <\/p>\n\n\n\n AI Scaling Hits Its Limits<\/strong><\/p>\n\n\n\n Power caps, rising token costs, and inference delays are reshaping enterprise AI. Join our exclusive salon to discover how top teams are:<\/p>\n\n\n\n Secure your spot to stay ahead<\/strong>: https:\/\/bit.ly\/4mwGngO<\/p>\n\n\n\n In releasing both synthetic and non-synthetic versions of the Seed-OSS-36B-Base model, the Seed Team sought to balance practical performance with research flexibility. <\/p>\n\n\n\n The synthetic-data variant,<\/strong> trained with additional instruction data, consistently delivers stronger scores on standard benchmarks<\/strong> and is intended as a higher-performing general-purpose option. <\/p>\n\n\n\n The non-synthetic model,<\/strong> by contrast, omits these augmentations, creating a cleaner foundation that avoids potential bias or distortion<\/strong> introduced by synthetic instruction data. <\/p>\n\n\n\n By providing both, the team gives applied users access to improved results while ensuring researchers retain a neutral baseline for studying post-training methods.<\/p>\n\n\n\n Meanwhile, the Seed-OSS-36B-Instruct model <\/strong>differs in that it is post-trained with instruction data<\/strong> to prioritize task execution and instruction following, rather than serving purely as a foundation model.<\/p>\n\n\n\n All three models are released under the Apache-2.0 license, allowing free use, modification, and redistribution by researchers and developers working for enterprises.<\/p>\n\n\n\n That means they can be used to power commercial applications, internal to a company or external\/customer-facing, without paying ByteDance any licensing fees or for application programming interface (API) usage.<\/strong><\/p>\n\n\n\n This continues the summer 2025 trend of Chinese companies shipping powerful open source models with OpenAI attempting to catch up with its own open source gpt-oss duet released earlier this month.<\/p>\n\n\n\n The Seed Team positions Seed-OSS for international applications<\/strong>, emphasizing versatility across reasoning, agent-like task execution, and multilingual settings.<\/p>\n\n\n\n The Seed Team, formed in 2023, has concentrated on building foundation models that can serve both research and applied use cases. <\/p>\n\n\n\n The architecture behind Seed-OSS-36B combines familiar design choices such as causal language modeling, grouped query attention, SwiGLU activation, RMSNorm, and RoPE positional encoding.<\/p>\n\n\n\n Each model carries 36 billion parameters across 64 layers and supports a vocabulary of 155,000 tokens.<\/p>\n\n\n\n One of the defining features is its native long-context capability, with a maximum length of 512,000 tokens,<\/strong> designed to process extended documents and reasoning chains without performance loss.<\/p>\n\n\n\n That\u2019s twice the length of OpenAI\u2019s new GPT-5 model family and is roughly equivalent to about 1,600 pages of text, <\/strong>the length of a Christian Bible. <\/p>\n\n\n\n Another distinguishing element is the introduction of a thinking budget<\/strong>, which lets developers specify how much reasoning the model should perform before delivering an answer. <\/p>\n\n\n\n It\u2019s something we\u2019ve seen from other recent open source models as well, including Nvidia\u2019s new Nemotron-Nano-9B-v2, also available on Hugging Face.<\/p>\n\n\n\n In practice, this means teams can tune performance depending on the complexity of the task and the efficiency requirements of deployment. <\/p>\n\n\n\n Budgets are recommended in multiples of 512 tokens, with 0 providing a direct response mode\/<\/p>\n\n\n\n Benchmarks published with the release position Seed-OSS-36B among the stronger large open-source models. The Instruct variant, in particular, posts state-of-the-art results in multiple areas.<\/p>\n\n\n\n The no-synthetic Base version, while slightly behind on many measures, proves competitive in its own right. <\/p>\n\n\n\n It outperforms its synthetic counterpart on GPQA-D,<\/strong> providing researchers with a cleaner, instruction-free baseline for experimentation.<\/p>\n\n\n\n For enterprises comparing open options, these results suggest Seed-OSS offers strong potential across math-heavy, coding, and long-context workloads<\/strong> while still providing flexibility for research use cases.<\/p>\n\n\n\n Beyond performance, the Seed Team highlights accessibility for developers and practitioners. The models can be deployed using Hugging Face Transformers<\/strong>, with quantization support in both 4-bit and 8-bit formats<\/strong> to reduce memory requirements. <\/p>\n\n\n\n They also integrate with vLLM for scalable serving<\/strong>, including configuration examples and API server instructions.<\/p>\n\n\n\n To lower barriers further, the team includes scripts for inference, prompt customization, and tool integration. <\/p>\n\n\n\n For technical leaders managing small teams or working under budget constraints<\/strong>, these provisions are positioned to make experimentation with 36-billion-parameter models more approachable.<\/p>\n\n\n\n With the models offered under Apache-2.0, organizations can adopt them without restrictive licensing terms, an important factor for teams balancing legal and operational concerns.<\/p>\n\n\n\n For decision makers evaluating the open-source landscape, the release brings three takeaways:<\/p>\n\n\n\n By placing strong performance and flexible deployment under an open license, ByteDance\u2019s Seed Team has added new options for enterprises, researchers, and developers alike. <\/p>\n
\n<\/div>
\n\n\n\n\n
\n<\/div>\n
Design and core features<\/h2>\n\n\n\n
Competitive performance on third-party benchmarks<\/h2>\n\n\n\n
\n
Access and deployment<\/h2>\n\n\n\n
Licensing and considerations for enterprise decision-makers<\/h2>\n\n\n\n
\n