Tiktok makes headlines again today The white house joined the favored social media application – but his parent company BytedanceA Chinese web giant also had a surprise announcement.
The company Sowing team from AI researchers Seed-Oss-36b published today The face hugs on the AI ​​code -Sharing.
Seed-Oss-36b is a brand new line of open source, large-scale models (LLM), which have been developed for advanced argument and developer USability with one longer token context – that’s, how much information can accept the models as inputs after which output in a single exchange – As many competing LLMs of US technology firmsEven leaders like Openaai and Anthropic.
The collection introduces three most important variants:
- Seeds-oss-36b-base With synthetic data
- Seeds-oss-36b-base without synthetic data
- Seed-oss-36b structure
When publishing each synthetic and non-synthetic versions of the Seed-Oss-36B-based model, the Seed team tried to reconcile practical performance with research flexibility.
The synthetic data variant, consistently trained with additional instruction data delivers stronger scores for traditional benchmarks And is taken into account an optional general option with higher performance.
The Non-synthetic model, Supports these augmentation, create A cleaner basis that avoids possible distortion or distortion Introduced by synthetic instruction data.
By providing the 2, the team grants applied users access to improved results and ensures that researchers retain a neutral baseline for the examination of methods after training.
Meanwhile the Seed-Oss-36B-Instructure model differs within the proven fact that it’s According to the training data in keeping with instruction data Prioritize the execution and instruction of the tasks as a substitute of only serving as a foundation model.
All three models are published under the license of Apache-2.0, which enable the free use, change and redistribution by researchers and developers who work for firms.
That means They could be used to provide industrial applications to pay the inner or external/customer-oriented employees, without Bytedance license fees or for the usage of application programming interfaces (API).
This continues the fort Summer 2025 Trend of Chinese firms that send powerful open source models With Openai, who tries to fulfill his own open source GPT-OĂź duett, which was released firstly of this month.
The seed setting positions Seed for international applicationsAnd emphasizes the flexibility when it comes to argumentation, agent -like task execution and multilingual attitudes.
The Seed team, founded in 2023, has focused on the development of foundation models that may serve each research and applied applications.
Design and core functions
The architecture behind Seed-Oss-36b combines familiar design options equivalent to causal voice modeling, grouped attention to queries, Swiglu activation, RMS standard and cable position coding.
Each model carries 36 billion parameters over 64 layers and supports a vocabulary of 155,000 tokens.
One of the defining features is his Native long context ability with a maximum length of 512,000 tokens, Developed to process prolonged documents and argument chains without lack of performance.
This is twice so long as the brand new GPT 5 model family from Openaai and is About 1,600 pages corresponds to text, The length of a Christian Bible.
Another distinction element is the introduction of A Budget thinkHow to point developers how much argument the model should perform before the output of a solution.
It is something that we also saw from other current open source models Available on the cuddling face.
In practice, which means teams can set the performance depending on the complexity of the duty and the efficiency requirements of the use.
Budgets are really useful in multiple of 512 tokens, with 0 forming a direct reply mode/
Competition performance at benchmarks from third -party providers
Benchmarks, which were published with the discharge position Seed-Oss-36b under the stronger large open source models. In particular, the instruction variant results in the most recent cutting-edge in several areas.
- Mathematics and argumentation: Seed-OĂź-36B-Instructure reached 91.7 percent on AIME24 And 65 on BeyondaimeBoth represent open source “state-of-the-art” (Sota).
- Coding: On LiveCodebech V6, the instruction model records 67.4Another Sota rating.
- Long context handling: It reaches via ruler at 128,000 context length 94.6mark the very best open source result.
- Basic model power: The basic variant of the synthetic data provides the fundamental variant 65.1 on MMLU-Pro And 81.7 in mathematicsBoth lead to their categories.
The non-synthetic basic version, which is well behind it in lots of measures, seems to be competitive.
It exceeds his synthetic counterpart to GPQA-D, Provision of researchers a cleaner, precedent -free baseline for experimentation.
For firms that compare open options, these results are Propose that Seed-Oss offers strong potential for mathematical, coded and long context-related workloads Still offers flexibility in the course of the flexibility for research usage cases.
Access and provision
In addition to performance, the Seed team emphasizes accessibility for developers and practitioners. The models could be used with hugs facial transformerswith Quantization support in each 4-bit and 8-bit formats Reduce memory requirements.
you too Integrate into Vllm for scalable portionsincluding configuration examples and API server instructions.
In order to further lower the obstacles, the team comprises scripts for inference, immediate adaptation and integration of tools.
For Technical managers who manage small teams or work under budget restrictionsThese provisions are positioned in such a way that they develop into more accessible with 36 billion parameter models.
License and considerations for decision -makers from firms
With the models offered as a part of Apache-2.0, organizations can take them without restrictive license conditions, a vital factor for teams that bring about legal and operational concerns.
For decision-makers who evaluate the open source landscape, the publication brings three snack bars:
- Modern benchmarks in mathematics, coding and long context -related argument.
- A balance between higher -perfect synthetic models and clean research base lines.
- Accessibility functions that lower the operational overhead for Lean Engineering teams.
Due to the strong performance and versatile use within the context of an open license, the seed team from bytedance has added recent options for firms, researchers and developers.

