
What's s1?
The s1 model introduced by Feifei Li's teamIt is an AI with powerful reasoning capabilitiesinference model. The model achieves a very low training cost (less than $50) compared to OpenAI-o1 andDeepSeek-R1equivalent performance of cutting-edge inference models such as the
The s1 model is based on Google'sGemini 2.0 Flash The Thinking Experimental model was conducteddistillate, and optimized by methods such as supervised fine-tuning (SFT) and test-time scaling. The s1 model shows excellent performance in math and coding ability tests, bringing new low-cost and high-efficiency solutions to the AI field.
s1 R&D background and characteristics
- R&D Background: The s1 model was introduced in response to the current problem of high technology development costs in the field of artificial intelligence. The high cost often restricts small and medium-sized enterprises and start-up teams from venturing into this field, resulting in further concentration and technological barriers in the industry. As a result, a team of researchers from Stanford University and the University of Washington worked to develop a low-cost, highly efficient AI model.
- Core featuresThe s1 model employs the technique of "distillation", which is a method of extracting the reasoning power of other, more powerful models by mimicking their answers. The successful application of this technique allows the s1 model to achieve powerful reasoning performance at a very low cost.
s1 technical details
- Training costs: The s1 model was trained at an extremely low cost, costing less than $50 in cloud computing. Training took less than 30 minutes with only 16 Nvidia H100 GPUs. This cost is much lower than the development cost of traditional AI models, demonstrating an extremely efficient use of resources.
- data set: The training dataset for the s1 model has been carefully selected to contain 1000 high-quality problems that cover a wide range of domains such as math competitions, PhD-level science problems, and Olympiads. These problems are equipped with reasoning trajectories and answers, and are validated by three criteria: difficulty, diversity and quality.
- training processThe s1 model is distilled from Google's reasoning model Gemini 2.0 Flash Thinking Experimental. During the training process, the s1 model has a self-checking mechanism that "waits" while reasoning to improve the accuracy of the model's answers. In addition, the s1 model also employs a supervised fine-tuning (SFT) approach, which utilizes a smaller dataset for self-imitation and tuning to further improve the model's performance.
s1 performance
- Math and Programming Skills: The s1 model demonstrated comparable levels of performance to the industry's top inference models, such as OpenAI's O1 and DeepSeek's R1, in tests of mathematical and programming ability. This performance demonstrates the excellence of the s1 model in its reasoning ability.
- Expansion during testing: The s1 model also has excellent performance in terms of scaling at test time. By controlling the amount of computation in the model at test time, the s1 model is able to improve the accuracy of the answers while maintaining efficiency.
s1 Impact and significance
- Technology Popularization: The successful launch of the s1 model has promoted the popularization of AI technology. Its low-cost and high-efficiency features have enabled more enterprises and research institutions to venture into the AI field, promoting the further development of the technology.
- market competition: The emergence of s1 models has intensified competition in the AI industry. Achieving powerful inference performance at a very low cost has challenged the competitive advantage of large technology companies. At the same time, the s1 model has provided lessons and references for other teams, promoting technological innovation and cooperation in the industry.
Paper Address:https://arxiv.org/abs/2501.19393
Open source address:https://github.com/simplescaling/s1
data statistics
Related Navigation

Shangtang Technology has launched a comprehensive big model system with powerful natural language processing, text-born diagrams and other multimodal capabilities, aiming to provide efficient AI solutions for enterprises.

ERNIE X1 Turbo
Baidu has launched a new generation of high-level AI assistants to disassemble complex tasks and automate the entire process with autonomous deep thinking, multimodal toolchain invocation and extreme cost advantages.

Nova Sonic
Amazon has introduced a new generation of generative AI speech models with unified model architecture, natural and smooth voice interaction, real-time two-way conversation capability and multi-language support, which can be widely used in multi-industry scenarios.

Xiaomi MiMo
Xiaomi's open-sourced 7 billion parameter inference macromodel, which outperforms models such as OpenAI o1-mini in mathematical reasoning and code competitions by a small margin.

Phi-3
A high-performance large-scale language model from Microsoft, tuned with instructions to support cross-platform operation, with excellent language comprehension and reasoning capabilities, especially suitable for multimodal application scenarios.

ZhiPu AI BM
The series of large models jointly developed by Tsinghua University and Smart Spectrum AI have powerful multimodal understanding and generation capabilities, and are widely used in natural language processing, code generation and other scenarios.

OpenAI o3-mini
OpenAI introduces small AI models with inference capabilities and cost-effective pricing, designed for developers and users to optimize application performance and efficiency.

DeepSeek-V3
Hangzhou Depth Seeker has launched an efficient open source language model with 67.1 billion parameters, using a hybrid expert architecture that excels at handling math, coding and multilingual tasks.
No comments...