Skip to content

  • Projects
  • Groups
  • Snippets
  • Help
    • Loading...
    • Help
    • Submit feedback
    • Contribute to GitLab
  • Sign in
A
amicimuseisiciliani
  • Project
    • Project
    • Details
    • Activity
    • Cycle Analytics
  • Issues 1
    • Issues 1
    • List
    • Board
    • Labels
    • Milestones
  • Merge Requests 0
    • Merge Requests 0
  • CI / CD
    • CI / CD
    • Pipelines
    • Jobs
    • Schedules
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Members
    • Members
  • Collapse sidebar
  • Activity
  • Create a new issue
  • Jobs
  • Issue Boards
  • Carlton Curiel
  • amicimuseisiciliani
  • Issues
  • #1

Closed
Open
Opened Mar 12, 2025 by Carlton Curiel@carltoncuriel
  • Report abuse
  • New issue
Report abuse New issue

Applied aI Tools


AI keeps getting cheaper with every passing day!

Just a couple of weeks back we had the DeepSeek V3 model pushing NVIDIA's stock into a downward spiral. Well, today we have this new expense effective design launched. At this rate of innovation, I am thinking about selling NVIDIA stocks lol.

Developed by scientists at Stanford and the University of Washington, their S1 AI design was trained for simple $50.

Yes - only $50.

This more challenges the dominance of multi-million-dollar designs like OpenAI's o1, DeepSeek's R1, and others.

This breakthrough highlights how innovation in AI no longer requires massive budgets, potentially equalizing access to innovative reasoning abilities.

Below, we explore s1's advancement, benefits, and ramifications for the AI engineering market.

Here's the original paper for your reference - s1: Simple test-time scaling

How s1 was developed: Breaking down the methodology

It is really intriguing to find out how researchers across the world are optimizing with limited resources to lower costs. And these efforts are working too.

I have tried to keep it simple and jargon-free to make it easy to comprehend, read on!

Knowledge distillation: The secret sauce

The s1 model uses a method called knowledge distillation.

Here, a smaller sized AI model mimics the reasoning processes of a larger, more sophisticated one.

Researchers trained s1 utilizing outputs from Google's Gemini 2.0 Flash Thinking Experimental, a reasoning-focused model available via Google AI Studio. The group prevented resource-heavy methods like support knowing. They utilized supervised fine-tuning (SFT) on a dataset of simply 1,000 curated questions. These concerns were paired with Gemini's responses and detailed reasoning.

What is monitored fine-tuning (SFT)?

Supervised Fine-Tuning (SFT) is an artificial intelligence technique. It is utilized to adapt a pre-trained Large Language Model (LLM) to a particular job. For this process, it utilizes labeled information, where each information point is identified with the correct output.

Adopting uniqueness in training has several advantages:

- SFT can enhance a design's efficiency on particular jobs
- Improves information performance
- Saves resources compared to training from scratch
- Allows for customization
- Improve a model's ability to manage edge cases and control its behavior.
This approach permitted s1 to replicate Gemini's analytical methods at a fraction of the cost. For comparison, DeepSeek's R1 design, created to rival OpenAI's o1, reportedly required pricey support discovering pipelines.

Cost and compute effectiveness

Training s1 took under thirty minutes using 16 NVIDIA H100 GPUs. This expense researchers roughly $20-$ 50 in cloud calculate credits!

By contrast, OpenAI's o1 and comparable models require thousands of dollars in compute resources. The base design for s1 was an off-the-shelf AI from Alibaba's Qwen, easily available on GitHub.

Here are some major aspects to think about that aided with attaining this expense efficiency:

Low-cost training: The s1 model attained remarkable results with less than $50 in cloud computing credits! Niklas Muennighoff is a Stanford scientist included in the job. He approximated that the required calculate power could be quickly rented for around $20. This showcases the task's extraordinary price and availability.
Minimal Resources: The team utilized an off-the-shelf base design. They fine-tuned it through distillation. They extracted reasoning capabilities from Google's Gemini 2.0 Flash Thinking Experimental.
Small Dataset: The s1 design was trained using a small dataset of just 1,000 curated questions and answers. It included the thinking behind each response from Google's Gemini 2.0.
Quick Training Time: The model was trained in less than thirty minutes using 16 Nvidia H100 GPUs.
Ablation Experiments: The low cost allowed scientists to run many ablation experiments. They made little variations in configuration to discover what works best. For example, they determined whether the model must use 'Wait' and not 'Hmm'.
Availability: The development of s1 offers an alternative to high-cost AI models like OpenAI's o1. This improvement brings the potential for powerful thinking models to a more comprehensive audience. The code, information, and training are available on GitHub.
These aspects challenge the notion that enormous investment is constantly necessary for producing capable AI designs. They democratize AI development, enabling smaller sized teams with minimal resources to attain significant results.

The 'Wait' Trick

A creative development in s1's design involves including the word "wait" throughout its reasoning process.

This simple timely extension forces the design to stop briefly and double-check its responses, enhancing precision without extra training.

The 'Wait' Trick is an example of how cautious prompt engineering can substantially improve AI model performance. This enhancement does not rely exclusively on increasing model size or training information.

Discover more about writing prompt - Why Structuring or Formatting Is Crucial In Prompt Engineering?

Advantages of s1 over industry leading AI designs

Let's comprehend why this development is essential for the AI engineering market:

1. Cost availability

OpenAI, Google, annunciogratis.net and Meta invest billions in AI infrastructure. However, s1 proves that high-performance reasoning models can be developed with minimal resources.

For instance:

OpenAI's o1: Developed using proprietary techniques and costly calculate.
DeepSeek's R1: Relied on massive reinforcement learning.
s1: Attained comparable outcomes for under $50 utilizing distillation and SFT.
2. Open-source openness

s1's code, training information, and design weights are openly available on GitHub, unlike closed-source designs like o1 or Claude. This openness cultivates community cooperation and scope of audits.

3. Performance on benchmarks

In tests measuring mathematical problem-solving and coding tasks, s1 matched the efficiency of leading models like o1. It likewise neared the performance of R1. For instance:

- The s1 model outshined OpenAI's o1-preview by up to 27% on competition math concerns from MATH and AIME24 datasets
- GSM8K (mathematics reasoning): s1 scored within 5% of o1.
- HumanEval (coding): s1 attained ~ 70% accuracy, comparable to R1.
- An essential function of S1 is its use of test-time scaling, which enhances its precision beyond preliminary abilities. For instance, it increased from 50% to 57% on AIME24 issues utilizing this technique.
s1 doesn't go beyond GPT-4 or Claude-v1 in raw ability. These designs excel in specialized domains like scientific oncology.

While distillation approaches can reproduce existing designs, some experts note they may not lead to breakthrough advancements in AI efficiency

Still, its cost-to-performance ratio is unrivaled!

s1 is challenging the status quo

What does the development of s1 mean for the world?

Commoditization of AI Models

s1's success raises existential concerns for AI giants.

If a little group can reproduce innovative thinking for $50, what distinguishes a $100 million model? This threatens the "moat" of proprietary AI systems, pushing companies to innovate beyond distillation.

Legal and ethical issues

OpenAI has earlier implicated competitors like DeepSeek of poorly harvesting data through API calls. But, s1 avoids this issue by using Google's Gemini 2.0 within its regards to service, which permits non-commercial research study.

Shifting power characteristics

s1 exemplifies the "democratization of AI", allowing startups and scientists to take on tech giants. Projects like Meta's LLaMA (which needs pricey fine-tuning) now deal with pressure from cheaper, purpose-built options.

The constraints of s1 model and future directions in AI engineering

Not all is best with s1 for now, and it is wrong to expect so with minimal resources. Here's the s1 model constraints you should know before adopting:

Scope of Reasoning

s1 masters tasks with clear detailed reasoning (e.g., math problems) but has a hard time with open-ended creativity or nuanced context. This mirrors constraints seen in models like LLaMA and PaLM 2.

Dependency on moms and dad designs

As a distilled design, s1's abilities are naturally bounded by Gemini 2.0's knowledge. It can not exceed the initial model's thinking, unlike OpenAI's o1, which was trained from scratch.

Scalability concerns

While s1 shows "test-time scaling" (extending its thinking actions), real innovation-like GPT-4's leap over GPT-3.5-still requires huge compute spending plans.

What next from here?

The s1 experiment highlights two essential patterns:

Distillation is democratizing AI: Small teams can now replicate high-end capabilities!
The value shift: Future competition may fixate data quality and special architectures, not simply calculate scale.
Meta, Google, and Microsoft are investing over $100 billion in AI infrastructure. Open-source tasks like s1 might a rebalancing. This change would enable innovation to thrive at both the grassroots and corporate levels.

s1 isn't a replacement for industry-leading models, however it's a wake-up call.

By slashing expenses and opening gain access to, it challenges the AI community to focus on performance and inclusivity.

Whether this causes a wave of low-cost rivals or tighter constraints from tech giants remains to be seen. One thing is clear: the period of "bigger is much better" in AI is being redefined.

Have you attempted the s1 model?

The world is moving quickly with AI engineering advancements - and this is now a matter of days, not months.

I will keep covering the current AI designs for you all to attempt. One should find out the optimizations made to lower costs or innovate. This is really an interesting space which I am delighting in to blog about.

If there is any issue, correction, or doubt, please comment. I would enjoy to repair it or clear any doubt you have.

At Applied AI Tools, we wish to make finding out available. You can find how to utilize the many available AI software for your personal and professional usage. If you have any questions - email to content@merrative.com and we will cover them in our guides and blog sites.

Learn more about AI ideas:

- 2 key insights on the future of software application development - Transforming Software Design with AI Agents
- Explore AI Agents - What is OpenAI o3-mini
- Learn what is tree of thoughts prompting approach
- Make the mos of Google Gemini - 6 most current Generative AI tools by Google to improve workplace productivity
- Learn what influencers and specialists think about AI's effect on future of work - 15+ Generative AI prices estimate on future of work, impact on jobs and workforce productivity
You can register for our newsletter to get notified when we publish brand-new guides!

Type your email ...

Subscribe

This post is composed utilizing resources of Merrative. We are a publishing talent marketplace that helps you develop publications and content libraries.

Get in touch if you wish to develop a material library like ours. We concentrate on the specific niche of Applied AI, Technology, Artificial Intelligence, or Data Science.

Assignee
Assign to
None
Milestone
None
Assign milestone
Time tracking
None
Due date
No due date
0
Labels
None
Assign labels
  • View project labels
Reference: carltoncuriel/amicimuseisiciliani#1