Skip to content

  • Projects
  • Groups
  • Snippets
  • Help
    • Loading...
    • Help
    • Submit feedback
    • Contribute to GitLab
  • Sign in
W
w-3ttich
  • Project
    • Project
    • Details
    • Activity
    • Cycle Analytics
  • Issues 1
    • Issues 1
    • List
    • Board
    • Labels
    • Milestones
  • Merge Requests 0
    • Merge Requests 0
  • CI / CD
    • CI / CD
    • Pipelines
    • Jobs
    • Schedules
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Members
    • Members
  • Collapse sidebar
  • Activity
  • Create a new issue
  • Jobs
  • Issue Boards
  • Sheena Dodery
  • w-3ttich
  • Issues
  • #1

Closed
Open
Opened Feb 09, 2025 by Sheena Dodery@sheenae7121134
  • Report abuse
  • New issue
Report abuse New issue

DeepSeek: the Chinese aI Model That's a Tech Breakthrough and A Security Risk


DeepSeek: at this phase, the only takeaway is that open-source models surpass proprietary ones. Everything else is problematic and I don't buy the general public numbers.

DeepSink was developed on top of open source Meta models (PyTorch, Llama) and ClosedAI is now in danger because its appraisal is outrageous.

To my understanding, no public paperwork links DeepSeek straight to a particular "Test Time Scaling" strategy, but that's extremely probable, raovatonline.org so permit me to simplify.

Test Time Scaling is utilized in maker discovering to scale the design's efficiency at test time instead of during training.

That means less GPU hours and less effective chips.

To put it simply, annunciogratis.net lower computational requirements and lower hardware expenses.

That's why Nvidia lost practically $600 billion in market cap, the greatest one-day loss in U.S. history!

Lots of people and macphersonwiki.mywikis.wiki institutions who shorted American AI stocks ended up being exceptionally abundant in a couple of hours since financiers now forecast we will require less powerful AI chips ...

Nvidia short-sellers just made a single-day profit of $6.56 billion according to research from S3 Partners. Nothing compared to the marketplace cap, I'm taking a look at the single-day amount. More than 6 billions in less than 12 hours is a lot in my book. And that's simply for Nvidia. Short sellers of chipmaker Broadcom made more than $2 billion in earnings in a few hours (the US stock market runs from 9:30 AM to 4:00 PM EST).

The Nvidia Short Interest Gradually information programs we had the 2nd highest level in January 2025 at $39B however this is dated because the last record date was Jan 15, 2025 -we have to wait for the current data!

A tweet I saw 13 hours after publishing my short article! Perfect summary Distilled language models

Small language models are trained on a smaller sized scale. What makes them various isn't simply the capabilities, it is how they have actually been constructed. A distilled language model is a smaller sized, more effective model developed by moving the knowledge from a bigger, more intricate design like the future ChatGPT 5.

Imagine we have a teacher design (GPT5), which is a large language design: a deep neural network trained on a lot of information. Highly resource-intensive when there's limited computational power or when you require speed.

The knowledge from this instructor design is then "distilled" into a trainee design. The trainee design is easier and has less parameters/layers, that makes it lighter: less memory usage and computational needs.

During distillation, the trainee design is trained not just on the raw information however likewise on the outputs or the "soft targets" (probabilities for each class instead of tough labels) produced by the instructor model.

With distillation, the trainee model gains from both the original data and the detailed forecasts (the "soft targets") made by the instructor design.

In other words, the design doesn't simply gain from "soft targets" however likewise from the exact same training information used for the teacher, but with the guidance of the instructor's outputs. That's how knowledge transfer is enhanced: dual learning from information and from the instructor's predictions!

Ultimately, the trainee imitates the instructor's decision-making procedure ... all while using much less computational power!

But here's the twist as I understand it: DeepSeek didn't simply extract content from a single large language model like ChatGPT 4. It depended on lots of large language designs, including open-source ones like Meta's Llama.

So now we are distilling not one LLM however numerous LLMs. That was one of the "genius" idea: mixing various architectures and datasets to create a seriously versatile and systemcheck-wiki.de robust little language model!

DeepSeek: Less guidance

Another necessary innovation: less human supervision/guidance.

The concern is: how far can models choose less human-labeled data?

R1-Zero learned "thinking" abilities through experimentation, it evolves, it has special "thinking behaviors" which can result in sound, endless repetition, and language mixing.

R1-Zero was speculative: there was no preliminary guidance from identified data.

DeepSeek-R1 is various: it utilized a structured training pipeline that includes both supervised fine-tuning and support knowing (RL). It started with preliminary fine-tuning, followed by RL to fine-tune and improve its reasoning capabilities.

Completion result? Less sound and no language blending, unlike R1-Zero.

R1 utilizes human-like reasoning patterns initially and it then advances through RL. The innovation here is less human-labeled data + RL to both guide and refine the design's efficiency.

My concern is: did DeepSeek really fix the problem knowing they drew out a lot of information from the datasets of LLMs, which all gained from human guidance? To put it simply, wavedream.wiki is the conventional dependence really broken when they relied on previously trained designs?

Let me reveal you a live real-world screenshot shared by Alexandre Blanc today. It shows training information drawn out from other models (here, ChatGPT) that have gained from human supervision ... I am not convinced yet that the standard dependence is broken. It is "simple" to not require massive amounts of top quality reasoning data for training when taking shortcuts ...

To be balanced and reveal the research study, I've uploaded the DeepSeek R1 Paper (downloadable PDF, 22 pages).

My concerns relating to DeepSink?

Both the web and mobile apps collect your IP, keystroke patterns, and gadget details, and everything is kept on servers in China.

Keystroke pattern analysis is a behavioral biometric technique utilized to recognize and validate individuals based on their unique typing patterns.

I can hear the "But 0p3n s0urc3 ...!" remarks.

Yes, demo.qkseo.in open source is terrific, but this reasoning is restricted due to the fact that it does NOT think about human psychology.

Regular users will never run designs in your area.

Most will merely want quick responses.

Technically unsophisticated users will use the web and mobile variations.

Millions have already downloaded the mobile app on their phone.

DeekSeek's models have a genuine edge which's why we see ultra-fast user adoption. In the meantime, they are exceptional to Google's Gemini or OpenAI's ChatGPT in numerous methods. R1 scores high up on objective standards, no doubt about that.

I recommend browsing for anything delicate that does not line up with the Party's propaganda on the web or mobile app, and the output will speak for itself ...

China vs America

Screenshots by T. Cassel. Freedom of speech is gorgeous. I could share horrible examples of propaganda and censorship but I won't. Just do your own research. I'll end with DeepSeek's privacy policy, which you can continue reading their site. This is a basic screenshot, wiki.dulovic.tech nothing more.

Feel confident, your code, ideas and conversations will never ever be archived! When it comes to the real financial investments behind DeepSeek, we have no idea if they remain in the hundreds of millions or in the billions. We just understand the $5.6 M quantity the media has been pressing left and right is misinformation!

Assignee
Assign to
None
Milestone
None
Assign milestone
Time tracking
None
Due date
No due date
0
Labels
None
Assign labels
  • View project labels
Reference: sheenae7121134/w-3ttich#1