Skip to content

  • Projects
  • Groups
  • Snippets
  • Help
    • Loading...
    • Help
    • Submit feedback
    • Contribute to GitLab
  • Sign in
W
web-3buzz
  • Project
    • Project
    • Details
    • Activity
    • Cycle Analytics
  • Issues 36
    • Issues 36
    • List
    • Board
    • Labels
    • Milestones
  • Merge Requests 0
    • Merge Requests 0
  • CI / CD
    • CI / CD
    • Pipelines
    • Jobs
    • Schedules
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Members
    • Members
  • Collapse sidebar
  • Activity
  • Create a new issue
  • Jobs
  • Issue Boards
  • Alison Randell
  • web-3buzz
  • Issues
  • #2

Closed
Open
Opened Feb 09, 2025 by Alison Randell@alisons8741073
  • Report abuse
  • New issue
Report abuse New issue

If there's Intelligent Life out There


Optimizing LLMs to be proficient at specific tests backfires on Meta, Stability.

-. -. -. -. -. -. -

When you buy through links on our website, we might make an affiliate commission. Here's how it works.

Hugging Face has actually launched its second LLM leaderboard to rank the very best language models it has actually tested. The brand-new leaderboard seeks to be a more tough consistent requirement for evaluating open large language design (LLM) efficiency across a variety of tasks. Alibaba's Qwen models appear dominant in the leaderboard's inaugural rankings, taking three spots in the top 10.

Pumped to reveal the brand name new open LLM leaderboard. We burned 300 H100 to re-run brand-new evaluations like MMLU-pro for all significant open LLMs!Some knowing:- Qwen 72B is the king and Chinese open designs are controling general- Previous evaluations have actually ended up being too simple for recent ... June 26, 2024

Hugging Face's 2nd leaderboard tests language designs across 4 jobs: understanding testing, reasoning on exceptionally long contexts, intricate math abilities, and direction following. Six benchmarks are used to check these qualities, with tests including resolving 1,000-word murder mysteries, explaining PhD-level questions in layman's terms, and the majority of overwhelming of all: high-school mathematics formulas. A full breakdown of the criteria used can be found on Hugging Face's blog site.

The frontrunner of the brand-new leaderboard is Qwen, Alibaba's LLM, which takes first, 3rd, and 10th location with its handful of variations. Also showing up are Llama3-70B, Meta's LLM, and a handful of smaller projects that managed to surpass the pack. Notably missing is any indication of ChatGPT; Hugging Face's leaderboard does not check closed-source models to make sure reproducibility of results.

Tests to qualify on the leaderboard are run exclusively on Hugging Face's own computers, which according to CEO Clem Delangue's Twitter, are powered by 300 Nvidia H100 GPUs. Because of Hugging Face's open-source and collaborative nature, anybody is complimentary to send new designs for screening and admission on the leaderboard, with a brand-new voting system focusing on popular brand-new entries for testing. The leaderboard can be filtered to show only a highlighted range of considerable designs to prevent a complicated excess of small LLMs.

As a pillar of the LLM area, Hugging Face has actually ended up being a relied on source for LLM learning and neighborhood collaboration. After its very first leaderboard was launched last year as a method to compare and replicate testing arise from numerous established LLMs, the board rapidly took off in appeal. Getting high ranks on the board became the goal of lots of designers, little and big, and as models have ended up being usually stronger, 'smarter,' and optimized for the particular tests of the first leaderboard, its results have become less and less significant, hence the development of a 2nd version.

Some LLMs, consisting of newer variations of Meta's Llama, badly underperformed in the brand-new leaderboard compared to their high marks in the very first. This originated from a trend of over-training LLMs just on the very first leaderboard's criteria, leading to falling back in real-world efficiency. This regression of efficiency, thanks to hyperspecific and self-referential information, follows a pattern of AI performance growing even worse over time, showing as soon as again as Google's AI answers have shown that LLM efficiency is just as good as its training information and that real synthetic "intelligence" is still numerous, numerous years away.

Remain on the Leading Edge: Get the Tom's Hardware Newsletter

Get Tom's Hardware's best news and thorough evaluations, straight to your inbox.

Dallin Grimm is a contributing author for Tom's Hardware. He has been constructing and breaking computers considering that 2017, working as the resident child at Tom's. From APUs to RGB, Dallin guides all the most recent tech news.

Moore Threads GPUs apparently show 'excellent' inference efficiency with DeepSeek designs

DeepSeek research study recommends Huawei's Ascend 910C delivers 60% of Nvidia H100 inference efficiency

Asus and MSI hike RTX 5090 and RTX 5080 GPU rates by as much as 18%

-. bit_user. LLM performance is just as good as its training data which real artificial "intelligence" is still numerous, lots of years away. First, this statement discount rates the role of network architecture.

The meaning of "intelligence" can not be whether something processes details precisely like humans do, archmageriseswiki.com or else the look for extra terrestrial intelligence would be entirely useless. If there's intelligent life out there, setiathome.berkeley.edu it most likely does not believe rather like we do. Machines that act and behave intelligently also needn't always do so, either. Reply

-. jp7189. I do not love the click-bait China vs. the world title. The reality is qwen is open source, open weights and can be run anywhere. It can (and has currently been) tweaked to add/remove bias. I praise hugging face's work to develop standardized tests for LLMs, and for putting the focus on open source, open weights first. Reply

-. jp7189. bit_user said:. First, this declaration discount rates the function of network architecture.

Second, intelligence isn't a binary thing - it's more like a spectrum. There are various classes cognitive tasks and capabilities you may be acquainted with, if you study child advancement or animal intelligence.

The definition of "intelligence" can not be whether something processes details precisely like humans do, or else the search for extra terrestrial intelligence would be entirely futile. If there's intelligent life out there, it most likely doesn't believe quite like we do. Machines that act and behave smartly also needn't always do so, either. We're producing a tools to help humans, therfore I would argue LLMs are more helpful if we grade them by human intelligence requirements. Reply

- View All 3 Comments

Most Popular

Tomshardware becomes part of Future US Inc, an international media group and leading digital publisher. Visit our business website.

- Conditions.

  • Contact Future's professionals.
  • Privacy policy. - Cookies policy. - Availability Statement. - Advertise with us.
  • About us. - Coupons.
  • Careers

    © Future US, Inc. Full 7th Floor, 130 West 42nd Street, New York City, NY 10036.
Assignee
Assign to
None
Milestone
None
Assign milestone
Time tracking
None
Due date
No due date
0
Labels
None
Assign labels
  • View project labels
Reference: alisons8741073/web-3buzz#2