Skip to content

  • Projects
  • Groups
  • Snippets
  • Help
    • Loading...
    • Help
    • Submit feedback
    • Contribute to GitLab
  • Sign in
A
angkor-stroy
  • Project
    • Project
    • Details
    • Activity
    • Cycle Analytics
  • Issues 23
    • Issues 23
    • List
    • Board
    • Labels
    • Milestones
  • Merge Requests 0
    • Merge Requests 0
  • CI / CD
    • CI / CD
    • Pipelines
    • Jobs
    • Schedules
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Members
    • Members
  • Collapse sidebar
  • Activity
  • Create a new issue
  • Jobs
  • Issue Boards
  • Alejandrina Leblanc
  • angkor-stroy
  • Issues
  • #10

Closed
Open
Opened Feb 12, 2025 by Alejandrina Leblanc@alejandrinaleb
  • Report abuse
  • New issue
Report abuse New issue

Hugging Face Clones OpenAI's Deep Research in 24 Hr


Open source "Deep Research" project shows that agent structures increase AI model ability.

On Tuesday, Hugging Face researchers released an open source AI research agent called "Open Deep Research," produced by an in-house group as a difficulty 24 hr after the launch of OpenAI's Deep Research feature, which can autonomously browse the web and develop research study reports. The task looks for to match Deep Research's performance while making the innovation freely available to designers.

"While effective LLMs are now freely available in open-source, OpenAI didn't disclose much about the agentic structure underlying Deep Research," writes Hugging Face on its announcement page. "So we chose to embark on a 24-hour mission to reproduce their results and open-source the needed structure along the method!"

Similar to both OpenAI's Deep Research and Google's implementation of its own "Deep Research" using Gemini (first presented in December-before OpenAI), Hugging Face's solution includes an "representative" framework to an existing AI model to allow it to carry out multi-step tasks, such as gathering details and building the report as it goes along that it presents to the user at the end.

The open source clone is already racking up similar benchmark outcomes. After only a day's work, Hugging Face's Open Deep Research has reached 55.15 percent accuracy on the General AI Assistants (GAIA) standard, which tests an AI model's capability to collect and synthesize details from multiple sources. OpenAI's Deep Research scored 67.36 percent accuracy on the exact same standard with a single-pass response (OpenAI's rating increased to 72.57 percent when 64 reactions were integrated utilizing a consensus mechanism).

As Hugging Face explains in its post, GAIA consists of intricate multi-step questions such as this one:

Which of the fruits shown in the 2008 painting "Embroidery from Uzbekistan" were served as part of the October 1949 breakfast menu for the ocean liner that was later on utilized as a floating prop for the movie "The Last Voyage"? Give the products as a comma-separated list, purchasing them in clockwise order based upon their plan in the painting beginning from the 12 o'clock position. Use the plural kind of each fruit.

To correctly address that type of concern, the AI agent need to look for multiple diverse sources and assemble them into a meaningful response. Much of the concerns in GAIA represent no easy task, even for clashofcryptos.trade a human, so they test agentic AI's guts quite well.

Choosing the ideal core AI model

An AI representative is absolutely nothing without some type of existing AI model at its core. In the meantime, Open Deep Research develops on OpenAI's big language designs (such as GPT-4o) or simulated reasoning designs (such as o1 and o3-mini) through an API. But it can also be adapted to open-weights AI models. The novel part here is the agentic structure that holds all of it together and permits an AI language model to autonomously complete a research job.

We spoke with Hugging Face's Aymeric Roucher, who leads the Open Deep Research job, about the team's choice of AI design. "It's not 'open weights' because we used a closed weights model even if it worked well, however we explain all the advancement procedure and show the code," he informed Ars Technica. "It can be switched to any other design, so [it] supports a completely open pipeline."

"I tried a lot of LLMs consisting of [Deepseek] R1 and o3-mini," Roucher adds. "And for this use case o1 worked best. But with the open-R1 initiative that we've launched, we may supplant o1 with a much better open design."

While the core LLM or SR model at the heart of the research representative is important, Open Deep Research shows that developing the ideal agentic layer is crucial, because benchmarks reveal that the multi-step agentic method improves big language model capability significantly: OpenAI's GPT-4o alone (without an agentic framework) ratings 29 percent on average on the GAIA criteria versus OpenAI Deep Research's 67 percent.

According to Roucher, a core component of Hugging Face's recreation makes the task work along with it does. They used Hugging Face's open source "smolagents" library to get a running start, which utilizes what they call "code agents" instead of JSON-based representatives. These code representatives write their actions in programming code, which reportedly makes them 30 percent more efficient at completing tasks. The approach permits the system to handle intricate sequences of actions more concisely.

The speed of open source AI

Like other open source AI applications, the designers behind Open Deep Research have wasted no time repeating the design, thanks partially to outdoors contributors. And like other open source jobs, the team constructed off of the work of others, which reduces development times. For example, Hugging Face used web surfing and text assessment tools obtained from Microsoft Research's Magnetic-One representative task from late 2024.

While the open source research representative does not yet match OpenAI's performance, its release gives developers open door wavedream.wiki to study and customize the . The task shows the research community's ability to rapidly replicate and openly share AI capabilities that were previously available just through commercial suppliers.

"I believe [the criteria are] quite indicative for hard concerns," said Roucher. "But in regards to speed and UX, our service is far from being as enhanced as theirs."

Roucher says future improvements to its research study agent may include assistance for townshipmarket.co.za more file formats and vision-based web browsing capabilities. And Hugging Face is currently dealing with cloning OpenAI's Operator, which can carry out other types of jobs (such as seeing computer screens and controlling mouse and wiki.eqoarevival.com keyboard inputs) within a web browser environment.

Hugging Face has actually posted its code publicly on GitHub and opened positions for engineers to help expand the task's abilities.

"The action has been fantastic," Roucher told Ars. "We've got lots of new contributors chiming in and proposing additions.

Assignee
Assign to
None
Milestone
None
Assign milestone
Time tracking
None
Due date
No due date
0
Labels
None
Assign labels
  • View project labels
Reference: alejandrinaleb/angkor-stroy#10