Overview

  • Founded Date February 15, 1974
  • Sectors Office
  • Posted Jobs 0
  • Viewed 8

Company Description

Open-R1: a Fully Open Reproduction Of DeepSeek-R1

Hey there! This post is an intro to the task, not a claim that we’ve replicated R1 yet. We’re integrating in the open, so as quickly as we have assessment numbers, we’ll share them. You can follow our progress on Hugging Face and GitHub.

True, however it appears like there’s nothing to be evaluated as of today. I presume the ultimate goal is to train a brand-new reasoning design and after that use the very same examination metrics as o1 and the DeepSeek-R1.

Well, there should be at least some peace of mind check and validation to guarantee the design was trained properly.

Oh yes, if you are talking about the examination number of deepseek’s model it’s coming soon!

As pointed out in the post there is no design called Open-R1 to evaluate at all … not yet anyhow. This is a blog site detailing that Hugging face will take the R1 Deepseek design, work out how it was built as described in the paper and from what they launched, and then reproduce that procedure.

in truth this is practically how science works … A comes up with a strategy, discovery or development and it is tested by B, C and D to see if it is reproduceable. Thats been the foundation of research now for a few centuries.

This blog site is not stating they have actually currently done so … Its a blog outlining an intent to start training a design like R1 and calling it Open-R1.

Also DeepSeek-R1 was only released recently, and even in their paper they described the compute hours required. While those are low calculate hours for a SOTA model this does not mean you can train said design in a week. I ‘d personally love to be able to train a transformer design in a week, but we might require to wait a while for that level of calculate technology.

So there are no standards for a design that has not been built yet right? As described in the blog site, and once again in reply to your concern.

However fear not, there is a GitHub Repo already and factors (hell I may join myself), some prelim work done, and a strategy of attack. An excellent beginning position.

n
@edbeeching
has assessed the released designs already

( src: https://x.com/edwardbeeching/status/1884273209136275742)

R1 simply trained on o1 outputs, so collectively …/ s. This is what the brand-new AI czars are saying

Hi! This article is an intro to the job, not a claim that we’ve recreated R1 yet. We will absolutely share the missing piece when we have them, you can anticipate the models and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo

That’s great and essential to understand this significant hype that does not have technical comprehension and description. Science has to do with recreation, and if they declare to be open, let them fullfill the open part.

Please do release the training cost.

We will!

Excalidraw Hi n
@bojan2501
thanks, we will undoubtedly be striving to ensure this training dish can work for little language models on customer hardware considering that not everyone has a cluster of H100s in your home:-RRB- The tool we used for the images was Excalidraw! https://excalidraw.com

eagerly anticipating it! WTF are your speaking about?

need to be a joke

It’s really cool to see how the entire open source community comes together!

Ops …

5.5 M is number press reporter in the deepseekv3 tech report (just the training, not the experiment afaik), for R1 tough to approximate tbh however much less than 5.5 M imo

Historically, they have actually never ever launched code or datasets of their LLM training, so I wouldn’t anticipate this time to be various. If they would release it that would be fantastic obviously!

Yes obviously!

So essentially you’re asking to change existing censorship with another flavour of censorship?

The code for the designs are inside the model repositories, e.g. for V3: https://huggingface.co/deepseek-ai/DeepSeek-V3/blob/main/modeling_deepseek.py

Hello Team, I’m Ray Bernard, the author and developer of EQUATOR. My research study team will be dealing with a paper focused on replicating certain components of DeepSeek R1. Our aim is to replicate the cold start and offer your group with a dataset that includes COT and other strategies to support these efforts. We like to contribute our work to assist. Please let me know if you discover this helpful. Best, Ray Bernard https://www.facebook.com/groups/1186310571520299/

Where is the evaluation numbers? without it you can’t call it recreation.

8 replies

True, however it seems like there’s nothing to be assessed since today. I presume the ultimate goal is to train a brand-new reasoning model and after that utilize the exact same evaluation metrics as o1 and the DeepSeek-R1.

That’s rather intriguing, I was asking myself why the concerns the author exposed here are not being asked by others? I believe the work they have actually done is remarkable but at the same time I wonder why they wouldn’t put these missing pieces on if they are expected to be fully open.
Why even without reproduction and understanding of the innovation they could impact a lot the marketplace in this method?

4 replies

Hi! This article is an intro to the job, not a claim that we have actually replicated R1 yet. We will totally share the missing piece when we have them, you can anticipate the designs and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo

Interesting read, and it is great that we see more effort into this direction: more optimization and less brute force.
Also wonder what tool did the author use for developing step diagram.

2 replies

Excalidraw I’m so pleased that effort like this currently exist, I’m gon na try to contribute:-RRB- 1 reply

eagerly anticipating it! So racist articel

2 replies

WTF are your speaking about?

Awesome to have this open recreation started!

For Step # 1 check out https://github.com/open-thoughts/open-thoughts!

https://x.com/ryanmart3n/status/1884284101265612856

Let’s do this thing!

1 reply

It’s truly cool to see how the entire open source community comes together!

Does anyone the real training cost of r1? I can’t find it in the paper or the statement post. Is the 6M expense reported by media just the number drawn from v3’s training expense?

2 replies

Ops …

Has anybody asked the DeepSeek group to publish their training data and code, or at least share them independently with an independent duplication job like this? Have they rejected such a request?

A faithful replication depends on using the same dataset and hyperparameters. Otherwise, any major disparities with the published standards would be tough to pin down-whether due to training data differences or the replication technique itself.

1 reply

Historically, they have actually never launched code or datasets of their LLM training, so I wouldn’t anticipate this time to be various. If they would launch it that would be remarkable obviously!

In the meantime we need to make finest guess quotes and see if we can arrive ourselves.

You provide great duplication process of Deepseek reasoning training. I will attempt something similar to it.

This is actually excellent information, can we fine tune with specific usage case when code is released?

1 reply

Yes obviously!

Please think about getting rid of biased, tainted or unaligned training data and make an effort to remove copyrighted works from the crawl from intake. This will make the model more usable. If you recycled anthropic curation checks, this might also help, eliminate obviouslybiased information will likely add a lot of value. We do not want another tainted, unaligned open source model, right? And no corporate would ever utilize deepseek or a model that reuses it, right?
We appreciate your work for the benefit of humanity, we hope.
Miike C from NJ

1 reply

So generally you’re asking to replace existing censorship with another flavour of censorship?

Can’t wait! Hopefully the model will be uncensored but whatever you can do is alright! Love seeing open source structure itself up. I’m not clever enough to actually help but I can contribute support lol

Hello guys, I am even just looking for code for DeepSeek-V2, in order to fully comprehend multi-head latent attention. You do not seem to have code in Hugging Face even for that. Or am I missing something? Don’t see anything in src/transformers/models. MLA is not appropriately explained in their paper, so it would be essential to have code for this.

Scroll to Top