
Cyberbizafrica
Overview
-
Founded Date julio 2, 1971
-
Sectors Tecnología
-
Posted Jobs 0
-
Viewed 21
Company Description
Open-R1: a Fully Open Reproduction Of DeepSeek-R1
Hey there! This article is an introduction to the job, not a claim that we have actually reproduced R1 yet. We’re constructing in the open, so as quickly as we have evaluation numbers, we’ll share them. You can follow our progress on Hugging Face and GitHub.
True, but it looks like there’s nothing to be assessed since right now. I assume the supreme goal is to train a new reasoning design and after that use the same assessment metrics as o1 and the DeepSeek-R1.
Well, there ought to be at least some sanity check and validation to make sure the design was trained properly.
Oh yes, if you are talking about the evaluation variety of deepseek’s model it’s coming soon!
As discussed in the post there is no model called Open-R1 to check at all … not yet anyhow. This is a blog describing that Hugging face will take the R1 Deepseek model, exercise how it was developed as detailed in the paper and from what they released, and after that reproduce that procedure.
in fact this is practically how science works … A develops a strategy, discovery or development and it is evaluated by B, C and D to see if it is reproduceable. Thats been the cornerstone of research study now for a couple of centuries.
This blog site is not stating they have actually already done so … Its a blog site detailing an intent to start training a design like R1 and calling it Open-R1.
Also DeepSeek-R1 was just released recently, and even in their paper they described the calculate hours required. While those are low compute hours for a SOTA design this does not imply you can train stated model in a week. I ‘d personally like to be able to train a transformer design in a week, however we might require to wait a while for that level of calculate technology.
So there are no criteria for a model that has not been developed yet right? As described in the blog, and once again in reply to your question.
However fear not, there is a GitHub Repo already and factors (hell I might join myself), some prelim work done, and a master plan. An excellent beginning position.
n
@edbeeching
has actually evaluated the launched models already
( src: https://x.com/edwardbeeching/status/1884273209136275742)
R1 simply trained on o1 outputs, so collectively …/ s. This is what the new AI czars are stating
Hi! This article is an intro to the project, not a claim that we have actually replicated R1 yet. We will totally share the missing out on piece when we have them, you can anticipate the models and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo
That’s good and essential to understand this remarkable buzz that lacks technical understanding and description. Science has to do with reproduction, and if they declare to be open, let them fullfill the open part.
Please do release the training expense.
We will!
Excalidraw Hi n
@bojan2501
thanks, we will indeed be working hard to make certain this training dish can work for little language models on consumer hardware considering that not everybody has a cluster of H100s in the house:-RRB- The tool we used for the images was Excalidraw! https://excalidraw.com
looking forward to it! WTF are your discussing?
should be a joke
It’s actually cool to see how the entire open source community comes together!
Ops …
5.5 M is number reporter in the deepseekv3 tech report (just the training, not the experiment afaik), for R1 difficult to estimate tbh but much less than 5.5 M imo
Historically, they have never launched code or datasets of their LLM training, so I would not expect this time to be various. If they would launch it that would be incredible naturally!
Yes of course!
So basically you’re asking to replace existing censorship with another flavour of censorship?
The code for the models are inside the design repositories, e.g. for V3: https://huggingface.co/deepseek-ai/DeepSeek-V3/blob/main/modeling_deepseek.py
Hello Team, I’m Ray Bernard, the author and developer of EQUATOR. My research study team will be dealing with a paper concentrated on reproducing particular elements of DeepSeek R1. Our goal is to recreate the cold start and offer your group with a dataset that includes COT and other strategies to support these efforts. We like to contribute our work to help. Please let me know if you discover this beneficial. Best, Ray Bernard https://www.facebook.com/groups/1186310571520299/
Where is the examination numbers? without it you can’t call it recreation.
8 replies
True, but it seems like there’s absolutely nothing to be examined as of right now. I assume the supreme goal is to train a brand-new reasoning design and after that utilize the very same assessment metrics as o1 and the DeepSeek-R1.
That’s quite intriguing, I was asking myself why the questions the author exposed here are not being asked by others? I think the work they have done is memorable but at the exact same time I wonder why they wouldn’t put these missing pieces on if they are supposed to be completely open.
Why even without reproduction and comprehension of the they could impact a lot the market in this method?
4 replies
Hi! This post is an introduction to the project, not a claim that we’ve reproduced R1 yet. We will completely share the missing piece when we have them, you can anticipate the models and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo
Interesting read, and it is great that we see more effort into this instructions: more optimization and less strength.
Also wonder what tool did the author use for developing action diagram.
2 replies
Excalidraw I’m so thankful that effort like this currently exist, I’m gon na try to contribute:-RRB- 1 reply
eagerly anticipating it! So racist articel
2 replies
WTF are your talking about?
Awesome to have this open recreation began!
For Step # 1 check out https://github.com/open-thoughts/open-thoughts!
https://x.com/ryanmart3n/status/1884284101265612856
Let’s do this thing!
1 reply
It’s actually cool to see how the entire open source community comes together!
Does anyone know the actual training expense of r1? I can’t find it in the paper or the announcement post. Is the 6M cost reported by media simply the number taken from v3’s training expense?
2 replies
Ops …
Has anybody asked the DeepSeek team to publish their training data and code, or at least share them independently with an independent duplication project like this? Have they rejected such a request?
A faithful replication depends upon using the very same dataset and hyperparameters. Otherwise, any major discrepancies with the published benchmarks would be difficult to pin down-whether due to training data distinctions or the replication technique itself.
1 reply
Historically, they have never released code or datasets of their LLM training, so I would not expect this time to be different. If they would launch it that would be amazing obviously!
In the meantime we need to make finest guess price quotes and see if we can get there ourselves.
You provide excellent replication process of Deepseek reasoning training. I will attempt something similar to it.
This is really great information, can we fine tune with specific usage case when code is released?
1 reply
Yes of course!
Please think about getting rid of prejudiced, tainted or unaligned training information and make an effort to eliminate copyrighted works from the crawl from consumption. This will make the design more usable. If you reused anthropic curation checks, this might also assist, remove obviouslybiased data will likely add a lot of worth. We don’t desire another polluted, unaligned open source design, right? And no corporate would ever utilize deepseek or a design that recycles it, right?
We value your work for the benefit of mankind, we hope.
Miike C from NJ
1 reply
So essentially you’re asking to replace existing censorship with another flavour of censorship?
Can’t wait! Hopefully the model will be uncensored however whatever you can do is alright! Love seeing open source structure itself up. I’m not smart enough to in fact assist however I can contribute support lol
Hello guys, I am even simply looking for code for DeepSeek-V2, in order to fully understand multi-head hidden attention. You do not seem to have code in Hugging Face even for that. Or am I missing something? Don’t see anything in src/transformers/models. MLA is not correctly described in their paper, so it would be necessary to have code for this.