Overview

  • Founded Date March 3, 2002
  • Sectors Automotive
  • Posted Jobs 0
  • Viewed 24

Company Description

Open-R1: a Fully Open Reproduction Of DeepSeek-R1

Hey there! This post is an intro to the job, not a claim that we’ve reproduced R1 yet. We’re building in the open, so as quickly as we have evaluation numbers, we’ll share them. You can follow our progress on Hugging Face and GitHub.

True, however it appears like there’s absolutely nothing to be examined since today. I assume the supreme objective is to train a brand-new reasoning design and then utilize the very same evaluation metrics as o1 and the DeepSeek-R1.

Well, there should be at least some peace of mind check and validation to guarantee the model was trained properly.

Oh yes, if you are speaking about the evaluation number of deepseek’s design it’s coming soon!

As mentioned in the blog site post there is no design called Open-R1 to evaluate at all … not yet anyway. This is a blog laying out that Hugging face will take the R1 Deepseek model, exercise how it was constructed as laid out in the paper and from what they launched, and after that duplicate that process.

in reality this is practically how science works … A creates a plan, discovery or development and it is tested by B, C and D to see if it is reproduceable. Thats been the cornerstone of research study now for a couple of centuries.

This blog site is not stating they have actually currently done so … Its a blog describing an intent to begin training a model like R1 and calling it Open-R1.

Also DeepSeek-R1 was only released recently, and even in their paper they outlined the calculate hours required. While those are low calculate hours for a SOTA design this does not suggest you can train stated model in a week. I ‘d personally love to be able to train a transformer model in a week, but we may need to wait a while for that level of calculate innovation.

So there are no criteria for a design that has not been built yet right? As detailed in the blog, and again in reply to your question.

However fear not, there is a GitHub Repo already and contributors (hell I might join myself), some prelim work done, and a strategy of attack. An excellent starting position.

n
@edbeeching
has actually assessed the released designs already

( src: https://x.com/edwardbeeching/status/1884273209136275742)

R1 simply trained on o1 outputs, so collectively …/ s. This is what the brand-new AI czars are stating

Hi! This post is an intro to the project, not a claim that we have actually recreated R1 yet. We will absolutely share the missing out on piece when we have them, you can expect the models and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo

That’s great and important to understand this significant buzz that lacks technical understanding and explanation. Science is about recreation, and if they claim to be open, let them fullfill the open part.

Please do publish the training expense.

We will!

Excalidraw Hi n
@bojan2501
thanks, we will undoubtedly be working hard to ensure this training dish can work for small language designs on consumer hardware since not everybody has a cluster of H100s in the house:-RRB- The tool we utilized for the images was Excalidraw! https://excalidraw.com

anticipating it! WTF are your speaking about?

must be a joke

It’s really cool to see how the entire open source neighborhood comes together!

Ops …

5.5 M is number reporter in the deepseekv3 tech report (simply the training, not the experiment afaik), for R1 hard to estimate tbh but much less than 5.5 M imo

Historically, they have never ever released code or datasets of their LLM training, so I would not anticipate this time to be different. If they would launch it that would be incredible of course!

Yes obviously!

So generally you’re asking to change existing censorship with another flavour of censorship?

The code for the models are inside the model repositories, e.g. for V3: https://huggingface.co/deepseek-ai/DeepSeek-V3/blob/main/modeling_deepseek.py

Hello Team, I’m Ray Bernard, the author and creator of EQUATOR. My research study group will be working on a paper concentrated on reproducing certain components of DeepSeek R1. Our goal is to replicate the cold start and offer your group with a dataset that includes COT and other techniques to support these efforts. We like to contribute our work to assist. Please let me know if you find this helpful. Best, Ray Bernard https://www.facebook.com/groups/1186310571520299/

Where is the assessment numbers? without it you can’t call it reproduction.

8 replies

True, but it looks like there’s absolutely nothing to be evaluated since right now. I assume the supreme objective is to train a new thinking design and then use the very same assessment metrics as o1 and the DeepSeek-R1.

That’s rather fascinating, I was asking myself why the concerns the author exposed here are not being asked by others? I think the work they have actually done is unforgettable but at the very same time I wonder why they would not put these missing out on pieces on if they are expected to be totally open.
Why even without recreation and comprehension of the development they could impact a lot the marketplace in this way?

4 replies

Hi! This article is an introduction to the task, not a claim that we have actually reproduced R1 yet. We will absolutely share the missing out on piece when we have them, you can expect the models and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo

Interesting read, and it is good that we see more effort into this direction: more optimization and less strength.
Also wonder what tool did the author usage for creating action diagram.

2 replies

Excalidraw I’m so happy that initiative like this currently exist, I’m gon na attempt to contribute:-RRB- 1 reply

eagerly anticipating it! So racist articel

2 replies

WTF are your speaking about?

Awesome to have this open recreation started!

For Step # 1 check out https://github.com/open-thoughts/open-thoughts!

https://x.com/ryanmart3n/status/1884284101265612856

Let’s do this thing!

1 reply

It’s actually cool to see how the whole open source community comes together!

Does anybody understand the real training expense of r1? I can’t discover it in the paper or the announcement post. Is the 6M cost reported by media just the number taken from v3’s training cost?

2 replies

Ops …

Has anyone asked the DeepSeek group to release their training information and code, or a minimum of share them privately with an independent duplication project like this? Have they declined such a demand?

A faithful duplication depends on utilizing the exact same dataset and hyperparameters. Otherwise, any major disparities with the published benchmarks would be hard to pin down-whether due to training information differences or the duplication method itself.

1 reply

Historically, they have actually never ever or datasets of their LLM training, so I wouldn’t anticipate this time to be various. If they would launch it that would be remarkable obviously!

In the meantime we need to make best guess quotes and see if we can arrive ourselves.

You provide excellent replication procedure of Deepseek thinking training. I will try something similar to it.

This is truly excellent info, can we fine tune with particular use case when code is launched?

1 reply

Yes of course!

Please consider getting rid of prejudiced, tainted or unaligned training data and make an effort to remove copyrighted works from the crawl from intake. This will make the model more functional. If you recycled anthropic curation checks, this might likewise assist, remove obviouslybiased data will likely include a great deal of worth. We don’t want another tainted, unaligned open source design, right? And no corporate would ever use deepseek or a design that reuses it, right?
We value your work for the advantage of humanity, we hope.
Miike C from NJ

1 reply

So essentially you’re asking to change existing censorship with another flavour of censorship?

Can’t wait! Hopefully the model will be uncensored but whatever you can do is alright! Love seeing open source structure itself up. I’m not smart sufficient to really assist but I can contribute support lol

Hello guys, I am even just searching for code for DeepSeek-V2, in order to fully understand multi-head hidden attention. You do not seem to have code in Hugging Face even for that. Or am I missing something? Don’t see anything in src/transformers/models. MLA is not appropriately explained in their paper, so it would be essential to have code for this.