Tech Talent Source

Dmuchane Zjezdzalnie

Overview

  • Founded Date September 4, 1946
  • Sectors Construction / Facilities
  • Posted Jobs 0
  • Viewed 17
Bottom Promo

Company Description

Open-R1: a Totally Open Reproduction Of DeepSeek-R1

Hey there! This blog post is an introduction to the job, not a claim that we have actually reproduced R1 yet. We’re constructing in the open, so as quickly as we have evaluation numbers, we’ll share them. You can follow our progress on Hugging Face and GitHub.

True, however it seems like there’s absolutely nothing to be assessed since right now. I presume the ultimate goal is to train a brand-new thinking model and then utilize the very same examination metrics as o1 and the DeepSeek-R1.

Well, there must be at least some peace of mind check and validation to make sure the design was trained properly.

Oh yes, if you are discussing the assessment number of deepseek’s model it’s coming soon!

As pointed out in the article there is no design called Open-R1 to test at all … not yet anyhow. This is a blog site outlining that Hugging face will take the R1 Deepseek design, work out how it was built as laid out in the paper and from what they released, and after that reproduce that procedure.

in fact this is basically how science works … A comes up with a plan, discovery or innovation and it is checked by B, C and D to see if it is reproduceable. Thats been the foundation of research now for a few centuries.

This blog site is not stating they have currently done so … Its a blog site describing an intent to start training a model like R1 and calling it Open-R1.

Also DeepSeek-R1 was only launched recently, and even in their paper they described the calculate hours needed. While those are low calculate hours for a SOTA model this does not imply you can train said model in a week. I ‘d personally like to be able to train a transformer design in a week, however we may require to wait a while for that level of calculate technology.

So there are no criteria for a design that has not been built yet right? As outlined in the blog site, and once again in reply to your question.

However fear not, there is a GitHub Repo currently and contributors (hell I may join myself), some prelim work done, and a strategy of attack. A great starting position.

n
@edbeeching
has actually assessed the released designs already

( src: https://x.com/edwardbeeching/status/1884273209136275742)

R1 just trained on o1 outputs, so collectively …/ s. This is what the new AI czars are stating

Hi! This article is an introduction to the task, not a claim that we have actually replicated R1 yet. We will absolutely share the missing out on piece when we have them, you can anticipate the designs and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo

That’s great and crucial to understand this incredible hype that does not have technical comprehension and description. Science is about reproduction, and if they claim to be open, let them fullfill the open part.

Please do publish the training cost.

We will!

Excalidraw Hi n
@bojan2501
thanks, we will indeed be to ensure this training recipe can work for small language designs on consumer hardware given that not everybody has a cluster of H100s in the house:-RRB- The tool we utilized for the images was Excalidraw! https://excalidraw.com

looking forward to it! WTF are your discussing?

should be a joke

It’s truly cool to see how the whole open source community comes together!

Ops …

5.5 M is number press reporter in the deepseekv3 tech report (just the training, not the experiment afaik), for R1 tough to estimate tbh however much less than 5.5 M imo

Historically, they have actually never released code or datasets of their LLM training, so I wouldn’t expect this time to be different. If they would launch it that would be incredible naturally!

Yes obviously!

So essentially you’re asking to change existing censorship with another flavour of censorship?

The code for the models are inside the model repositories, e.g. for V3: https://huggingface.co/deepseek-ai/DeepSeek-V3/blob/main/modeling_deepseek.py

Hello Team, I’m Ray Bernard, the author and developer of EQUATOR. My research study team will be working on a paper focused on duplicating certain components of DeepSeek R1. Our objective is to recreate the cold start and provide your group with a dataset that includes COT and other methods to support these efforts. We like to contribute our work to help. Please let me understand if you find this useful. Best, Ray Bernard https://www.facebook.com/groups/1186310571520299/

Where is the assessment numbers? without it you can’t call it reproduction.

8 replies

True, however it looks like there’s absolutely nothing to be assessed as of right now. I assume the supreme objective is to train a brand-new reasoning design and after that use the very same evaluation metrics as o1 and the DeepSeek-R1.

That’s quite intriguing, I was asking myself why the concerns the author exposed here are not being asked by others? I think the work they have actually done is unforgettable but at the same time I question why they would not put these missing out on pieces on if they are supposed to be completely open.
Why even without reproduction and comprehension of the development they could impact so much the market in this way?

4 replies

Hi! This article is an intro to the project, not a claim that we have actually recreated R1 yet. We will totally share the missing piece when we have them, you can anticipate the designs and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo

Interesting read, and it is great that we see more effort into this direction: more optimization and less strength.
Also wonder what tool did the author use for developing step diagram.

2 replies

Excalidraw I’m so delighted that effort like this already exist, I’m gon na attempt to contribute:-RRB- 1 reply

anticipating it! So racist articel

2 replies

WTF are your speaking about?

Awesome to have this open recreation started!

For Step # 1 check out https://github.com/open-thoughts/open-thoughts!

https://x.com/ryanmart3n/status/1884284101265612856

Let’s do this thing!

1 reply

It’s really cool to see how the entire open source community comes together!

Does anybody understand the actual training cost of r1? I can’t discover it in the paper or the announcement post. Is the 6M cost reported by media simply the number taken from v3’s training cost?

2 replies

Ops …

Has anyone asked the DeepSeek team to release their training data and code, or a minimum of share them independently with an independent replication project like this? Have they declined such a demand?

A loyal duplication depends on using the very same dataset and hyperparameters. Otherwise, any significant inconsistencies with the published benchmarks would be hard to pin down-whether due to training data distinctions or the replication method itself.

1 reply

Historically, they have never released code or datasets of their LLM training, so I would not expect this time to be different. If they would release it that would be amazing obviously!

In the meantime we need to make best guess estimates and see if we can get there ourselves.

You offer great replication process of Deepseek reasoning training. I will attempt something similar to it.

This is actually excellent information, can we tweak with particular usage case when code is launched?

1 reply

Yes naturally!

Please think about getting rid of biased, polluted or unaligned training information and make an effort to remove copyrighted works from the crawl from intake. This will make the design more usable. If you recycled anthropic curation checks, this might likewise assist, remove obviouslybiased data will likely add a lot of worth. We don’t want another polluted, unaligned open source model, right? And no business would ever use deepseek or a model that recycles it, right?
We appreciate your work for the benefit of mankind, we hope.
Miike C from NJ

1 reply

So basically you’re asking to replace existing censorship with another flavour of censorship?

Can’t wait! Hopefully the model will be uncensored however whatever you can do is alright! Love seeing open source building itself up. I’m not smart adequate to actually help however I can contribute moral support lol

Hello guys, I am even just trying to discover code for DeepSeek-V2, in order to completely understand multi-head hidden attention. You do not seem to have code in Hugging Face even for that. Or am I missing out on something? Don’t see anything in src/transformers/models. MLA is not appropriately explained in their paper, so it would be essential to have code for this.

Bottom Promo
Bottom Promo
Top Promo