Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
3
1
oussama yousr
oussama
Follow
0 followers
·
3 following
AI & ML interests
None yet
Recent Activity
reacted
to
prithivMLmods
's
post
with 🚀
about 19 hours ago
Reasoning SmolLM2 🚀 🎯Fine-tuning SmolLM2 on a lightweight synthetic reasoning dataset for reasoning-specific tasks. Future updates will focus on lightweight, blazing-fast reasoning models. Until then, check out the blog for fine-tuning details. 🔥Blog : https://huggingface.co/blog/prithivMLmods/smollm2-ft 🔼 Models : + SmolLM2-CoT-360M : https://huggingface.co/prithivMLmods/SmolLM2-CoT-360M + Reasoning-SmolLM2-135M : https://huggingface.co/prithivMLmods/Reasoning-SmolLM2-135M + SmolLM2-CoT-360M-GGUF : https://huggingface.co/prithivMLmods/SmolLM2-CoT-360M-GGUF 🤠 Other Details : + Demo : https://huggingface.co/prithivMLmods/SmolLM2-CoT-360M/blob/main/Demo/SmolLM2%20Demo.ipynb + Fine-tune nB : https://huggingface.co/prithivMLmods/SmolLM2-CoT-360M/blob/main/finetune/SmolLM-FT.ipynb
reacted
to
fdaudens
's
post
with ❤️
about 19 hours ago
Yes, DeepSeek R1's release is impressive. But the real story is what happened in just 7 days after: - Original release: 8 models, 540K downloads. Just the beginning... - The community turned those open-weight models into +550 NEW models on Hugging Face. Total downloads? 2.5M—nearly 5X the originals. The reason? DeepSeek models are open-weight, letting anyone build on top of them. Interesting to note that the community focused on quantized versions for better efficiency & accessibility. They want models that use less memory, run faster, and are more energy-efficient. When you empower builders, innovation explodes. For everyone. 🚀 The most popular community model? @bartowski's DeepSeek-R1-Distill-Qwen-32B-GGUF version — 1M downloads alone.
reacted
to
lewtun
's
post
with ❤️
about 19 hours ago
Introducing OpenR1-Math-220k! https://huggingface.co/datasets/open-r1/OpenR1-Math-220k The community has been busy distilling DeepSeek-R1 from inference providers, but we decided to have a go at doing it ourselves from scratch 💪 What’s new compared to existing reasoning datasets? ♾ Based on https://huggingface.co/datasets/AI-MO/NuminaMath-1.5: we focus on math reasoning traces and generate answers for problems in NuminaMath 1.5, an improved version of the popular NuminaMath-CoT dataset. 🐳 800k R1 reasoning traces: We generate two answers for 400k problems using DeepSeek R1. The filtered dataset contains 220k problems with correct reasoning traces. 📀 512 H100s running locally: Instead of relying on an API, we leverage vLLM and SGLang to run generations locally on our science cluster, generating 180k reasoning traces per day. ⏳ Automated filtering: We apply Math Verify to only retain problems with at least one correct answer. We also leverage Llama3.3-70B-Instruct as a judge to retrieve more correct examples (e.g for cases with malformed answers that can’t be verified with a rules-based parser) 📊 We match the performance of DeepSeek-Distill-Qwen-7B by finetuning Qwen-7B-Math-Instruct on our dataset. 🔎 Read our blog post for all the nitty gritty details: https://huggingface.co/blog/open-r1/update-2
View all activity
Organizations
None yet
oussama
's activity
All
Models
Datasets
Spaces
Papers
Collections
Community
Posts
Upvotes
Likes
Articles
upvoted
an
article
about 19 hours ago
view article
Article
Open R1: Update #3
By
open-r1
and 9 others
•
3 days ago
•
204