Marco Zocca

ocramz
·

AI & ML interests

Program understanding, languages and compilers

Recent Activity

liked a model 4 days ago
Qwen/Qwen2.5-Coder-3B-Instruct
liked a model 4 days ago
Qwen/Qwen2.5-VL-3B-Instruct
liked a model 5 days ago
manycore-research/SpatialLM-Llama-1B
View all activity

Organizations

UnfoldML's profile picture BigCode's profile picture

ocramz's activity

reacted to onekq's post with 👍 2 months ago
view post
Post
2306
So 🐋DeepSeek🐋 hits the mainstream media. But it has been a star in our little cult for at least 6 months. Its meteoric success is not overnight, but two years in the making.

To learn their history, just look at their 🤗 repo deepseek-ai

* End of 2023, they launched the first model (pretrained by themselves) following Llama 2 architecture
* June 2024, v2 (MoE architecture) surpassed Gemini 1.5, but behind Mistral
* September, v2.5 surpassed GPT 4o mini
* December, v3 surpassed GPT 4o
* Now R1 surpassed o1

Most importantly, if you think DeepSeek success is singular and unrivaled, that's WRONG. The following models are also near or equal the o1 bar.

* Minimax-01
* Kimi k1.5
* Doubao 1.5 pro
  • 1 reply
·