Highly advanced based model for training:

  • Sequence Length: 131072
  • Parm 2 ultra: trained for 2 hours on 1 Million OpenO1 chats, 180k sonnet 3.5, 130k qwq messages.
Downloads last month
8
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for Pinkstack/PARM-v2-ULTRA-o1-3B-vLLM

Base model

Qwen/Qwen2.5-3B
Finetuned
(1)
this model

Collection including Pinkstack/PARM-v2-ULTRA-o1-3B-vLLM