BokantLM 0.1β0.5B
BokantLM β "Small but Supreme in Its Domain"
BokantLM is not a general-purpose model that tries to do everything well.
Instead, it is an ultra-lightweight LLM designed to focus on a single domain, delivering the highest possible efficiency and performance in that area.
Overview
- Model Name: BokantLM 0.1β0.5B
- Base Model: Qwen2.5-0.5B
- Fine-tuning Dataset:
greengerong/leetcode
Philosophy
While most LLMs aim for versatility by learning across many fields,
BokantLM is built to achieve top efficiency and performance within a specific domain.
This 0.1β0.5B release is specialized in coding and algorithm problem solving,
with a particular focus on LeetCode-style challenges.
The reason I created this model
I created this model based on the idea that if I focus intensively on learning only Python , even a small model could become very good at Python programming.
Future Plans
- β Coding(Python)-specialized model release (current version)
- π Mathematics problem-solving specialized version
- π Domain-specific ultra-lightweight models for law, medicine, science, etc.
- π Attempt at applying large LLM knowledge distillation
- Downloads last month
- 34