|
--- |
|
tags: |
|
- autotrain |
|
- text-generation-inference |
|
- text-generation |
|
library_name: transformers |
|
base_model: Qwen/Qwen2.5-0.5B |
|
widget: |
|
- messages: |
|
- role: user |
|
content: What is your favorite condiment? |
|
license: apache-2.0 |
|
datasets: |
|
- greengerong/leetcode |
|
--- |
|
|
|
.png) |
|
|
|
# BokantLM 0.1β0.5B |
|
|
|
## BokantLM β "Small but Supreme in Its Domain" |
|
|
|
BokantLM is **not** a general-purpose model that tries to do everything well. |
|
Instead, it is an **ultra-lightweight LLM** designed to focus on a single domain, delivering the **highest possible efficiency and performance** in that area. |
|
|
|
--- |
|
|
|
## Overview |
|
|
|
- **Model Name:** BokantLM 0.1β0.5B |
|
- **Base Model:** [Qwen2.5-0.5B](https://huggingface.co/Qwen/Qwen2.5-0.5B) |
|
- **Fine-tuning Dataset:** [`greengerong/leetcode`](https://huggingface.co/datasets/greengerong/leetcode) |
|
|
|
--- |
|
|
|
## Philosophy |
|
|
|
While most LLMs aim for versatility by learning across many fields, |
|
BokantLM is built to **achieve top efficiency and performance within a specific domain**. |
|
|
|
This **0.1β0.5B release** is specialized in **coding and algorithm problem solving**, |
|
with a particular focus on **LeetCode-style challenges**. |
|
|
|
--- |
|
|
|
## The reason I created this model |
|
|
|
I created this model based on the idea that **if I focus intensively on learning only Python** , even a **small model** could become **very good at Python programming.** |
|
|
|
--- |
|
|
|
## Future Plans |
|
|
|
- β
**Coding(Python)-specialized** model release (current version) |
|
- π Mathematics problem-solving specialized version |
|
- π Domain-specific ultra-lightweight models for **law, medicine, science**, etc. |
|
- π **Attempt** at applying large LLM knowledge distillation |
|
|
|
--- |
|
|