llaa33219 commited on
Commit
1ae81e0
Β·
verified Β·
1 Parent(s): 28140e6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -22
README.md CHANGED
@@ -14,34 +14,50 @@ datasets:
14
  - greengerong/leetcode
15
  ---
16
 
17
- # Model Trained Using AutoTrain
18
 
19
- This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain).
20
 
21
- # Usage
22
 
23
- ```python
24
 
25
- from transformers import AutoModelForCausalLM, AutoTokenizer
 
 
 
 
 
 
 
 
 
 
26
 
27
- model_path = "PATH_TO_THIS_REPO"
 
28
 
29
- tokenizer = AutoTokenizer.from_pretrained(model_path)
30
- model = AutoModelForCausalLM.from_pretrained(
31
- model_path,
32
- device_map="auto",
33
- torch_dtype='auto'
34
- ).eval()
35
 
36
- # Prompt content: "hi"
37
- messages = [
38
- {"role": "user", "content": "hi"}
39
- ]
 
 
 
 
 
 
 
 
 
 
 
40
 
41
- input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt')
42
- output_ids = model.generate(input_ids.to('cuda'))
43
- response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True)
44
 
45
- # Model response: "Hello! How can I assist you today?"
46
- print(response)
47
- ```
 
14
  - greengerong/leetcode
15
  ---
16
 
17
+ ![BokantLM 둜고](https://raw.githubusercontent.com/llaa33219/BokantLM-logo/refs/heads/main/New%20Project%20(6).png)
18
 
19
+ # BokantLM 0.1-0.5B
20
 
21
+ **BokantLM**λŠ” *"Bokant"* (μ•„ν”„λ¦¬μΉΈμŠ€μ–΄λ‘œ "μœ„μ—" β†’ "졜고")λΌλŠ” μ΄λ¦„μ²˜λŸΌ, **μž‘μ§€λ§Œ νŠΉμ • μž‘μ—…μ—μ„œ 졜고의 μ„±λŠ₯을 λͺ©ν‘œλ‘œ ν•˜λŠ” LLM**μž…λ‹ˆλ‹€.
22
 
23
+ 이 ν”„λ‘œμ νŠΈμ˜ 핡심 철학은 **"λͺ¨λ“  것을 μž˜ν•˜λŠ” λͺ¨λΈ"이 μ•„λ‹ˆλΌ, "ν•˜λ‚˜μ— 집쀑해 μž˜ν•˜λŠ” λͺ¨λΈ"**을 λ§Œλ“œλŠ” κ²ƒμž…λ‹ˆλ‹€.
24
 
25
+ ---
26
+
27
+ ## ​ λͺ¨λΈ κ°œμš”
28
+
29
+ - **λͺ¨λΈλͺ…:** BokantLM 0.1-0.5B
30
+ - **기반 λͺ¨λΈ:** [Qwen2.5-0.5B](https://huggingface.co/Qwen/Qwen2.5-0.5B)
31
+ - **νŒŒμΈνŠœλ‹ 데이터:** [`greengerong/leetcode`](https://huggingface.co/datasets/greengerong/leetcode)
32
+
33
+ ---
34
+
35
+ ## ​ μ˜λ„μ™€ μ² ν•™
36
 
37
+ λŒ€λΆ€λΆ„μ˜ LLM은 λ²”μš©μ„±μ„ μœ„ν•΄ λ‹€μ–‘ν•œ λΆ„μ•Όλ₯Ό ν•™μŠ΅ν•˜μ§€λ§Œ,
38
+ BokantLM은 **νŠΉμ • λΆ„μ•Ό ν•œμ •μœΌλ‘œ μ΅œμƒμ˜ 효율과 μ„±λŠ₯**을 λ…Έλ¦½λ‹ˆλ‹€.
39
 
40
+ 이번 **0.1-0.5B 버전**은 μ½”λ”©/μ•Œκ³ λ¦¬μ¦˜ λΆ„μ•Όλ₯Ό 집쀑 ν•™μŠ΅ν•˜μ˜€μœΌλ©°,
41
+ 특히 **LeetCode μŠ€νƒ€μΌμ˜ 문제 ν•΄κ²°**에 μ΅œμ ν™”λ˜μ–΄ μžˆμŠ΅λ‹ˆλ‹€.
42
+
43
+ ---
 
 
44
 
45
+ ## ​ ν–₯ν›„ κ³„νš
46
+
47
+ - βœ… **μ½”λ”© νŠΉν™”** λͺ¨λΈ μΆœμ‹œ (ν˜„μž¬ 버전)
48
+ - πŸ”„ μˆ˜ν•™ 문제 풀이 νŠΉν™” 버전
49
+ - πŸ”„ 법λ₯ , μ˜ν•™, κ³Όν•™ λ“± **뢄야별 μ΄ˆκ²½λŸ‰ νŠΉν™” λͺ¨λΈ**
50
+ - πŸ”„ **λŒ€ν˜• LLM 지식 증λ₯˜(Distillation) 적용 μ‹œλ„**
51
+
52
+ ---
53
+
54
+ ## ​ λΌμ΄μ„ μŠ€
55
+
56
+ 이 μ €μž₯μ†Œ 및 λͺ¨λΈμ€ 연ꡬ 및 비상업적 μš©λ„λ‘œλ§Œ μ‚¬μš© κ°€λŠ₯ν•©λ‹ˆλ‹€.
57
+ 상업적 μ‚¬μš©μ€ 별도 문의 λ°”λžλ‹ˆλ‹€.
58
+
59
+ ---
60
 
61
+ ## BokantLM - "Small but Supreme in Its Domain"
 
 
62
 
63
+ BokantLM은 λͺ¨λ“  것을 μž˜ν•˜λŠ” λ²”μš© λͺ¨λΈμ΄ μ•„λ‹ˆλΌ, **ν•œ 뢄야에 집쀑해 졜고의 효율과 μ„±λŠ₯을 λ‚΄λŠ” 것을 λͺ©ν‘œλ‘œ ν•˜λŠ” μ΄ˆκ²½λŸ‰ LLM λͺ¨λΈ**μž…λ‹ˆλ‹€.