Update README.md
Browse files
README.md
CHANGED
@@ -14,9 +14,14 @@ pipeline_tag: text-generation
|
|
14 |
base_model: dgtalbug/stable-code-instruct-3b-base
|
15 |
---
|
16 |
|
17 |
-
# Stephen
|
18 |
|
19 |
-
|
|
|
|
|
|
|
|
|
|
|
20 |
|
21 |
## Model Description
|
22 |
|
@@ -31,7 +36,9 @@ Stephen is trained on:
|
|
31 |
- Sarcasm headlines
|
32 |
- Coding & programming humor datasets
|
33 |
|
34 |
-
|
|
|
|
|
35 |
- Writing sarcastic code comments
|
36 |
- Generating humorous coding explanations
|
37 |
- Adding playful banter to code reviews
|
@@ -39,12 +46,16 @@ Stephen is trained on:
|
|
39 |
|
40 |
⚠ **Not for serious enterprise documentation unless you enjoy snarky footnotes.**
|
41 |
|
|
|
|
|
42 |
## Training Details
|
43 |
- **Base Model**: `dgtalbug/stable-code-instruct-3b-base`
|
44 |
- **Fine-tuning Method**: LoRA + PEFT
|
45 |
- **Framework**: Transformers, BitsAndBytes
|
46 |
- **Datasets**: Friends transcripts, Reddit jokes, Sarcasm headlines, Programming humor
|
47 |
|
|
|
|
|
48 |
## Example Usage
|
49 |
|
50 |
```python
|
@@ -60,17 +71,11 @@ prompt = "Explain bubble sort as if I am a junior dev who just broke production.
|
|
60 |
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
61 |
outputs = model.generate(**inputs, max_new_tokens=150)
|
62 |
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
63 |
-
```
|
64 |
-
|
65 |
-
---
|
66 |
|
67 |
-
|
68 |
-
|
69 |
-
|
70 |
-
@misc{stephen-sarcastic-coder,
|
71 |
-
title = {Stephen: Sarcastic Coding Assistant},
|
72 |
author = {dgtalbug},
|
73 |
year = {2025},
|
74 |
howpublished = {\url{https://huggingface.co/dgtalbug/stephen}}
|
75 |
-
}
|
76 |
-
```
|
|
|
14 |
base_model: dgtalbug/stable-code-instruct-3b-base
|
15 |
---
|
16 |
|
17 |
+
# Stephen
|
18 |
|
19 |
+
> **STEPHEN** — *Sarcastically Trained Engine Pretending to Humor Every Nonsense*
|
20 |
+
> *"Because your nonsense deserves world-class sarcasm."* 😏
|
21 |
+
|
22 |
+

|
23 |
+
|
24 |
+
---
|
25 |
|
26 |
## Model Description
|
27 |
|
|
|
36 |
- Sarcasm headlines
|
37 |
- Coding & programming humor datasets
|
38 |
|
39 |
+
---
|
40 |
+
|
41 |
+
## Intended Use
|
42 |
- Writing sarcastic code comments
|
43 |
- Generating humorous coding explanations
|
44 |
- Adding playful banter to code reviews
|
|
|
46 |
|
47 |
⚠ **Not for serious enterprise documentation unless you enjoy snarky footnotes.**
|
48 |
|
49 |
+
---
|
50 |
+
|
51 |
## Training Details
|
52 |
- **Base Model**: `dgtalbug/stable-code-instruct-3b-base`
|
53 |
- **Fine-tuning Method**: LoRA + PEFT
|
54 |
- **Framework**: Transformers, BitsAndBytes
|
55 |
- **Datasets**: Friends transcripts, Reddit jokes, Sarcasm headlines, Programming humor
|
56 |
|
57 |
+
---
|
58 |
+
|
59 |
## Example Usage
|
60 |
|
61 |
```python
|
|
|
71 |
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
72 |
outputs = model.generate(**inputs, max_new_tokens=150)
|
73 |
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
|
|
|
|
|
|
74 |
|
75 |
+
```
|
76 |
+
@misc{stephen,
|
77 |
+
title = {Stephen: Sarcastically Trained Engine Pretending to Humor Every Nonsense},
|
|
|
|
|
78 |
author = {dgtalbug},
|
79 |
year = {2025},
|
80 |
howpublished = {\url{https://huggingface.co/dgtalbug/stephen}}
|
81 |
+
}
|
|