Update README.md
Browse files
README.md
CHANGED
|
@@ -12,6 +12,11 @@ tags:
|
|
| 12 |
- text-generation-inference
|
| 13 |
---
|
| 14 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
|
| 16 |
```py
|
| 17 |
pip install transformers
|
|
@@ -52,4 +57,24 @@ generated_text = generate_text(prompt, max_length=68)
|
|
| 52 |
|
| 53 |
# Print the generated text
|
| 54 |
print(generated_text)
|
| 55 |
-
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
- text-generation-inference
|
| 13 |
---
|
| 14 |
|
| 15 |
+
# **Gpt2-Wikitext-9180**
|
| 16 |
+
|
| 17 |
+
> **Gpt2-Wikitext-9180**, fine-tuned from GPT-2, is a Transformer-based language model trained on a large English corpus (WikiText) using self-supervised learning. This means it was trained on raw, unlabeled text data, using an automated process to create inputs and labels by predicting the next word in a sentence. No manual annotation was involved, allowing the model to leverage a vast amount of publicly available data.
|
| 18 |
+
|
| 19 |
+
## Demo Inference
|
| 20 |
|
| 21 |
```py
|
| 22 |
pip install transformers
|
|
|
|
| 57 |
|
| 58 |
# Print the generated text
|
| 59 |
print(generated_text)
|
| 60 |
+
```
|
| 61 |
+
|
| 62 |
+
|
| 63 |
+
---
|
| 64 |
+
|
| 65 |
+
### **Intended Use Case**
|
| 66 |
+
|
| 67 |
+
* **Text Generation**: Auto-completion, story generation, or dialogue simulation.
|
| 68 |
+
* **Language Modeling**: Understanding language structure and context for downstream NLP tasks.
|
| 69 |
+
* **Educational and Research Use**: Exploring fine-tuning techniques, language understanding, or benchmarking language models.
|
| 70 |
+
* **Prototyping**: Quick deployment of language-based features in applications and interfaces.
|
| 71 |
+
|
| 72 |
+
---
|
| 73 |
+
|
| 74 |
+
### **Limitations**
|
| 75 |
+
|
| 76 |
+
* **Factual Inaccuracy**: May generate plausible-sounding but incorrect or outdated information.
|
| 77 |
+
* **Bias and Toxicity**: Can reflect biases present in training data (e.g., stereotypes, offensive language).
|
| 78 |
+
* **Context Length**: Limited context window inherited from GPT-2 architecture.
|
| 79 |
+
* **Not Real-Time Aware**: Lacks access to current events or updates beyond its training data.
|
| 80 |
+
* **Lack of Understanding**: Generates text based on patterns, not genuine comprehension or reasoning.
|