File size: 4,305 Bytes
df5a66d 5fb9d5b df5a66d 5fb9d5b f5b1185 5fb9d5b df5a66d 5fb9d5b df5a66d 591c412 df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 94138d6 df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d 5fb9d5b df5a66d ce29b23 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 |
---
language:
- en
tags:
- qwen2
- text-generation
- summarization
- key-points
- blog-summarization
datasets:
- ncls-p/blog-key-points
license: cc-by-4.0
base_model: Qwen/Qwen2.5-3B-Instruct
---
# Qwen2.5-3B-blog-key-points
This model is fine-tuned from [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct) on the [ncls-p/blog-key-points](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct). It specializes in extracting key points from blog articles and web content, providing concise bullet-point summaries that capture the essential information.
## Model Description
**Qwen2.5-3B-blog-key-points** is a 3B parameter model fine-tuned specifically for the task of extracting key points from articles. It can process a full article and generate a concise, bullet-point summary highlighting the most important information.
### Model Details
- **Model Type:** Qwen2.5 (3B parameters)
- **Base Model:** [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct)
- **Training Dataset:** [ncls-p/blog-key-points](https://huggingface.co/datasets/ncls-p/blog-key-points)
- **Language:** English
- **License:** [CC-BY-4.0](https://creativecommons.org/licenses/by/4.0/)
- **Finetuning Approach:** Instruction fine-tuning on article-summary pairs
## Uses
### Direct Use
This model is designed for extracting key points from articles. You can use it directly for:
- Summarizing blog posts
- Extracting important information from news articles
- Creating bullet-point summaries of long-form content
- Generating concise overviews of research papers
### Example Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "ncls-p/Qwen2.5-3B-blog-key-points"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)
article = """
[Your article text here]
"""
prompt = f"""
Extract the key points from the following article:
{article}
"""
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=1024)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
```
## Training
The model was fine-tuned on the [blog-key-points dataset](https://huggingface.co/datasets/ncls-p/blog-key-points), which contains 200 article-summary pairs. Each pair consists of a full article and a bullet-point summary of key points extracted using AI.
### Training Procedure
- **Fine-tuning Framework:** [Unsloth](https://github.com/unslothai/unsloth)
- **Training Data Format:**
```json
{
"instruction": "",
"input": "Full article content",
"output": "Here are the key points of the article:\n* Key point 1\n* Key point 2\n* Key point 3\n..."
}
```
## Evaluation
The model was evaluated on its ability to extract relevant key points from articles not seen during training. Evaluation metrics focused on:
1. **Relevance:** How well the extracted points capture the main ideas of the article
2. **Conciseness:** The ability to summarize information in a clear, bullet-point format
3. **Completeness:** Whether all important information is captured in the summary
## Limitations and Biases
- The model may inherit biases present in the training data, including potential biases in the source articles or in the key point extraction process.
- Performance may vary depending on the length, complexity, and domain of the input article.
- The model is primarily trained on English-language content and may not perform well on content in other languages.
- As with any summarization model, there is a risk of omitting important information or misrepresenting the original content.
## How to Cite
If you use this model in your research, please cite:
```bibtex
@misc{qwen25-3b-blog-key-points,
author = {ncls-p},
title = {Qwen2.5-3B-blog-key-points},
year = {2024},
publisher = {Hugging Face},
journal = {Hugging Face model repository},
howpublished = {\url{https://huggingface.co/ncls-p/Qwen2.5-3B-blog-key-points}},
}
```
## Dataset Creation
The dataset used to train this model was created using the [llm-to-blog-key-points-dataset](https://github.com/ncls-p/llm-to-blog-key-points-dataset), a CLI tool that extracts key points from web articles using AI and adds them to a dataset in a structured format. |