Description
This model is a specialized adaptation of the facebook/bart-large-xsum, fine-tuned for enhanced performance on dialogue summarization using the SamSum dataset.
Development
- Kaggle Notebook: Text Summarization with Large Language Models
Usage
from transformers import pipeline
model = pipeline("summarization", model="luisotorres/bart-finetuned-samsum")
conversation = '''Sarah: Do you think it's a good idea to invest in Bitcoin?
Emily: I'm skeptical. The market is very volatile, and you could lose money.
Sarah: True. But there's also a high upside, right?
'''
model(conversation)
Training Parameters
evaluation_strategy = "epoch",
save_strategy = 'epoch',
load_best_model_at_end = True,
metric_for_best_model = 'eval_loss',
seed = 42,
learning_rate=2e-5,
per_device_train_batch_size=4,
per_device_eval_batch_size=4,
gradient_accumulation_steps=2,
weight_decay=0.01,
save_total_limit=2,
num_train_epochs=4,
predict_with_generate=True,
fp16=True,
report_to="none"
Reference
This model is based on the original BART architecture, as detailed in:
Lewis et al. (2019). BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. arXiv:1910.13461
- Downloads last month
- 278
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Dataset used to train luisotorres/bart-finetuned-samsum
Spaces using luisotorres/bart-finetuned-samsum 6
Evaluation results
- Validation ROUGE-1 on SamSumself-reported53.880
- Validation ROUGE-2 on SamSumself-reported29.233
- Validation ROUGE-L on SamSumself-reported44.774
- Validation ROUGE-L Sum on SamSumself-reported49.825
- Test ROUGE-1 on SamSumself-reported52.816
- Test ROUGE-2 on SamSumself-reported28.126
- Test ROUGE-L on SamSumself-reported43.715
- Test ROUGE-L Sum on SamSumself-reported48.571