|
--- |
|
language: |
|
- en |
|
license: apache-2.0 |
|
task_categories: |
|
- other |
|
tags: |
|
- topic-modeling |
|
- llm-evaluation |
|
- benchmark |
|
- legislation |
|
- wikipedia |
|
--- |
|
|
|
# Dataset Overview |
|
|
|
This repository contains benchmark datasets for evaluating Large Language Model (LLM)-based topic discovery methods and comparing them against traditional topic models. These datasets provide a valuable resource for researchers studying topic modeling and LLM capabilities in this domain. The work is described in the following paper: [Large Language Models Struggle to Describe the Haystack without Human Help: Human-in-the-loop Evaluation of LLMs](https://arxiv.org/abs/2502.14748). Original data source: [GitHub](https://github.com/ahoho/topics?tab=readme-ov-file#download-data) |
|
|
|
## [Bills Dataset](https://huggingface.co/datasets/zli12321/Bills) |
|
|
|
The Bills Dataset is a collection of legislative documents containing 32,661 bill summaries (train) from the 110th–114th U.S. Congresses, categorized into 21 top-level and 112 secondary-level topics. A test split of 15.2K summaries is also included. |
|
|
|
### Loading the Bills Dataset |
|
```python |
|
from datasets import load_dataset |
|
|
|
# Load the train and test splits |
|
train_dataset = load_dataset('zli12321/Bills', split='train') |
|
test_dataset = load_dataset('zli12321/Bills', split='test') |
|
``` |
|
|
|
## [Wiki Dataset](https://huggingface.co/datasets/zli12321/Wiki) |
|
|
|
The Wiki dataset consists of 14,290 articles spanning 15 high-level and 45 mid-level topics, including widely recognized public topics such as music and anime. A test split of 8.02K summaries is included. |
|
|
|
## Synthetic Science Fiction (Pending internal clearance process) |
|
|
|
Please cite the relevant papers below if you find the data useful. Do not hesitate to create an issue or email us if you have problems! |
|
|
|
|
|
**Citation:** |
|
|
|
If you find LLM-based topic generation has hallucination or instability, and coherence not applicable to LLM-based topic models: |
|
``` |
|
@misc{li2025largelanguagemodelsstruggle, |
|
title={Large Language Models Struggle to Describe the Haystack without Human Help: Human-in-the-loop Evaluation of LLMs}, |
|
author={Zongxia Li and Lorena Calvo-Bartolomé and Alexander Hoyle and Paiheng Xu and Alden Dima and Juan Francisco Fung and Jordan Boyd-Graber}, |
|
year={2025}, |
|
eprint={2502.14748}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.CL}, |
|
url={https://arxiv.org/abs/2502.14748}, |
|
} |
|
``` |
|
|
|
If you use the human annotations or preprocessing: |
|
``` |
|
@inproceedings{li-etal-2024-improving, |
|
title = "Improving the {TENOR} of Labeling: Re-evaluating Topic Models for Content Analysis", |
|
author = "Li, Zongxia and |
|
Mao, Andrew and |
|
Stephens, Daniel and |
|
Goel, Pranav and |
|
Walpole, Emily and |
|
Dima, Alden and |
|
Fung, Juan and |
|
Boyd-Graber, Jordan", |
|
editor = "Graham, Yvette and |
|
Purver, Matthew", |
|
booktitle = "Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)", |
|
month = mar, |
|
year = "2024", |
|
address = "St. Julian{'}s, Malta", |
|
publisher = "Association for Computational Linguistics", |
|
url = "https://aclanthology.org/2024.eacl-long.51/", |
|
pages = "840--859" |
|
} |
|
``` |
|
|
|
If you want to use the claim coherence does not generalize to neural topic models: |
|
``` |
|
@inproceedings{hoyle-etal-2021-automated, |
|
title = "Is Automated Topic Evaluation Broken? The Incoherence of Coherence", |
|
author = "Hoyle, Alexander Miserlis and |
|
Goel, Pranav and |
|
Hian-Cheong, Andrew and |
|
Peskov, Denis and |
|
Boyd-Graber, Jordan and |
|
Resnik, Philip", |
|
booktitle = "Advances in Neural Information Processing Systems", |
|
year = "2021", |
|
url = "https://arxiv.org/abs/2107.02173", |
|
} |
|
``` |
|
|
|
|
|
If you evaluate ground-truth evaluations or stability: |
|
``` |
|
@inproceedings{hoyle-etal-2022-neural, |
|
title = "Are Neural Topic Models Broken?", |
|
author = "Hoyle, Alexander Miserlis and |
|
Goel, Pranav and |
|
Sarkar, Rupak and |
|
Resnik, Philip", |
|
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2022", |
|
year = "2022", |
|
publisher = "Association for Computational Linguistics", |
|
url = "https://aclanthology.org/2022.findings-emnlp.390", |
|
doi = "10.18653/v1/2022.findings-emnlp.390", |
|
pages = "5321--5344", |
|
} |
|
``` |