language:
- en
license: cc-by-sa-3.0
tags:
- treecorpus
- wikipedia
- encyclopedia
- knowledge-base
- factual-knowledge
- training-data
- conversational-ai
- nlp
- language-model
- text-corpus
- qa-dataset
- structured-data
- large-scale
pretty_name: 'TreeCorpus: Wikipedia Knowledge for AI Models'
size_categories:
- 10M<n<100M
TreeCorpus
TreeCorpus is a comprehensive, structured dataset derived from the latest Wikipedia dumps, specially processed to serve as high-quality training data for conversational AI models. This dataset transforms Wikipedia's encyclopedic knowledge into a format optimized for natural language understanding and generation tasks.
Dataset Statistics
- Size: 26.27 GB (26,272,580,250 bytes)
- Examples: 2,882,766 articles
- Download Size: 13.33 GB (13,326,529,312 bytes)
- Language: English
Data Structure
Each entry in the dataset contains:
id
(string): Unique Wikipedia article identifiertitle
(string): Article titletext
(string): Clean, processed text contenturl
(string): Source Wikipedia URLtimestamp
(string): Processing timestamp
Key Features
- Clean, Structured Content: Meticulously processed to remove markup, templates, references, and other non-content elements while preserving the informational value of Wikipedia articles.
- Rich Metadata: Each entry includes article ID, title, clean text content, source URL, and timestamp.
- Comprehensive Coverage: Incorporates the full spectrum of Wikipedia's knowledge base, spanning nearly 3 million articles across countless topics.
- Conversational Optimization: Content is processed specifically to support training of dialogue systems, conversational agents, and knowledge-grounded language models.
- Regular Updates: Built from the latest Wikipedia dumps to ensure current information.
Usage
This dataset is ideal for:
- Training large language models requiring broad knowledge bases
- Fine-tuning conversational agents for knowledge-intensive tasks
- Question-answering systems that need factual grounding
- Research in knowledge representation and retrieval in natural language
License and Citation
TreeCorpus is derived from Wikipedia content available under the CC BY-SA 3.0 license. When using this dataset, please provide appropriate attribution to both this dataset and Wikipedia.
Dataset Configuration
The dataset is configured with a default split:
- Split name: train
- Data files pattern: data/train-*
Creation Process
TreeCorpus was created using a specialized pipeline that:
- Downloads the latest Wikipedia dumps
- Processes XML content to extract articles
- Cleans and standardizes text by removing markup, templates, and non-content elements
- Structures data in a consistent, machine-readable format
- Filters out redirects, stubs, and non-article content
For more details on the methodology and processing pipeline, please see the accompanying code documentation.