|
--- |
|
library_name: transformers |
|
license: apache-2.0 |
|
language: |
|
- ja |
|
pipeline_tag: fill-mask |
|
widget: |
|
- text: 彼のダンス、めっちゃ[MASK]!😂 |
|
--- |
|
|
|
# BERT for Japanese Twitter |
|
|
|
This is a base BERT model that has been adapted for Japanese Twitter. |
|
|
|
It was adapted from [Japanese BERT](https://huggingface.co/tohoku-nlp/bert-base-japanese-v3) by preparing a specialized vocabulary and continuing pretraining on a Twitter corpus. |
|
|
|
This model is reccomended for Japanese SNS tasks, like [sentiment analysis](https://huggingface.co/datasets/shunk031/wrime) and [defamation detection](https://huggingface.co/datasets/kubota/defamation-japanese-twitter). |
|
|
|
|
|
This model has been used to finetune a series of models. The main ones are [BERT for Japanese Twitter Sentiment](https://huggingface.co/LoneWolfgang/bert-for-japanese-twitter-sentiment) and [BERT for Japanese Twitter Emotion](https://huggingface.co/LoneWolfgang/bert-for-japanese-twitter-emotion). |
|
## Training Data |
|
|
|
The Twitter API was used to collect Japanese tweets from June 2022 to April 2023. |
|
|
|
N-gram based deduplication was used to reduce spam content and improve the diversity of the training corpus. |
|
The refined training corpus was 28 million tweets. |
|
|
|
## Tokenization |
|
|
|
The vocabulary was prepared using the [WordPieceTrainer](https://huggingface.co/docs/tokenizers/api/trainers) with the Twitter training corpus. |
|
It shares 60% of its vocabulary with Japanese BERT. |
|
|
|
The vocabulary includes colloquialisms, neologisms, emoji and kaomoji expressions that are common on Twitter. |