File size: 485 Bytes
72fa063 acc410b 402cd40 f520aa9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 |
---
license: apache-2.0
---
The model is pretrained on the OSCAR dataset for Bangla, English and Hindi.
The base model is Distil-BERT and the intended use for this model is for the datasets that contain a mix of these languages.
To Cite:
@article{raihan2023mixed,
title={Mixed-Distil-BERT: Code-mixed Language Modeling for Bangla, English, and Hindi},
author={Raihan, Md Nishat and Goswami, Dhiman and Mahmud, Antara},
journal={arXiv preprint arXiv:2309.10272},
year={2023}
} |