|
--- |
|
library_name: transformers |
|
license: apache-2.0 |
|
base_model: distilbert/distilbert-base-uncased |
|
tags: |
|
- masked_language_modeling, |
|
- domain_addaption |
|
model-index: |
|
- name: distilbert-med-v2 |
|
results: [] |
|
datasets: |
|
- FreedomIntelligence/medical-o1-reasoning-SFT |
|
language: |
|
- en |
|
--- |
|
|
|
# Model Card: Fine-Tuned Language Model for Medical Reasoning |
|
|
|
## **Model Overview** |
|
This language model has been fine-tuned on the [FreedomIntelligence/medical-o1-reasoning-SFT](https://huggingface.co/datasets/FreedomIntelligence/medical-o1-reasoning-SFT). The fine-tuning process focused on enhancing the model's ability to reason and provide contextually accurate predictions within the medical domain. Domain adaptation was achieved through fine-tuning the pretrained model on the medical reasoning dataset using the Masked Language Modeling (MLM) objective. This approach allowed the model to better understand domain-specific terminology and context by predicting masked tokens based on surrounding words. This model can assist in tasks requiring reasoning over medical text, such as medical Q&A, report generation, or other domain-specific NLP applications. |
|
|
|
|
|
--- |
|
|
|
## **Intended Use** |
|
### **Primary Applications** |
|
- Medical reasoning tasks. |
|
- Text completion or generation in the medical domain. |
|
- Context-aware masked token prediction for medical content. |
|
|
|
### **Limitations** |
|
- This model is fine-tuned for **medical reasoning** and might not generalize well to other domains. |
|
- It is not a substitute for professional medical advice or diagnosis. |
|
|
|
### **Usage Warnings** |
|
- Outputs should always be reviewed by qualified professionals. |
|
- The model may produce incorrect or misleading results when provided with ambiguous or incomplete context. |
|
|
|
--- |
|
|
|
## **Fine-Tuning Details** |
|
### **Pretrained Model** |
|
- Base model:[ `distilbert/distilbert-base-uncased`](https://huggingface.co/distilbert/distilbert-base-uncased) |
|
|
|
### **Fine-Tuning Objective** |
|
The model was fine-tuned using the **Masked Language Modeling (MLM)** objective. Both **subword masking** and **whole word masking** strategies were experimented with, but no significant performance difference was observed between the two approaches. |
|
|
|
|
|
### **Dataset** |
|
- Dataset: [FreedomIntelligence/medical-o1-reasoning-SFT](https://huggingface.co/datasets/FreedomIntelligence/medical-o1-reasoning-SFT) |
|
- Dataset Description: This dataset is designed for reasoning tasks in the medical domain and includes diverse medical scenarios. |
|
|
|
### **Training Setup** |
|
- **Hardware**: T4 GPU (Google Colab) |
|
- **Software**: - Transformers 4.48.1 |
|
- Pytorch 2.5.1+cu121 |
|
- Datasets 3.2.0 |
|
- Tokenizers 0.21.0 |
|
|
|
--- |
|
|
|
## **Performance Metrics** |
|
- **Evaluation Metrics**: Cross-entropy loss and perplexity were used to evaluate model performance. |
|
- **Results**: |
|
- Cross-Entropy Loss: 1.2834 |
|
- Perplexity: 3.6 |
|
|
|
### **Observations** |
|
- Both subword masking and whole word masking yielded similar performance metrics, indicating no significant advantage of one masking strategy over the other. |
|
|
|
--- |
|
|
|
### **Limitations and Biases** |
|
- The model's performance is dependent on the quality and representativeness of the dataset. |
|
- Potential biases in the dataset may propagate into the model outputs. |
|
- The model might not handle rare or out-of-distribution medical terms effectively. |
|
|
|
|
|
|
|
|
|
## Training results |
|
|
|
| Training Loss | Epoch | Step | Validation Loss | |
|
|:-------------:|:-----:|:----:|:---------------:| |
|
| 1.4437 | 1.0 | 263 | 1.4178 | |
|
| 1.4236 | 2.0 | 526 | 1.3673 | |
|
| 1.4316 | 3.0 | 789 | 1.3320 | |
|
| 1.4174 | 4.0 | 1052 | 1.3123 | |
|
| 1.4015 | 5.0 | 1315 | 1.2988 | |
|
| 1.3922 | 6.0 | 1578 | 1.2834 | |
|
|