distilbert-med-v2 / README.md
sfarrukh's picture
Update README.md
401fe72 verified
---
library_name: transformers
license: apache-2.0
base_model: distilbert/distilbert-base-uncased
tags:
- masked_language_modeling,
- domain_addaption
model-index:
- name: distilbert-med-v2
results: []
datasets:
- FreedomIntelligence/medical-o1-reasoning-SFT
language:
- en
---
# Model Card: Fine-Tuned Language Model for Medical Reasoning
## **Model Overview**
This language model has been fine-tuned on the [FreedomIntelligence/medical-o1-reasoning-SFT](https://huggingface.co/datasets/FreedomIntelligence/medical-o1-reasoning-SFT). The fine-tuning process focused on enhancing the model's ability to reason and provide contextually accurate predictions within the medical domain. Domain adaptation was achieved through fine-tuning the pretrained model on the medical reasoning dataset using the Masked Language Modeling (MLM) objective. This approach allowed the model to better understand domain-specific terminology and context by predicting masked tokens based on surrounding words. This model can assist in tasks requiring reasoning over medical text, such as medical Q&A, report generation, or other domain-specific NLP applications.
---
## **Intended Use**
### **Primary Applications**
- Medical reasoning tasks.
- Text completion or generation in the medical domain.
- Context-aware masked token prediction for medical content.
### **Limitations**
- This model is fine-tuned for **medical reasoning** and might not generalize well to other domains.
- It is not a substitute for professional medical advice or diagnosis.
### **Usage Warnings**
- Outputs should always be reviewed by qualified professionals.
- The model may produce incorrect or misleading results when provided with ambiguous or incomplete context.
---
## **Fine-Tuning Details**
### **Pretrained Model**
- Base model:[ `distilbert/distilbert-base-uncased`](https://huggingface.co/distilbert/distilbert-base-uncased)
### **Fine-Tuning Objective**
The model was fine-tuned using the **Masked Language Modeling (MLM)** objective. Both **subword masking** and **whole word masking** strategies were experimented with, but no significant performance difference was observed between the two approaches.
### **Dataset**
- Dataset: [FreedomIntelligence/medical-o1-reasoning-SFT](https://huggingface.co/datasets/FreedomIntelligence/medical-o1-reasoning-SFT)
- Dataset Description: This dataset is designed for reasoning tasks in the medical domain and includes diverse medical scenarios.
### **Training Setup**
- **Hardware**: T4 GPU (Google Colab)
- **Software**: - Transformers 4.48.1
- Pytorch 2.5.1+cu121
- Datasets 3.2.0
- Tokenizers 0.21.0
---
## **Performance Metrics**
- **Evaluation Metrics**: Cross-entropy loss and perplexity were used to evaluate model performance.
- **Results**:
- Cross-Entropy Loss: 1.2834
- Perplexity: 3.6
### **Observations**
- Both subword masking and whole word masking yielded similar performance metrics, indicating no significant advantage of one masking strategy over the other.
---
### **Limitations and Biases**
- The model's performance is dependent on the quality and representativeness of the dataset.
- Potential biases in the dataset may propagate into the model outputs.
- The model might not handle rare or out-of-distribution medical terms effectively.
## Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 1.4437 | 1.0 | 263 | 1.4178 |
| 1.4236 | 2.0 | 526 | 1.3673 |
| 1.4316 | 3.0 | 789 | 1.3320 |
| 1.4174 | 4.0 | 1052 | 1.3123 |
| 1.4015 | 5.0 | 1315 | 1.2988 |
| 1.3922 | 6.0 | 1578 | 1.2834 |