SafetyBERT
SafetyBERT is a BERT model fine-tuned on occupational safety data from MSHA, OSHA, NTSB, and other safety organizations, as well as a large corpus of occupational safety-related Abstracts.
Quick Start
from transformers import AutoTokenizer, AutoModelForMaskedLM
tokenizer = AutoTokenizer.from_pretrained("bert-base-cased")
model = AutoModelForMaskedLM.from_pretrained("adanish91/safetybert")
# Example usage
text = "The worker failed to wear proper [MASK] equipment."
inputs = tokenizer(text, return_tensors="pt")
outputs = model(**inputs)
Model Details
- Base Model: bert-base-uncased
- Parameters: 110M
- Training Data: 2.4M safety documents from multiple sources
- Specialization: Mining, construction, transportation safety
Performance
Significantly outperforms BERT-base on safety classification tasks:
- 76.9% improvement in pseudo-perplexity
- Superior performance on Occupational safety-related downstream tasks
Applications
- Safety document analysis
- Incident report classification
- Downloads last month
- 11
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for adanish91/safetybert
Base model
google-bert/bert-base-uncased