This is just a preview model. It is a finetuned gemma-2b with added chatml tokens.
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 54.63 |
AI2 Reasoning Challenge (25-Shot) | 52.56 |
HellaSwag (10-Shot) | 73.65 |
MMLU (5-Shot) | 46.02 |
TruthfulQA (0-shot) | 51.25 |
Winogrande (5-shot) | 66.38 |
GSM8k (5-shot) | 37.91 |
- Downloads last month
- 9
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Model tree for pansophic/gemma-2b-sft-preview
Base model
google/gemma-2bDatasets used to train pansophic/gemma-2b-sft-preview
Evaluation results
- normalized accuracy on AI2 Reasoning Challenge (25-Shot)test set Open LLM Leaderboard52.560
- normalized accuracy on HellaSwag (10-Shot)validation set Open LLM Leaderboard73.650
- accuracy on MMLU (5-Shot)test set Open LLM Leaderboard46.020
- mc2 on TruthfulQA (0-shot)validation set Open LLM Leaderboard51.250
- accuracy on Winogrande (5-shot)validation set Open LLM Leaderboard66.380
- accuracy on GSM8k (5-shot)test set Open LLM Leaderboard37.910