palmer

palmer

a better base model

This model is palmer-002-2401 scaled to 32k by merging and fine-tuning with TinyLlama-1.1B-32k-Instruct by Doctor-Shotgun

evaluation πŸ§ͺ

note that this is a zero-shot setting as opposite to open llm leaderboard's few-shot evals

   model           ARC-C     OBQA   HellaSwag  PIQA  Winogrande Average
tinyllama        | 0.3029 | 0.3600 | 0.5935 | 0.7329 | 0.5959 | 0.5170 |
palmer-002-2401  | 0.3294 | 0.3700 | 0.5950 | 0.7399 | 0.5896 | 0.5247 | 
palmer-002-32k   | 0.3268 | 0.3780 | 0.5785 | 0.7492 | 0.6251 | 0.5315 | (this)
babbage-002      | 0.3285 | 0.3620 | 0.6380 | 0.7606 | 0.6085 | 0.5395 |

This model's performance is close to openai's one while being capable of using 2x the context size.

prompt πŸ“

no prompt πŸš€

Buy Me A Coffee

Downloads last month
118
Safetensors
Model size
1.1B params
Tensor type
FP16
Β·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for appvoid/palmer-002-32k

Merges
12 models