Trained the eos_token into the lm_head.
This should allow qlora finetunes with 24 or even 16 GB of vram.
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
HF Inference API was unable to determine this model's library.
Model tree for gghfez/Qwen2.5-14B-Base-lm_head-LoRA
Base model
Qwen/Qwen2.5-14B