Commit History
support to disable exllama for gptq (#604)
faecff9
unverified
Delete duplicate lines (#606)
aa656e0
unverified
btlm and falcon monkey patches for flash attn (#566)
6b9b229
unverified
make phi training work with Loras (#588)
62eaee7
unverified
don't resize embeddings if it's already large enough (#577)
3607882
unverified
Support Sample packing for phi arch (#586)
12a2dbb
unverified
Add training callback to send predictions to WandB table (#521)
5b67ea9
unverified
fix for quant config from model (#540)
a94f9cb
unverified
Add support for GPTQ using native transformers/peft (#468)
3355706
unverified
fix: bad dtype for full finetune (#504)
1991946
unverified
Refactor train cfg cli (#499)
125cccb
unverified
simplify linear layer locator
267b7b2
fsdp requires params be the same type too (#493)
98bf76e
unverified
Fix(tokenizer): Make sure to add pad for CodeLlamaTokenizer (#489)
4c37bd0
unverified
fix condition and add logging
3a011ea
rename var and reformat
f319b0b
Update src/axolotl/utils/models.py
7fd662d
unverified
Update src/axolotl/utils/models.py
9e69968
unverified
ignore: address pr review
d03887f
unverified
Maxime
commited on
ignore: linter
a184549
unverified
Maxime
commited on
fix: finetune model inference needs the dtype fix to work with flash-attn
f311df9
unverified
Maxime
commited on