Commit History
Fix: Warn when fullfinetune without adapter (#770)
44c9d01
unverified
convert exponential notation lr to floats (#771)
ca84cca
unverified
Fix: eval table conflict with eval_sample_packing (#769)
9923b72
unverified
Implement fused modules (#747)
15d3a65
unverified
refactor to set eval_batch_size earlier if unset, so we can warn if mismatched (#662)
2642cae
unverified
Make dataset_processes configurable (#651)
9ec2077
unverified
Fix bug when using pretokenized datasets (#652)
590d603
unverified
ich
commited on
Feat: Add example for Mistral (#644)
eb41f76
unverified
Fix(cfg): Add validation for save_strategy and eval_strategy (#633)
383f88d
unverified
use fastchat conversations template (#578)
e7d3e2d
unverified
Feat: Add support for upstream FA2 (#626)
19a600a
unverified
Fix: Fail bf16 check when running on cpu during merge (#631)
cfbce02
unverified
add bf16 check (#587)
131afdb
unverified
make phi training work with Loras (#588)
62eaee7
unverified
E2e device cuda (#575)
2414673
unverified
Model parallel (#538)
f6060a6
unverified
Add training callback to send predictions to WandB table (#521)
5b67ea9
unverified
Fix pretraining with iterable/streaming Dataset (#556)
2f586d1
unverified
Jan Philipp Harries
Jan Philipp Harries
commited on