Commit History
Fix(cfg): Add validation for save_strategy and eval_strategy (#633)
383f88d
unverified
use fastchat conversations template (#578)
e7d3e2d
unverified
Fix: Fail bf16 check when running on cpu during merge (#631)
cfbce02
unverified
better handling and logging of empty sharegpt turns (#603)
a363604
unverified
misc fixes to add gptq tests (#621)
03e5907
unverified
Support Sample packing for phi arch (#586)
12a2dbb
unverified
E2e device cuda (#575)
2414673
unverified
e2e testing (#574)
9218ebe
unverified
Fix pretraining with iterable/streaming Dataset (#556)
2f586d1
unverified
Jan Philipp Harries
Jan Philipp Harries
commited on
workaround for md5 variations (#533)
0b4cf5b
unverified
recommend padding when using sample packing (#531)
3437149
unverified
fix test fixture b/c hf trainer tokenization changed (#464)
d5dcf9c
unverified
fix fixture for new tokenizer handling in transformers (#428)
8cace80
unverified
simplify `load_tokenizer`
efb3b2c
extract module for working with cfg
8cec513
fix DefaultDict.__or__
a13e45d
Attention mask and position id fixes for packing (#285)
2bb0b78
unverified
experimental llama 2 chat support (#296)
3392270
unverified
Jan Philipp Harries
Jan Philipp Harries
commited on