Commit History
add support for https remote yamls (#1277)
9bca7db
unverified
Peft deepspeed resume (#1227)
c67fb71
unverified
Peft lotfq (#1222)
4cb7900
unverified
Revert "run PR e2e docker CI tests in Modal" (#1220) [skip ci]
8da1633
unverified
run PR e2e docker CI tests in Modal (#1217) [skip ci]
36d053f
unverified
Update deps 202401 (#1204) [skip ci]
a01b998
unverified
upgrade deepspeed to 0.13.1 for mixtral fixes (#1189) [skip ci]
8a49309
unverified
Qwen2 (#1166)
f5a828a
unverified
Remove fused-dense-lib from requirements.txt (#1087)
91502b9
unverified
fix: warn user to install mamba_ssm package (#1019)
d69ba2b
unverified
pin accelerate for deepspeed fix (#1080)
9e3f0cb
unverified
Separate AutoGPTQ dep to `pip install -e .[auto-gptq]` (#1077)
9be92d1
unverified
paired kto support (#1069)
d7057cc
unverified
update peft to 0.7.0 (#1073)
768d348
unverified
Add: mlflow for experiment tracking (#1059) [skip ci]
090c24d
unverified
Phi2 rewrite (#1058)
732851f
unverified
RL/DPO (#935)
f243c21
bump transformers and update attention class map name (#1023)
bcc78d8
unverified
chore: Update transformers to latest (#986)
7d4185f
unverified
update transformers to fix checkpoint saving (#963)
f28e755
unverified
dumpmemory
commited on
Mixtral official (#942)
7fabc4d
unverified
Update requirements.txt (#940)
9a5eb39
unverified
update to latest transformers for mixstral support (#929)
35f9b0f
unverified
update datasets version to cut down the warnings due to pyarrow arg change (#897)
6a4562a
unverified
try #2: pin hf transformers and accelerate to latest release, don't reinstall pytorch (#867)
0de1457
unverified
Feat: Add dataset loading from S3, GCS (#765)
3cc67d2
unverified
add e2e tests for checking functionality of resume from checkpoint (#865)
b3a61e8
unverified
Pin optimum package (#838)
105d0b3
unverified
Bryan Thornbury
commited on
don't compile deepspeed or bitsandbytes from source (#837)
f544ab2
unverified
Feat: Added Gradio support (#812)
738a057
unverified
fix: pin autogptq (#818)
6459ac7
unverified
chore: bump transformers to v4.34.1 to fix tokenizer issue (#745)
8966a6f
unverified
pin xformers >= 0.0.22 (#724)
bfbdba8
unverified
Fix(version): Update FA to work with Mistral SWA (#673)
43856c0
unverified
Feat: Allow usage of native Mistral FA when no sample_packing (#669)
697c50d
unverified
removed duplicate on requirements.txt (#661)
a7e56d8
unverified
Napuh
commited on
add mistral e2e tests (#649)
5b0bc48
unverified
Mistral flash attn packing (#646)
b6ab8aa
unverified
use fastchat conversations template (#578)
e7d3e2d
unverified
Feat: Add support for upstream FA2 (#626)
19a600a
unverified
update README w deepspeed info (#605)
c25ba79
unverified
Update requirements.txt (#610)
ec0958f
unverified
Javier
commited on
fix wandb so mypy doesn't complain (#562)
bf08044
unverified
Update requirements.txt (#543)
c1921c9
unverified
dongxiaolong
commited on
update readme to point to direct link to runpod template, cleanup install instrucitons (#532)
34c0a86
unverified
Add support for GPTQ using native transformers/peft (#468)
3355706
unverified
add eval benchmark callback (#441)
7657632
unverified
customizable ascii art (#506)
548787d
unverified
Fix missing 'packaging' wheel (#482)
c500d02
unverified
Maxime
commited on